Apr 30 12:42:12.161771 kernel: Booting Linux on physical CPU 0x0000120000 [0x413fd0c1] Apr 30 12:42:12.161794 kernel: Linux version 6.6.88-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT Tue Apr 29 22:28:35 -00 2025 Apr 30 12:42:12.161803 kernel: KASLR enabled Apr 30 12:42:12.161809 kernel: efi: EFI v2.7 by American Megatrends Apr 30 12:42:12.161814 kernel: efi: ACPI 2.0=0xec080000 SMBIOS 3.0=0xf0a1ff98 ESRT=0xea465818 RNG=0xebf10018 MEMRESERVE=0xe4624e18 Apr 30 12:42:12.161819 kernel: random: crng init done Apr 30 12:42:12.161826 kernel: secureboot: Secure boot disabled Apr 30 12:42:12.161832 kernel: esrt: Reserving ESRT space from 0x00000000ea465818 to 0x00000000ea465878. Apr 30 12:42:12.161839 kernel: ACPI: Early table checksum verification disabled Apr 30 12:42:12.161845 kernel: ACPI: RSDP 0x00000000EC080000 000024 (v02 Ampere) Apr 30 12:42:12.161851 kernel: ACPI: XSDT 0x00000000EC070000 0000A4 (v01 Ampere Altra 00000000 AMI 01000013) Apr 30 12:42:12.161857 kernel: ACPI: FACP 0x00000000EC050000 000114 (v06 Ampere Altra 00000000 INTL 20190509) Apr 30 12:42:12.161863 kernel: ACPI: DSDT 0x00000000EBFF0000 019B57 (v02 Ampere Jade 00000001 INTL 20200717) Apr 30 12:42:12.161869 kernel: ACPI: DBG2 0x00000000EC060000 00005C (v00 Ampere Altra 00000000 INTL 20190509) Apr 30 12:42:12.161877 kernel: ACPI: GTDT 0x00000000EC040000 000110 (v03 Ampere Altra 00000000 INTL 20190509) Apr 30 12:42:12.161883 kernel: ACPI: SSDT 0x00000000EC030000 00002D (v02 Ampere Altra 00000001 INTL 20190509) Apr 30 12:42:12.161890 kernel: ACPI: FIDT 0x00000000EBFE0000 00009C (v01 ALASKA A M I 01072009 AMI 00010013) Apr 30 12:42:12.161896 kernel: ACPI: SPCR 0x00000000EBFD0000 000050 (v02 ALASKA A M I 01072009 AMI 0005000F) Apr 30 12:42:12.161902 kernel: ACPI: BGRT 0x00000000EBFC0000 000038 (v01 ALASKA A M I 01072009 AMI 00010013) Apr 30 12:42:12.161908 kernel: ACPI: MCFG 0x00000000EBFB0000 0000AC (v01 Ampere Altra 00000001 AMP. 01000013) Apr 30 12:42:12.161914 kernel: ACPI: IORT 0x00000000EBFA0000 000610 (v00 Ampere Altra 00000000 AMP. 01000013) Apr 30 12:42:12.161920 kernel: ACPI: PPTT 0x00000000EBF80000 006E60 (v02 Ampere Altra 00000000 AMP. 01000013) Apr 30 12:42:12.161926 kernel: ACPI: SLIT 0x00000000EBF70000 00002D (v01 Ampere Altra 00000000 AMP. 01000013) Apr 30 12:42:12.161932 kernel: ACPI: SRAT 0x00000000EBF60000 0006D0 (v03 Ampere Altra 00000000 AMP. 01000013) Apr 30 12:42:12.161940 kernel: ACPI: APIC 0x00000000EBF90000 0019F4 (v05 Ampere Altra 00000003 AMI 01000013) Apr 30 12:42:12.161946 kernel: ACPI: PCCT 0x00000000EBF40000 000576 (v02 Ampere Altra 00000003 AMP. 01000013) Apr 30 12:42:12.161964 kernel: ACPI: WSMT 0x00000000EBF30000 000028 (v01 ALASKA A M I 01072009 AMI 00010013) Apr 30 12:42:12.161971 kernel: ACPI: FPDT 0x00000000EBF20000 000044 (v01 ALASKA A M I 01072009 AMI 01000013) Apr 30 12:42:12.161977 kernel: ACPI: SPCR: console: pl011,mmio32,0x100002600000,115200 Apr 30 12:42:12.161983 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x88300000-0x883fffff] Apr 30 12:42:12.161989 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x90000000-0xffffffff] Apr 30 12:42:12.161995 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0x8007fffffff] Apr 30 12:42:12.162001 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80100000000-0x83fffffffff] Apr 30 12:42:12.162008 kernel: NUMA: NODE_DATA [mem 0x83fdffca800-0x83fdffcffff] Apr 30 12:42:12.162013 kernel: Zone ranges: Apr 30 12:42:12.162021 kernel: DMA [mem 0x0000000088300000-0x00000000ffffffff] Apr 30 12:42:12.162027 kernel: DMA32 empty Apr 30 12:42:12.162033 kernel: Normal [mem 0x0000000100000000-0x0000083fffffffff] Apr 30 12:42:12.162039 kernel: Movable zone start for each node Apr 30 12:42:12.162045 kernel: Early memory node ranges Apr 30 12:42:12.162054 kernel: node 0: [mem 0x0000000088300000-0x00000000883fffff] Apr 30 12:42:12.162061 kernel: node 0: [mem 0x0000000090000000-0x0000000091ffffff] Apr 30 12:42:12.162068 kernel: node 0: [mem 0x0000000092000000-0x0000000093ffffff] Apr 30 12:42:12.162075 kernel: node 0: [mem 0x0000000094000000-0x00000000eba36fff] Apr 30 12:42:12.162081 kernel: node 0: [mem 0x00000000eba37000-0x00000000ebeadfff] Apr 30 12:42:12.162088 kernel: node 0: [mem 0x00000000ebeae000-0x00000000ebeaefff] Apr 30 12:42:12.162094 kernel: node 0: [mem 0x00000000ebeaf000-0x00000000ebeccfff] Apr 30 12:42:12.162100 kernel: node 0: [mem 0x00000000ebecd000-0x00000000ebecdfff] Apr 30 12:42:12.162107 kernel: node 0: [mem 0x00000000ebece000-0x00000000ebecffff] Apr 30 12:42:12.162113 kernel: node 0: [mem 0x00000000ebed0000-0x00000000ec0effff] Apr 30 12:42:12.162119 kernel: node 0: [mem 0x00000000ec0f0000-0x00000000ec0fffff] Apr 30 12:42:12.162126 kernel: node 0: [mem 0x00000000ec100000-0x00000000ee53ffff] Apr 30 12:42:12.162134 kernel: node 0: [mem 0x00000000ee540000-0x00000000f765ffff] Apr 30 12:42:12.162140 kernel: node 0: [mem 0x00000000f7660000-0x00000000f784ffff] Apr 30 12:42:12.162146 kernel: node 0: [mem 0x00000000f7850000-0x00000000f7fdffff] Apr 30 12:42:12.162153 kernel: node 0: [mem 0x00000000f7fe0000-0x00000000ffc8efff] Apr 30 12:42:12.162159 kernel: node 0: [mem 0x00000000ffc8f000-0x00000000ffc8ffff] Apr 30 12:42:12.162165 kernel: node 0: [mem 0x00000000ffc90000-0x00000000ffffffff] Apr 30 12:42:12.162172 kernel: node 0: [mem 0x0000080000000000-0x000008007fffffff] Apr 30 12:42:12.162178 kernel: node 0: [mem 0x0000080100000000-0x0000083fffffffff] Apr 30 12:42:12.162184 kernel: Initmem setup node 0 [mem 0x0000000088300000-0x0000083fffffffff] Apr 30 12:42:12.162191 kernel: On node 0, zone DMA: 768 pages in unavailable ranges Apr 30 12:42:12.162197 kernel: On node 0, zone DMA: 31744 pages in unavailable ranges Apr 30 12:42:12.162205 kernel: psci: probing for conduit method from ACPI. Apr 30 12:42:12.162212 kernel: psci: PSCIv1.1 detected in firmware. Apr 30 12:42:12.162218 kernel: psci: Using standard PSCI v0.2 function IDs Apr 30 12:42:12.162225 kernel: psci: MIGRATE_INFO_TYPE not supported. Apr 30 12:42:12.162231 kernel: psci: SMC Calling Convention v1.2 Apr 30 12:42:12.162237 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Apr 30 12:42:12.162244 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100 -> Node 0 Apr 30 12:42:12.162250 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x10000 -> Node 0 Apr 30 12:42:12.162256 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x10100 -> Node 0 Apr 30 12:42:12.162263 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x20000 -> Node 0 Apr 30 12:42:12.162269 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x20100 -> Node 0 Apr 30 12:42:12.162275 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x30000 -> Node 0 Apr 30 12:42:12.162283 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x30100 -> Node 0 Apr 30 12:42:12.162290 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x40000 -> Node 0 Apr 30 12:42:12.162296 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x40100 -> Node 0 Apr 30 12:42:12.162302 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x50000 -> Node 0 Apr 30 12:42:12.162309 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x50100 -> Node 0 Apr 30 12:42:12.162315 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x60000 -> Node 0 Apr 30 12:42:12.162321 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x60100 -> Node 0 Apr 30 12:42:12.162327 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x70000 -> Node 0 Apr 30 12:42:12.162334 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x70100 -> Node 0 Apr 30 12:42:12.162340 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x80000 -> Node 0 Apr 30 12:42:12.162346 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x80100 -> Node 0 Apr 30 12:42:12.162352 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x90000 -> Node 0 Apr 30 12:42:12.162360 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x90100 -> Node 0 Apr 30 12:42:12.162366 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xa0000 -> Node 0 Apr 30 12:42:12.162373 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xa0100 -> Node 0 Apr 30 12:42:12.162379 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xb0000 -> Node 0 Apr 30 12:42:12.162386 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xb0100 -> Node 0 Apr 30 12:42:12.162392 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xc0000 -> Node 0 Apr 30 12:42:12.162399 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xc0100 -> Node 0 Apr 30 12:42:12.162405 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xd0000 -> Node 0 Apr 30 12:42:12.162411 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xd0100 -> Node 0 Apr 30 12:42:12.162418 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xe0000 -> Node 0 Apr 30 12:42:12.162424 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xe0100 -> Node 0 Apr 30 12:42:12.162432 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xf0000 -> Node 0 Apr 30 12:42:12.162439 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0xf0100 -> Node 0 Apr 30 12:42:12.162445 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100000 -> Node 0 Apr 30 12:42:12.162451 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x100100 -> Node 0 Apr 30 12:42:12.162458 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x110000 -> Node 0 Apr 30 12:42:12.162464 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x110100 -> Node 0 Apr 30 12:42:12.162470 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x120000 -> Node 0 Apr 30 12:42:12.162477 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x120100 -> Node 0 Apr 30 12:42:12.162483 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x130000 -> Node 0 Apr 30 12:42:12.162489 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x130100 -> Node 0 Apr 30 12:42:12.162495 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x140000 -> Node 0 Apr 30 12:42:12.162502 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x140100 -> Node 0 Apr 30 12:42:12.162510 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x150000 -> Node 0 Apr 30 12:42:12.162516 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x150100 -> Node 0 Apr 30 12:42:12.162522 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x160000 -> Node 0 Apr 30 12:42:12.162529 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x160100 -> Node 0 Apr 30 12:42:12.162535 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x170000 -> Node 0 Apr 30 12:42:12.162542 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x170100 -> Node 0 Apr 30 12:42:12.162548 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x180000 -> Node 0 Apr 30 12:42:12.162554 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x180100 -> Node 0 Apr 30 12:42:12.162567 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x190000 -> Node 0 Apr 30 12:42:12.162573 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x190100 -> Node 0 Apr 30 12:42:12.162582 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1a0000 -> Node 0 Apr 30 12:42:12.162588 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1a0100 -> Node 0 Apr 30 12:42:12.162595 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1b0000 -> Node 0 Apr 30 12:42:12.162602 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1b0100 -> Node 0 Apr 30 12:42:12.162609 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1c0000 -> Node 0 Apr 30 12:42:12.162615 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1c0100 -> Node 0 Apr 30 12:42:12.162623 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1d0000 -> Node 0 Apr 30 12:42:12.162630 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1d0100 -> Node 0 Apr 30 12:42:12.162637 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1e0000 -> Node 0 Apr 30 12:42:12.162644 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1e0100 -> Node 0 Apr 30 12:42:12.162650 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1f0000 -> Node 0 Apr 30 12:42:12.162657 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1f0100 -> Node 0 Apr 30 12:42:12.162664 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x200000 -> Node 0 Apr 30 12:42:12.162671 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x200100 -> Node 0 Apr 30 12:42:12.162677 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x210000 -> Node 0 Apr 30 12:42:12.162684 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x210100 -> Node 0 Apr 30 12:42:12.162691 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x220000 -> Node 0 Apr 30 12:42:12.162697 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x220100 -> Node 0 Apr 30 12:42:12.162705 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x230000 -> Node 0 Apr 30 12:42:12.162712 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x230100 -> Node 0 Apr 30 12:42:12.162719 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x240000 -> Node 0 Apr 30 12:42:12.162726 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x240100 -> Node 0 Apr 30 12:42:12.162732 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x250000 -> Node 0 Apr 30 12:42:12.162739 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x250100 -> Node 0 Apr 30 12:42:12.162746 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x260000 -> Node 0 Apr 30 12:42:12.162752 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x260100 -> Node 0 Apr 30 12:42:12.162759 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x270000 -> Node 0 Apr 30 12:42:12.162766 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x270100 -> Node 0 Apr 30 12:42:12.162773 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Apr 30 12:42:12.162781 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Apr 30 12:42:12.162788 kernel: pcpu-alloc: [0] 00 [0] 01 [0] 02 [0] 03 [0] 04 [0] 05 [0] 06 [0] 07 Apr 30 12:42:12.162795 kernel: pcpu-alloc: [0] 08 [0] 09 [0] 10 [0] 11 [0] 12 [0] 13 [0] 14 [0] 15 Apr 30 12:42:12.162802 kernel: pcpu-alloc: [0] 16 [0] 17 [0] 18 [0] 19 [0] 20 [0] 21 [0] 22 [0] 23 Apr 30 12:42:12.162808 kernel: pcpu-alloc: [0] 24 [0] 25 [0] 26 [0] 27 [0] 28 [0] 29 [0] 30 [0] 31 Apr 30 12:42:12.162815 kernel: pcpu-alloc: [0] 32 [0] 33 [0] 34 [0] 35 [0] 36 [0] 37 [0] 38 [0] 39 Apr 30 12:42:12.162822 kernel: pcpu-alloc: [0] 40 [0] 41 [0] 42 [0] 43 [0] 44 [0] 45 [0] 46 [0] 47 Apr 30 12:42:12.162829 kernel: pcpu-alloc: [0] 48 [0] 49 [0] 50 [0] 51 [0] 52 [0] 53 [0] 54 [0] 55 Apr 30 12:42:12.162836 kernel: pcpu-alloc: [0] 56 [0] 57 [0] 58 [0] 59 [0] 60 [0] 61 [0] 62 [0] 63 Apr 30 12:42:12.162842 kernel: pcpu-alloc: [0] 64 [0] 65 [0] 66 [0] 67 [0] 68 [0] 69 [0] 70 [0] 71 Apr 30 12:42:12.162849 kernel: pcpu-alloc: [0] 72 [0] 73 [0] 74 [0] 75 [0] 76 [0] 77 [0] 78 [0] 79 Apr 30 12:42:12.162857 kernel: Detected PIPT I-cache on CPU0 Apr 30 12:42:12.162864 kernel: CPU features: detected: GIC system register CPU interface Apr 30 12:42:12.162870 kernel: CPU features: detected: Virtualization Host Extensions Apr 30 12:42:12.162877 kernel: CPU features: detected: Hardware dirty bit management Apr 30 12:42:12.162884 kernel: CPU features: detected: Spectre-v4 Apr 30 12:42:12.162891 kernel: CPU features: detected: Spectre-BHB Apr 30 12:42:12.162897 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 30 12:42:12.162904 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 30 12:42:12.162911 kernel: CPU features: detected: ARM erratum 1418040 Apr 30 12:42:12.162918 kernel: CPU features: detected: SSBS not fully self-synchronizing Apr 30 12:42:12.162924 kernel: alternatives: applying boot alternatives Apr 30 12:42:12.162932 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=packet flatcar.autologin verity.usrhash=984055eb0c340c9cf0fb51b368030ed72e75b7f2e065edc13766888ef0b42074 Apr 30 12:42:12.162941 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Apr 30 12:42:12.162948 kernel: printk: log_buf_len individual max cpu contribution: 4096 bytes Apr 30 12:42:12.162982 kernel: printk: log_buf_len total cpu_extra contributions: 323584 bytes Apr 30 12:42:12.162989 kernel: printk: log_buf_len min size: 262144 bytes Apr 30 12:42:12.162995 kernel: printk: log_buf_len: 1048576 bytes Apr 30 12:42:12.163002 kernel: printk: early log buf free: 249864(95%) Apr 30 12:42:12.163009 kernel: Dentry cache hash table entries: 16777216 (order: 15, 134217728 bytes, linear) Apr 30 12:42:12.163016 kernel: Inode-cache hash table entries: 8388608 (order: 14, 67108864 bytes, linear) Apr 30 12:42:12.163023 kernel: Fallback order for Node 0: 0 Apr 30 12:42:12.163030 kernel: Built 1 zonelists, mobility grouping on. Total pages: 65996028 Apr 30 12:42:12.163038 kernel: Policy zone: Normal Apr 30 12:42:12.163045 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 30 12:42:12.163052 kernel: software IO TLB: area num 128. Apr 30 12:42:12.163059 kernel: software IO TLB: mapped [mem 0x00000000fbc8f000-0x00000000ffc8f000] (64MB) Apr 30 12:42:12.163066 kernel: Memory: 262923412K/268174336K available (10368K kernel code, 2186K rwdata, 8100K rodata, 38336K init, 897K bss, 5250924K reserved, 0K cma-reserved) Apr 30 12:42:12.163073 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=80, Nodes=1 Apr 30 12:42:12.163080 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 30 12:42:12.163088 kernel: rcu: RCU event tracing is enabled. Apr 30 12:42:12.163095 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=80. Apr 30 12:42:12.163102 kernel: Trampoline variant of Tasks RCU enabled. Apr 30 12:42:12.163109 kernel: Tracing variant of Tasks RCU enabled. Apr 30 12:42:12.163116 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 30 12:42:12.163124 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=80 Apr 30 12:42:12.163131 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 30 12:42:12.163138 kernel: GICv3: GIC: Using split EOI/Deactivate mode Apr 30 12:42:12.163145 kernel: GICv3: 672 SPIs implemented Apr 30 12:42:12.163152 kernel: GICv3: 0 Extended SPIs implemented Apr 30 12:42:12.163159 kernel: Root IRQ handler: gic_handle_irq Apr 30 12:42:12.163166 kernel: GICv3: GICv3 features: 16 PPIs Apr 30 12:42:12.163172 kernel: GICv3: CPU0: found redistributor 120000 region 0:0x00001001005c0000 Apr 30 12:42:12.163179 kernel: SRAT: PXM 0 -> ITS 0 -> Node 0 Apr 30 12:42:12.163186 kernel: SRAT: PXM 0 -> ITS 1 -> Node 0 Apr 30 12:42:12.163193 kernel: SRAT: PXM 0 -> ITS 2 -> Node 0 Apr 30 12:42:12.163199 kernel: SRAT: PXM 0 -> ITS 3 -> Node 0 Apr 30 12:42:12.163207 kernel: SRAT: PXM 0 -> ITS 4 -> Node 0 Apr 30 12:42:12.163214 kernel: SRAT: PXM 0 -> ITS 5 -> Node 0 Apr 30 12:42:12.163221 kernel: SRAT: PXM 0 -> ITS 6 -> Node 0 Apr 30 12:42:12.163228 kernel: SRAT: PXM 0 -> ITS 7 -> Node 0 Apr 30 12:42:12.163234 kernel: ITS [mem 0x100100040000-0x10010005ffff] Apr 30 12:42:12.163241 kernel: ITS@0x0000100100040000: allocated 8192 Devices @80000270000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163248 kernel: ITS@0x0000100100040000: allocated 32768 Interrupt Collections @80000280000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163255 kernel: ITS [mem 0x100100060000-0x10010007ffff] Apr 30 12:42:12.163262 kernel: ITS@0x0000100100060000: allocated 8192 Devices @800002a0000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163269 kernel: ITS@0x0000100100060000: allocated 32768 Interrupt Collections @800002b0000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163276 kernel: ITS [mem 0x100100080000-0x10010009ffff] Apr 30 12:42:12.163284 kernel: ITS@0x0000100100080000: allocated 8192 Devices @800002d0000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163292 kernel: ITS@0x0000100100080000: allocated 32768 Interrupt Collections @800002e0000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163298 kernel: ITS [mem 0x1001000a0000-0x1001000bffff] Apr 30 12:42:12.163305 kernel: ITS@0x00001001000a0000: allocated 8192 Devices @80000300000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163312 kernel: ITS@0x00001001000a0000: allocated 32768 Interrupt Collections @80000310000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163319 kernel: ITS [mem 0x1001000c0000-0x1001000dffff] Apr 30 12:42:12.163326 kernel: ITS@0x00001001000c0000: allocated 8192 Devices @80000330000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163333 kernel: ITS@0x00001001000c0000: allocated 32768 Interrupt Collections @80000340000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163340 kernel: ITS [mem 0x1001000e0000-0x1001000fffff] Apr 30 12:42:12.163347 kernel: ITS@0x00001001000e0000: allocated 8192 Devices @80000360000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163354 kernel: ITS@0x00001001000e0000: allocated 32768 Interrupt Collections @80000370000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163362 kernel: ITS [mem 0x100100100000-0x10010011ffff] Apr 30 12:42:12.163369 kernel: ITS@0x0000100100100000: allocated 8192 Devices @80000390000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163376 kernel: ITS@0x0000100100100000: allocated 32768 Interrupt Collections @800003a0000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163383 kernel: ITS [mem 0x100100120000-0x10010013ffff] Apr 30 12:42:12.163390 kernel: ITS@0x0000100100120000: allocated 8192 Devices @800003c0000 (indirect, esz 8, psz 64K, shr 1) Apr 30 12:42:12.163397 kernel: ITS@0x0000100100120000: allocated 32768 Interrupt Collections @800003d0000 (flat, esz 2, psz 64K, shr 1) Apr 30 12:42:12.163404 kernel: GICv3: using LPI property table @0x00000800003e0000 Apr 30 12:42:12.163411 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000800003f0000 Apr 30 12:42:12.163417 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Apr 30 12:42:12.163424 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163431 kernel: ACPI GTDT: found 1 memory-mapped timer block(s). Apr 30 12:42:12.163439 kernel: arch_timer: cp15 and mmio timer(s) running at 25.00MHz (phys/phys). Apr 30 12:42:12.163447 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 30 12:42:12.163453 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 30 12:42:12.163460 kernel: Console: colour dummy device 80x25 Apr 30 12:42:12.163468 kernel: printk: console [tty0] enabled Apr 30 12:42:12.163475 kernel: ACPI: Core revision 20230628 Apr 30 12:42:12.163482 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 30 12:42:12.163489 kernel: pid_max: default: 81920 minimum: 640 Apr 30 12:42:12.163496 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Apr 30 12:42:12.163503 kernel: landlock: Up and running. Apr 30 12:42:12.163511 kernel: SELinux: Initializing. Apr 30 12:42:12.163518 kernel: Mount-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.163525 kernel: Mountpoint-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.163532 kernel: RCU Tasks: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=80. Apr 30 12:42:12.163540 kernel: RCU Tasks Trace: Setting shift to 7 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=80. Apr 30 12:42:12.163547 kernel: rcu: Hierarchical SRCU implementation. Apr 30 12:42:12.163554 kernel: rcu: Max phase no-delay instances is 400. Apr 30 12:42:12.163561 kernel: Platform MSI: ITS@0x100100040000 domain created Apr 30 12:42:12.163568 kernel: Platform MSI: ITS@0x100100060000 domain created Apr 30 12:42:12.163576 kernel: Platform MSI: ITS@0x100100080000 domain created Apr 30 12:42:12.163583 kernel: Platform MSI: ITS@0x1001000a0000 domain created Apr 30 12:42:12.163589 kernel: Platform MSI: ITS@0x1001000c0000 domain created Apr 30 12:42:12.163596 kernel: Platform MSI: ITS@0x1001000e0000 domain created Apr 30 12:42:12.163603 kernel: Platform MSI: ITS@0x100100100000 domain created Apr 30 12:42:12.163610 kernel: Platform MSI: ITS@0x100100120000 domain created Apr 30 12:42:12.163617 kernel: PCI/MSI: ITS@0x100100040000 domain created Apr 30 12:42:12.163624 kernel: PCI/MSI: ITS@0x100100060000 domain created Apr 30 12:42:12.163630 kernel: PCI/MSI: ITS@0x100100080000 domain created Apr 30 12:42:12.163638 kernel: PCI/MSI: ITS@0x1001000a0000 domain created Apr 30 12:42:12.163645 kernel: PCI/MSI: ITS@0x1001000c0000 domain created Apr 30 12:42:12.163652 kernel: PCI/MSI: ITS@0x1001000e0000 domain created Apr 30 12:42:12.163659 kernel: PCI/MSI: ITS@0x100100100000 domain created Apr 30 12:42:12.163666 kernel: PCI/MSI: ITS@0x100100120000 domain created Apr 30 12:42:12.163673 kernel: Remapping and enabling EFI services. Apr 30 12:42:12.163680 kernel: smp: Bringing up secondary CPUs ... Apr 30 12:42:12.163686 kernel: Detected PIPT I-cache on CPU1 Apr 30 12:42:12.163693 kernel: GICv3: CPU1: found redistributor 1a0000 region 0:0x00001001007c0000 Apr 30 12:42:12.163700 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000080000800000 Apr 30 12:42:12.163709 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163716 kernel: CPU1: Booted secondary processor 0x00001a0000 [0x413fd0c1] Apr 30 12:42:12.163723 kernel: Detected PIPT I-cache on CPU2 Apr 30 12:42:12.163730 kernel: GICv3: CPU2: found redistributor 140000 region 0:0x0000100100640000 Apr 30 12:42:12.163737 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000080000810000 Apr 30 12:42:12.163744 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163751 kernel: CPU2: Booted secondary processor 0x0000140000 [0x413fd0c1] Apr 30 12:42:12.163758 kernel: Detected PIPT I-cache on CPU3 Apr 30 12:42:12.163765 kernel: GICv3: CPU3: found redistributor 1c0000 region 0:0x0000100100840000 Apr 30 12:42:12.163773 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000080000820000 Apr 30 12:42:12.163780 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163787 kernel: CPU3: Booted secondary processor 0x00001c0000 [0x413fd0c1] Apr 30 12:42:12.163794 kernel: Detected PIPT I-cache on CPU4 Apr 30 12:42:12.163801 kernel: GICv3: CPU4: found redistributor 100000 region 0:0x0000100100540000 Apr 30 12:42:12.163808 kernel: GICv3: CPU4: using allocated LPI pending table @0x0000080000830000 Apr 30 12:42:12.163815 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163821 kernel: CPU4: Booted secondary processor 0x0000100000 [0x413fd0c1] Apr 30 12:42:12.163828 kernel: Detected PIPT I-cache on CPU5 Apr 30 12:42:12.163835 kernel: GICv3: CPU5: found redistributor 180000 region 0:0x0000100100740000 Apr 30 12:42:12.163844 kernel: GICv3: CPU5: using allocated LPI pending table @0x0000080000840000 Apr 30 12:42:12.163851 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163858 kernel: CPU5: Booted secondary processor 0x0000180000 [0x413fd0c1] Apr 30 12:42:12.163865 kernel: Detected PIPT I-cache on CPU6 Apr 30 12:42:12.163872 kernel: GICv3: CPU6: found redistributor 160000 region 0:0x00001001006c0000 Apr 30 12:42:12.163879 kernel: GICv3: CPU6: using allocated LPI pending table @0x0000080000850000 Apr 30 12:42:12.163886 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163892 kernel: CPU6: Booted secondary processor 0x0000160000 [0x413fd0c1] Apr 30 12:42:12.163899 kernel: Detected PIPT I-cache on CPU7 Apr 30 12:42:12.163908 kernel: GICv3: CPU7: found redistributor 1e0000 region 0:0x00001001008c0000 Apr 30 12:42:12.163915 kernel: GICv3: CPU7: using allocated LPI pending table @0x0000080000860000 Apr 30 12:42:12.163921 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163928 kernel: CPU7: Booted secondary processor 0x00001e0000 [0x413fd0c1] Apr 30 12:42:12.163935 kernel: Detected PIPT I-cache on CPU8 Apr 30 12:42:12.163942 kernel: GICv3: CPU8: found redistributor a0000 region 0:0x00001001003c0000 Apr 30 12:42:12.163951 kernel: GICv3: CPU8: using allocated LPI pending table @0x0000080000870000 Apr 30 12:42:12.163959 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.163965 kernel: CPU8: Booted secondary processor 0x00000a0000 [0x413fd0c1] Apr 30 12:42:12.163973 kernel: Detected PIPT I-cache on CPU9 Apr 30 12:42:12.163981 kernel: GICv3: CPU9: found redistributor 220000 region 0:0x00001001009c0000 Apr 30 12:42:12.163989 kernel: GICv3: CPU9: using allocated LPI pending table @0x0000080000880000 Apr 30 12:42:12.163996 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164002 kernel: CPU9: Booted secondary processor 0x0000220000 [0x413fd0c1] Apr 30 12:42:12.164009 kernel: Detected PIPT I-cache on CPU10 Apr 30 12:42:12.164016 kernel: GICv3: CPU10: found redistributor c0000 region 0:0x0000100100440000 Apr 30 12:42:12.164023 kernel: GICv3: CPU10: using allocated LPI pending table @0x0000080000890000 Apr 30 12:42:12.164030 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164037 kernel: CPU10: Booted secondary processor 0x00000c0000 [0x413fd0c1] Apr 30 12:42:12.164046 kernel: Detected PIPT I-cache on CPU11 Apr 30 12:42:12.164053 kernel: GICv3: CPU11: found redistributor 240000 region 0:0x0000100100a40000 Apr 30 12:42:12.164060 kernel: GICv3: CPU11: using allocated LPI pending table @0x00000800008a0000 Apr 30 12:42:12.164067 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164074 kernel: CPU11: Booted secondary processor 0x0000240000 [0x413fd0c1] Apr 30 12:42:12.164081 kernel: Detected PIPT I-cache on CPU12 Apr 30 12:42:12.164088 kernel: GICv3: CPU12: found redistributor 80000 region 0:0x0000100100340000 Apr 30 12:42:12.164095 kernel: GICv3: CPU12: using allocated LPI pending table @0x00000800008b0000 Apr 30 12:42:12.164102 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164109 kernel: CPU12: Booted secondary processor 0x0000080000 [0x413fd0c1] Apr 30 12:42:12.164117 kernel: Detected PIPT I-cache on CPU13 Apr 30 12:42:12.164124 kernel: GICv3: CPU13: found redistributor 200000 region 0:0x0000100100940000 Apr 30 12:42:12.164131 kernel: GICv3: CPU13: using allocated LPI pending table @0x00000800008c0000 Apr 30 12:42:12.164138 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164145 kernel: CPU13: Booted secondary processor 0x0000200000 [0x413fd0c1] Apr 30 12:42:12.164152 kernel: Detected PIPT I-cache on CPU14 Apr 30 12:42:12.164159 kernel: GICv3: CPU14: found redistributor e0000 region 0:0x00001001004c0000 Apr 30 12:42:12.164166 kernel: GICv3: CPU14: using allocated LPI pending table @0x00000800008d0000 Apr 30 12:42:12.164173 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164182 kernel: CPU14: Booted secondary processor 0x00000e0000 [0x413fd0c1] Apr 30 12:42:12.164189 kernel: Detected PIPT I-cache on CPU15 Apr 30 12:42:12.164196 kernel: GICv3: CPU15: found redistributor 260000 region 0:0x0000100100ac0000 Apr 30 12:42:12.164203 kernel: GICv3: CPU15: using allocated LPI pending table @0x00000800008e0000 Apr 30 12:42:12.164210 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164216 kernel: CPU15: Booted secondary processor 0x0000260000 [0x413fd0c1] Apr 30 12:42:12.164223 kernel: Detected PIPT I-cache on CPU16 Apr 30 12:42:12.164230 kernel: GICv3: CPU16: found redistributor 20000 region 0:0x00001001001c0000 Apr 30 12:42:12.164237 kernel: GICv3: CPU16: using allocated LPI pending table @0x00000800008f0000 Apr 30 12:42:12.164254 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164263 kernel: CPU16: Booted secondary processor 0x0000020000 [0x413fd0c1] Apr 30 12:42:12.164270 kernel: Detected PIPT I-cache on CPU17 Apr 30 12:42:12.164277 kernel: GICv3: CPU17: found redistributor 40000 region 0:0x0000100100240000 Apr 30 12:42:12.164285 kernel: GICv3: CPU17: using allocated LPI pending table @0x0000080000900000 Apr 30 12:42:12.164292 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164299 kernel: CPU17: Booted secondary processor 0x0000040000 [0x413fd0c1] Apr 30 12:42:12.164306 kernel: Detected PIPT I-cache on CPU18 Apr 30 12:42:12.164314 kernel: GICv3: CPU18: found redistributor 0 region 0:0x0000100100140000 Apr 30 12:42:12.164321 kernel: GICv3: CPU18: using allocated LPI pending table @0x0000080000910000 Apr 30 12:42:12.164330 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164337 kernel: CPU18: Booted secondary processor 0x0000000000 [0x413fd0c1] Apr 30 12:42:12.164344 kernel: Detected PIPT I-cache on CPU19 Apr 30 12:42:12.164351 kernel: GICv3: CPU19: found redistributor 60000 region 0:0x00001001002c0000 Apr 30 12:42:12.164359 kernel: GICv3: CPU19: using allocated LPI pending table @0x0000080000920000 Apr 30 12:42:12.164366 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164375 kernel: CPU19: Booted secondary processor 0x0000060000 [0x413fd0c1] Apr 30 12:42:12.164382 kernel: Detected PIPT I-cache on CPU20 Apr 30 12:42:12.164389 kernel: GICv3: CPU20: found redistributor 130000 region 0:0x0000100100600000 Apr 30 12:42:12.164397 kernel: GICv3: CPU20: using allocated LPI pending table @0x0000080000930000 Apr 30 12:42:12.164404 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164411 kernel: CPU20: Booted secondary processor 0x0000130000 [0x413fd0c1] Apr 30 12:42:12.164419 kernel: Detected PIPT I-cache on CPU21 Apr 30 12:42:12.164426 kernel: GICv3: CPU21: found redistributor 1b0000 region 0:0x0000100100800000 Apr 30 12:42:12.164433 kernel: GICv3: CPU21: using allocated LPI pending table @0x0000080000940000 Apr 30 12:42:12.164442 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164449 kernel: CPU21: Booted secondary processor 0x00001b0000 [0x413fd0c1] Apr 30 12:42:12.164456 kernel: Detected PIPT I-cache on CPU22 Apr 30 12:42:12.164464 kernel: GICv3: CPU22: found redistributor 150000 region 0:0x0000100100680000 Apr 30 12:42:12.164471 kernel: GICv3: CPU22: using allocated LPI pending table @0x0000080000950000 Apr 30 12:42:12.164478 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164486 kernel: CPU22: Booted secondary processor 0x0000150000 [0x413fd0c1] Apr 30 12:42:12.164493 kernel: Detected PIPT I-cache on CPU23 Apr 30 12:42:12.164500 kernel: GICv3: CPU23: found redistributor 1d0000 region 0:0x0000100100880000 Apr 30 12:42:12.164509 kernel: GICv3: CPU23: using allocated LPI pending table @0x0000080000960000 Apr 30 12:42:12.164516 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164523 kernel: CPU23: Booted secondary processor 0x00001d0000 [0x413fd0c1] Apr 30 12:42:12.164531 kernel: Detected PIPT I-cache on CPU24 Apr 30 12:42:12.164538 kernel: GICv3: CPU24: found redistributor 110000 region 0:0x0000100100580000 Apr 30 12:42:12.164545 kernel: GICv3: CPU24: using allocated LPI pending table @0x0000080000970000 Apr 30 12:42:12.164552 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164559 kernel: CPU24: Booted secondary processor 0x0000110000 [0x413fd0c1] Apr 30 12:42:12.164566 kernel: Detected PIPT I-cache on CPU25 Apr 30 12:42:12.164574 kernel: GICv3: CPU25: found redistributor 190000 region 0:0x0000100100780000 Apr 30 12:42:12.164584 kernel: GICv3: CPU25: using allocated LPI pending table @0x0000080000980000 Apr 30 12:42:12.164593 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164600 kernel: CPU25: Booted secondary processor 0x0000190000 [0x413fd0c1] Apr 30 12:42:12.164607 kernel: Detected PIPT I-cache on CPU26 Apr 30 12:42:12.164615 kernel: GICv3: CPU26: found redistributor 170000 region 0:0x0000100100700000 Apr 30 12:42:12.164622 kernel: GICv3: CPU26: using allocated LPI pending table @0x0000080000990000 Apr 30 12:42:12.164629 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164636 kernel: CPU26: Booted secondary processor 0x0000170000 [0x413fd0c1] Apr 30 12:42:12.164644 kernel: Detected PIPT I-cache on CPU27 Apr 30 12:42:12.164652 kernel: GICv3: CPU27: found redistributor 1f0000 region 0:0x0000100100900000 Apr 30 12:42:12.164660 kernel: GICv3: CPU27: using allocated LPI pending table @0x00000800009a0000 Apr 30 12:42:12.164667 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164674 kernel: CPU27: Booted secondary processor 0x00001f0000 [0x413fd0c1] Apr 30 12:42:12.164681 kernel: Detected PIPT I-cache on CPU28 Apr 30 12:42:12.164688 kernel: GICv3: CPU28: found redistributor b0000 region 0:0x0000100100400000 Apr 30 12:42:12.164696 kernel: GICv3: CPU28: using allocated LPI pending table @0x00000800009b0000 Apr 30 12:42:12.164703 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164710 kernel: CPU28: Booted secondary processor 0x00000b0000 [0x413fd0c1] Apr 30 12:42:12.164717 kernel: Detected PIPT I-cache on CPU29 Apr 30 12:42:12.164726 kernel: GICv3: CPU29: found redistributor 230000 region 0:0x0000100100a00000 Apr 30 12:42:12.164733 kernel: GICv3: CPU29: using allocated LPI pending table @0x00000800009c0000 Apr 30 12:42:12.164741 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164748 kernel: CPU29: Booted secondary processor 0x0000230000 [0x413fd0c1] Apr 30 12:42:12.164755 kernel: Detected PIPT I-cache on CPU30 Apr 30 12:42:12.164763 kernel: GICv3: CPU30: found redistributor d0000 region 0:0x0000100100480000 Apr 30 12:42:12.164770 kernel: GICv3: CPU30: using allocated LPI pending table @0x00000800009d0000 Apr 30 12:42:12.164777 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164784 kernel: CPU30: Booted secondary processor 0x00000d0000 [0x413fd0c1] Apr 30 12:42:12.164793 kernel: Detected PIPT I-cache on CPU31 Apr 30 12:42:12.164801 kernel: GICv3: CPU31: found redistributor 250000 region 0:0x0000100100a80000 Apr 30 12:42:12.164808 kernel: GICv3: CPU31: using allocated LPI pending table @0x00000800009e0000 Apr 30 12:42:12.164815 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164822 kernel: CPU31: Booted secondary processor 0x0000250000 [0x413fd0c1] Apr 30 12:42:12.164829 kernel: Detected PIPT I-cache on CPU32 Apr 30 12:42:12.164837 kernel: GICv3: CPU32: found redistributor 90000 region 0:0x0000100100380000 Apr 30 12:42:12.164844 kernel: GICv3: CPU32: using allocated LPI pending table @0x00000800009f0000 Apr 30 12:42:12.164851 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164858 kernel: CPU32: Booted secondary processor 0x0000090000 [0x413fd0c1] Apr 30 12:42:12.164867 kernel: Detected PIPT I-cache on CPU33 Apr 30 12:42:12.164875 kernel: GICv3: CPU33: found redistributor 210000 region 0:0x0000100100980000 Apr 30 12:42:12.164882 kernel: GICv3: CPU33: using allocated LPI pending table @0x0000080000a00000 Apr 30 12:42:12.164889 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164896 kernel: CPU33: Booted secondary processor 0x0000210000 [0x413fd0c1] Apr 30 12:42:12.164904 kernel: Detected PIPT I-cache on CPU34 Apr 30 12:42:12.164911 kernel: GICv3: CPU34: found redistributor f0000 region 0:0x0000100100500000 Apr 30 12:42:12.164918 kernel: GICv3: CPU34: using allocated LPI pending table @0x0000080000a10000 Apr 30 12:42:12.164926 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164934 kernel: CPU34: Booted secondary processor 0x00000f0000 [0x413fd0c1] Apr 30 12:42:12.164941 kernel: Detected PIPT I-cache on CPU35 Apr 30 12:42:12.164951 kernel: GICv3: CPU35: found redistributor 270000 region 0:0x0000100100b00000 Apr 30 12:42:12.164959 kernel: GICv3: CPU35: using allocated LPI pending table @0x0000080000a20000 Apr 30 12:42:12.164966 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.164974 kernel: CPU35: Booted secondary processor 0x0000270000 [0x413fd0c1] Apr 30 12:42:12.164981 kernel: Detected PIPT I-cache on CPU36 Apr 30 12:42:12.164988 kernel: GICv3: CPU36: found redistributor 30000 region 0:0x0000100100200000 Apr 30 12:42:12.164996 kernel: GICv3: CPU36: using allocated LPI pending table @0x0000080000a30000 Apr 30 12:42:12.165003 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165012 kernel: CPU36: Booted secondary processor 0x0000030000 [0x413fd0c1] Apr 30 12:42:12.165019 kernel: Detected PIPT I-cache on CPU37 Apr 30 12:42:12.165026 kernel: GICv3: CPU37: found redistributor 50000 region 0:0x0000100100280000 Apr 30 12:42:12.165034 kernel: GICv3: CPU37: using allocated LPI pending table @0x0000080000a40000 Apr 30 12:42:12.165041 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165049 kernel: CPU37: Booted secondary processor 0x0000050000 [0x413fd0c1] Apr 30 12:42:12.165056 kernel: Detected PIPT I-cache on CPU38 Apr 30 12:42:12.165063 kernel: GICv3: CPU38: found redistributor 10000 region 0:0x0000100100180000 Apr 30 12:42:12.165071 kernel: GICv3: CPU38: using allocated LPI pending table @0x0000080000a50000 Apr 30 12:42:12.165080 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165087 kernel: CPU38: Booted secondary processor 0x0000010000 [0x413fd0c1] Apr 30 12:42:12.165095 kernel: Detected PIPT I-cache on CPU39 Apr 30 12:42:12.165103 kernel: GICv3: CPU39: found redistributor 70000 region 0:0x0000100100300000 Apr 30 12:42:12.165110 kernel: GICv3: CPU39: using allocated LPI pending table @0x0000080000a60000 Apr 30 12:42:12.165117 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165125 kernel: CPU39: Booted secondary processor 0x0000070000 [0x413fd0c1] Apr 30 12:42:12.165132 kernel: Detected PIPT I-cache on CPU40 Apr 30 12:42:12.165141 kernel: GICv3: CPU40: found redistributor 120100 region 0:0x00001001005e0000 Apr 30 12:42:12.165148 kernel: GICv3: CPU40: using allocated LPI pending table @0x0000080000a70000 Apr 30 12:42:12.165156 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165163 kernel: CPU40: Booted secondary processor 0x0000120100 [0x413fd0c1] Apr 30 12:42:12.165170 kernel: Detected PIPT I-cache on CPU41 Apr 30 12:42:12.165177 kernel: GICv3: CPU41: found redistributor 1a0100 region 0:0x00001001007e0000 Apr 30 12:42:12.165185 kernel: GICv3: CPU41: using allocated LPI pending table @0x0000080000a80000 Apr 30 12:42:12.165192 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165199 kernel: CPU41: Booted secondary processor 0x00001a0100 [0x413fd0c1] Apr 30 12:42:12.165206 kernel: Detected PIPT I-cache on CPU42 Apr 30 12:42:12.165215 kernel: GICv3: CPU42: found redistributor 140100 region 0:0x0000100100660000 Apr 30 12:42:12.165222 kernel: GICv3: CPU42: using allocated LPI pending table @0x0000080000a90000 Apr 30 12:42:12.165230 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165237 kernel: CPU42: Booted secondary processor 0x0000140100 [0x413fd0c1] Apr 30 12:42:12.165244 kernel: Detected PIPT I-cache on CPU43 Apr 30 12:42:12.165251 kernel: GICv3: CPU43: found redistributor 1c0100 region 0:0x0000100100860000 Apr 30 12:42:12.165259 kernel: GICv3: CPU43: using allocated LPI pending table @0x0000080000aa0000 Apr 30 12:42:12.165266 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165273 kernel: CPU43: Booted secondary processor 0x00001c0100 [0x413fd0c1] Apr 30 12:42:12.165282 kernel: Detected PIPT I-cache on CPU44 Apr 30 12:42:12.165290 kernel: GICv3: CPU44: found redistributor 100100 region 0:0x0000100100560000 Apr 30 12:42:12.165297 kernel: GICv3: CPU44: using allocated LPI pending table @0x0000080000ab0000 Apr 30 12:42:12.165304 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165311 kernel: CPU44: Booted secondary processor 0x0000100100 [0x413fd0c1] Apr 30 12:42:12.165318 kernel: Detected PIPT I-cache on CPU45 Apr 30 12:42:12.165326 kernel: GICv3: CPU45: found redistributor 180100 region 0:0x0000100100760000 Apr 30 12:42:12.165333 kernel: GICv3: CPU45: using allocated LPI pending table @0x0000080000ac0000 Apr 30 12:42:12.165341 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165348 kernel: CPU45: Booted secondary processor 0x0000180100 [0x413fd0c1] Apr 30 12:42:12.165357 kernel: Detected PIPT I-cache on CPU46 Apr 30 12:42:12.165364 kernel: GICv3: CPU46: found redistributor 160100 region 0:0x00001001006e0000 Apr 30 12:42:12.165371 kernel: GICv3: CPU46: using allocated LPI pending table @0x0000080000ad0000 Apr 30 12:42:12.165379 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165386 kernel: CPU46: Booted secondary processor 0x0000160100 [0x413fd0c1] Apr 30 12:42:12.165393 kernel: Detected PIPT I-cache on CPU47 Apr 30 12:42:12.165400 kernel: GICv3: CPU47: found redistributor 1e0100 region 0:0x00001001008e0000 Apr 30 12:42:12.165408 kernel: GICv3: CPU47: using allocated LPI pending table @0x0000080000ae0000 Apr 30 12:42:12.165415 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165424 kernel: CPU47: Booted secondary processor 0x00001e0100 [0x413fd0c1] Apr 30 12:42:12.165431 kernel: Detected PIPT I-cache on CPU48 Apr 30 12:42:12.165438 kernel: GICv3: CPU48: found redistributor a0100 region 0:0x00001001003e0000 Apr 30 12:42:12.165446 kernel: GICv3: CPU48: using allocated LPI pending table @0x0000080000af0000 Apr 30 12:42:12.165453 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165460 kernel: CPU48: Booted secondary processor 0x00000a0100 [0x413fd0c1] Apr 30 12:42:12.165467 kernel: Detected PIPT I-cache on CPU49 Apr 30 12:42:12.165475 kernel: GICv3: CPU49: found redistributor 220100 region 0:0x00001001009e0000 Apr 30 12:42:12.165482 kernel: GICv3: CPU49: using allocated LPI pending table @0x0000080000b00000 Apr 30 12:42:12.165491 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165498 kernel: CPU49: Booted secondary processor 0x0000220100 [0x413fd0c1] Apr 30 12:42:12.165505 kernel: Detected PIPT I-cache on CPU50 Apr 30 12:42:12.165512 kernel: GICv3: CPU50: found redistributor c0100 region 0:0x0000100100460000 Apr 30 12:42:12.165520 kernel: GICv3: CPU50: using allocated LPI pending table @0x0000080000b10000 Apr 30 12:42:12.165527 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165536 kernel: CPU50: Booted secondary processor 0x00000c0100 [0x413fd0c1] Apr 30 12:42:12.165543 kernel: Detected PIPT I-cache on CPU51 Apr 30 12:42:12.165550 kernel: GICv3: CPU51: found redistributor 240100 region 0:0x0000100100a60000 Apr 30 12:42:12.165557 kernel: GICv3: CPU51: using allocated LPI pending table @0x0000080000b20000 Apr 30 12:42:12.165566 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165574 kernel: CPU51: Booted secondary processor 0x0000240100 [0x413fd0c1] Apr 30 12:42:12.165581 kernel: Detected PIPT I-cache on CPU52 Apr 30 12:42:12.165588 kernel: GICv3: CPU52: found redistributor 80100 region 0:0x0000100100360000 Apr 30 12:42:12.165595 kernel: GICv3: CPU52: using allocated LPI pending table @0x0000080000b30000 Apr 30 12:42:12.165603 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165610 kernel: CPU52: Booted secondary processor 0x0000080100 [0x413fd0c1] Apr 30 12:42:12.165617 kernel: Detected PIPT I-cache on CPU53 Apr 30 12:42:12.165625 kernel: GICv3: CPU53: found redistributor 200100 region 0:0x0000100100960000 Apr 30 12:42:12.165633 kernel: GICv3: CPU53: using allocated LPI pending table @0x0000080000b40000 Apr 30 12:42:12.165641 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165648 kernel: CPU53: Booted secondary processor 0x0000200100 [0x413fd0c1] Apr 30 12:42:12.165655 kernel: Detected PIPT I-cache on CPU54 Apr 30 12:42:12.165662 kernel: GICv3: CPU54: found redistributor e0100 region 0:0x00001001004e0000 Apr 30 12:42:12.165670 kernel: GICv3: CPU54: using allocated LPI pending table @0x0000080000b50000 Apr 30 12:42:12.165677 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165685 kernel: CPU54: Booted secondary processor 0x00000e0100 [0x413fd0c1] Apr 30 12:42:12.165692 kernel: Detected PIPT I-cache on CPU55 Apr 30 12:42:12.165699 kernel: GICv3: CPU55: found redistributor 260100 region 0:0x0000100100ae0000 Apr 30 12:42:12.165708 kernel: GICv3: CPU55: using allocated LPI pending table @0x0000080000b60000 Apr 30 12:42:12.165715 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165722 kernel: CPU55: Booted secondary processor 0x0000260100 [0x413fd0c1] Apr 30 12:42:12.165729 kernel: Detected PIPT I-cache on CPU56 Apr 30 12:42:12.165737 kernel: GICv3: CPU56: found redistributor 20100 region 0:0x00001001001e0000 Apr 30 12:42:12.165744 kernel: GICv3: CPU56: using allocated LPI pending table @0x0000080000b70000 Apr 30 12:42:12.165752 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165759 kernel: CPU56: Booted secondary processor 0x0000020100 [0x413fd0c1] Apr 30 12:42:12.165766 kernel: Detected PIPT I-cache on CPU57 Apr 30 12:42:12.165775 kernel: GICv3: CPU57: found redistributor 40100 region 0:0x0000100100260000 Apr 30 12:42:12.165782 kernel: GICv3: CPU57: using allocated LPI pending table @0x0000080000b80000 Apr 30 12:42:12.165791 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165798 kernel: CPU57: Booted secondary processor 0x0000040100 [0x413fd0c1] Apr 30 12:42:12.165805 kernel: Detected PIPT I-cache on CPU58 Apr 30 12:42:12.165813 kernel: GICv3: CPU58: found redistributor 100 region 0:0x0000100100160000 Apr 30 12:42:12.165820 kernel: GICv3: CPU58: using allocated LPI pending table @0x0000080000b90000 Apr 30 12:42:12.165828 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165835 kernel: CPU58: Booted secondary processor 0x0000000100 [0x413fd0c1] Apr 30 12:42:12.165842 kernel: Detected PIPT I-cache on CPU59 Apr 30 12:42:12.165851 kernel: GICv3: CPU59: found redistributor 60100 region 0:0x00001001002e0000 Apr 30 12:42:12.165859 kernel: GICv3: CPU59: using allocated LPI pending table @0x0000080000ba0000 Apr 30 12:42:12.165866 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165873 kernel: CPU59: Booted secondary processor 0x0000060100 [0x413fd0c1] Apr 30 12:42:12.165880 kernel: Detected PIPT I-cache on CPU60 Apr 30 12:42:12.165888 kernel: GICv3: CPU60: found redistributor 130100 region 0:0x0000100100620000 Apr 30 12:42:12.165895 kernel: GICv3: CPU60: using allocated LPI pending table @0x0000080000bb0000 Apr 30 12:42:12.165903 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165910 kernel: CPU60: Booted secondary processor 0x0000130100 [0x413fd0c1] Apr 30 12:42:12.165919 kernel: Detected PIPT I-cache on CPU61 Apr 30 12:42:12.165926 kernel: GICv3: CPU61: found redistributor 1b0100 region 0:0x0000100100820000 Apr 30 12:42:12.165933 kernel: GICv3: CPU61: using allocated LPI pending table @0x0000080000bc0000 Apr 30 12:42:12.165941 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165948 kernel: CPU61: Booted secondary processor 0x00001b0100 [0x413fd0c1] Apr 30 12:42:12.165957 kernel: Detected PIPT I-cache on CPU62 Apr 30 12:42:12.165965 kernel: GICv3: CPU62: found redistributor 150100 region 0:0x00001001006a0000 Apr 30 12:42:12.165972 kernel: GICv3: CPU62: using allocated LPI pending table @0x0000080000bd0000 Apr 30 12:42:12.165979 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.165987 kernel: CPU62: Booted secondary processor 0x0000150100 [0x413fd0c1] Apr 30 12:42:12.165996 kernel: Detected PIPT I-cache on CPU63 Apr 30 12:42:12.166003 kernel: GICv3: CPU63: found redistributor 1d0100 region 0:0x00001001008a0000 Apr 30 12:42:12.166010 kernel: GICv3: CPU63: using allocated LPI pending table @0x0000080000be0000 Apr 30 12:42:12.166018 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166025 kernel: CPU63: Booted secondary processor 0x00001d0100 [0x413fd0c1] Apr 30 12:42:12.166032 kernel: Detected PIPT I-cache on CPU64 Apr 30 12:42:12.166039 kernel: GICv3: CPU64: found redistributor 110100 region 0:0x00001001005a0000 Apr 30 12:42:12.166047 kernel: GICv3: CPU64: using allocated LPI pending table @0x0000080000bf0000 Apr 30 12:42:12.166054 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166063 kernel: CPU64: Booted secondary processor 0x0000110100 [0x413fd0c1] Apr 30 12:42:12.166070 kernel: Detected PIPT I-cache on CPU65 Apr 30 12:42:12.166077 kernel: GICv3: CPU65: found redistributor 190100 region 0:0x00001001007a0000 Apr 30 12:42:12.166085 kernel: GICv3: CPU65: using allocated LPI pending table @0x0000080000c00000 Apr 30 12:42:12.166092 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166099 kernel: CPU65: Booted secondary processor 0x0000190100 [0x413fd0c1] Apr 30 12:42:12.166107 kernel: Detected PIPT I-cache on CPU66 Apr 30 12:42:12.166114 kernel: GICv3: CPU66: found redistributor 170100 region 0:0x0000100100720000 Apr 30 12:42:12.166121 kernel: GICv3: CPU66: using allocated LPI pending table @0x0000080000c10000 Apr 30 12:42:12.166130 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166137 kernel: CPU66: Booted secondary processor 0x0000170100 [0x413fd0c1] Apr 30 12:42:12.166144 kernel: Detected PIPT I-cache on CPU67 Apr 30 12:42:12.166151 kernel: GICv3: CPU67: found redistributor 1f0100 region 0:0x0000100100920000 Apr 30 12:42:12.166159 kernel: GICv3: CPU67: using allocated LPI pending table @0x0000080000c20000 Apr 30 12:42:12.166166 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166173 kernel: CPU67: Booted secondary processor 0x00001f0100 [0x413fd0c1] Apr 30 12:42:12.166180 kernel: Detected PIPT I-cache on CPU68 Apr 30 12:42:12.166187 kernel: GICv3: CPU68: found redistributor b0100 region 0:0x0000100100420000 Apr 30 12:42:12.166195 kernel: GICv3: CPU68: using allocated LPI pending table @0x0000080000c30000 Apr 30 12:42:12.166203 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166211 kernel: CPU68: Booted secondary processor 0x00000b0100 [0x413fd0c1] Apr 30 12:42:12.166218 kernel: Detected PIPT I-cache on CPU69 Apr 30 12:42:12.166225 kernel: GICv3: CPU69: found redistributor 230100 region 0:0x0000100100a20000 Apr 30 12:42:12.166232 kernel: GICv3: CPU69: using allocated LPI pending table @0x0000080000c40000 Apr 30 12:42:12.166240 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166247 kernel: CPU69: Booted secondary processor 0x0000230100 [0x413fd0c1] Apr 30 12:42:12.166254 kernel: Detected PIPT I-cache on CPU70 Apr 30 12:42:12.166261 kernel: GICv3: CPU70: found redistributor d0100 region 0:0x00001001004a0000 Apr 30 12:42:12.166270 kernel: GICv3: CPU70: using allocated LPI pending table @0x0000080000c50000 Apr 30 12:42:12.166278 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166285 kernel: CPU70: Booted secondary processor 0x00000d0100 [0x413fd0c1] Apr 30 12:42:12.166292 kernel: Detected PIPT I-cache on CPU71 Apr 30 12:42:12.166299 kernel: GICv3: CPU71: found redistributor 250100 region 0:0x0000100100aa0000 Apr 30 12:42:12.166307 kernel: GICv3: CPU71: using allocated LPI pending table @0x0000080000c60000 Apr 30 12:42:12.166314 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166321 kernel: CPU71: Booted secondary processor 0x0000250100 [0x413fd0c1] Apr 30 12:42:12.166328 kernel: Detected PIPT I-cache on CPU72 Apr 30 12:42:12.166336 kernel: GICv3: CPU72: found redistributor 90100 region 0:0x00001001003a0000 Apr 30 12:42:12.166344 kernel: GICv3: CPU72: using allocated LPI pending table @0x0000080000c70000 Apr 30 12:42:12.166352 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166359 kernel: CPU72: Booted secondary processor 0x0000090100 [0x413fd0c1] Apr 30 12:42:12.166367 kernel: Detected PIPT I-cache on CPU73 Apr 30 12:42:12.166374 kernel: GICv3: CPU73: found redistributor 210100 region 0:0x00001001009a0000 Apr 30 12:42:12.166381 kernel: GICv3: CPU73: using allocated LPI pending table @0x0000080000c80000 Apr 30 12:42:12.166389 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166396 kernel: CPU73: Booted secondary processor 0x0000210100 [0x413fd0c1] Apr 30 12:42:12.166403 kernel: Detected PIPT I-cache on CPU74 Apr 30 12:42:12.166412 kernel: GICv3: CPU74: found redistributor f0100 region 0:0x0000100100520000 Apr 30 12:42:12.166419 kernel: GICv3: CPU74: using allocated LPI pending table @0x0000080000c90000 Apr 30 12:42:12.166426 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166434 kernel: CPU74: Booted secondary processor 0x00000f0100 [0x413fd0c1] Apr 30 12:42:12.166441 kernel: Detected PIPT I-cache on CPU75 Apr 30 12:42:12.166448 kernel: GICv3: CPU75: found redistributor 270100 region 0:0x0000100100b20000 Apr 30 12:42:12.166456 kernel: GICv3: CPU75: using allocated LPI pending table @0x0000080000ca0000 Apr 30 12:42:12.166463 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166470 kernel: CPU75: Booted secondary processor 0x0000270100 [0x413fd0c1] Apr 30 12:42:12.166478 kernel: Detected PIPT I-cache on CPU76 Apr 30 12:42:12.166486 kernel: GICv3: CPU76: found redistributor 30100 region 0:0x0000100100220000 Apr 30 12:42:12.166494 kernel: GICv3: CPU76: using allocated LPI pending table @0x0000080000cb0000 Apr 30 12:42:12.166501 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166508 kernel: CPU76: Booted secondary processor 0x0000030100 [0x413fd0c1] Apr 30 12:42:12.166515 kernel: Detected PIPT I-cache on CPU77 Apr 30 12:42:12.166523 kernel: GICv3: CPU77: found redistributor 50100 region 0:0x00001001002a0000 Apr 30 12:42:12.166530 kernel: GICv3: CPU77: using allocated LPI pending table @0x0000080000cc0000 Apr 30 12:42:12.166537 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166544 kernel: CPU77: Booted secondary processor 0x0000050100 [0x413fd0c1] Apr 30 12:42:12.166553 kernel: Detected PIPT I-cache on CPU78 Apr 30 12:42:12.166560 kernel: GICv3: CPU78: found redistributor 10100 region 0:0x00001001001a0000 Apr 30 12:42:12.166568 kernel: GICv3: CPU78: using allocated LPI pending table @0x0000080000cd0000 Apr 30 12:42:12.166575 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166582 kernel: CPU78: Booted secondary processor 0x0000010100 [0x413fd0c1] Apr 30 12:42:12.166590 kernel: Detected PIPT I-cache on CPU79 Apr 30 12:42:12.166597 kernel: GICv3: CPU79: found redistributor 70100 region 0:0x0000100100320000 Apr 30 12:42:12.166604 kernel: GICv3: CPU79: using allocated LPI pending table @0x0000080000ce0000 Apr 30 12:42:12.166611 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 30 12:42:12.166618 kernel: CPU79: Booted secondary processor 0x0000070100 [0x413fd0c1] Apr 30 12:42:12.166627 kernel: smp: Brought up 1 node, 80 CPUs Apr 30 12:42:12.166634 kernel: SMP: Total of 80 processors activated. Apr 30 12:42:12.166642 kernel: CPU features: detected: 32-bit EL0 Support Apr 30 12:42:12.166649 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 30 12:42:12.166657 kernel: CPU features: detected: Common not Private translations Apr 30 12:42:12.166664 kernel: CPU features: detected: CRC32 instructions Apr 30 12:42:12.166671 kernel: CPU features: detected: Enhanced Virtualization Traps Apr 30 12:42:12.166678 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 30 12:42:12.166686 kernel: CPU features: detected: LSE atomic instructions Apr 30 12:42:12.166695 kernel: CPU features: detected: Privileged Access Never Apr 30 12:42:12.166702 kernel: CPU features: detected: RAS Extension Support Apr 30 12:42:12.166709 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Apr 30 12:42:12.166717 kernel: CPU: All CPU(s) started at EL2 Apr 30 12:42:12.166724 kernel: alternatives: applying system-wide alternatives Apr 30 12:42:12.166731 kernel: devtmpfs: initialized Apr 30 12:42:12.166738 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 30 12:42:12.166746 kernel: futex hash table entries: 32768 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.166753 kernel: pinctrl core: initialized pinctrl subsystem Apr 30 12:42:12.166762 kernel: SMBIOS 3.4.0 present. Apr 30 12:42:12.166769 kernel: DMI: GIGABYTE R272-P30-JG/MP32-AR0-JG, BIOS F17a (SCP: 1.07.20210713) 07/22/2021 Apr 30 12:42:12.166777 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 30 12:42:12.166784 kernel: DMA: preallocated 4096 KiB GFP_KERNEL pool for atomic allocations Apr 30 12:42:12.166792 kernel: DMA: preallocated 4096 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 30 12:42:12.166799 kernel: DMA: preallocated 4096 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 30 12:42:12.166807 kernel: audit: initializing netlink subsys (disabled) Apr 30 12:42:12.166814 kernel: audit: type=2000 audit(0.042:1): state=initialized audit_enabled=0 res=1 Apr 30 12:42:12.166823 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 30 12:42:12.166830 kernel: cpuidle: using governor menu Apr 30 12:42:12.166837 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 30 12:42:12.166845 kernel: ASID allocator initialised with 32768 entries Apr 30 12:42:12.166852 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 30 12:42:12.166859 kernel: Serial: AMBA PL011 UART driver Apr 30 12:42:12.166867 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Apr 30 12:42:12.166874 kernel: Modules: 0 pages in range for non-PLT usage Apr 30 12:42:12.166881 kernel: Modules: 509264 pages in range for PLT usage Apr 30 12:42:12.166888 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Apr 30 12:42:12.166897 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Apr 30 12:42:12.166905 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Apr 30 12:42:12.166912 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Apr 30 12:42:12.166919 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Apr 30 12:42:12.166926 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Apr 30 12:42:12.166934 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Apr 30 12:42:12.166941 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Apr 30 12:42:12.166950 kernel: ACPI: Added _OSI(Module Device) Apr 30 12:42:12.166958 kernel: ACPI: Added _OSI(Processor Device) Apr 30 12:42:12.166967 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Apr 30 12:42:12.166974 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 30 12:42:12.166981 kernel: ACPI: 2 ACPI AML tables successfully acquired and loaded Apr 30 12:42:12.166988 kernel: ACPI: Interpreter enabled Apr 30 12:42:12.166996 kernel: ACPI: Using GIC for interrupt routing Apr 30 12:42:12.167003 kernel: ACPI: MCFG table detected, 8 entries Apr 30 12:42:12.167010 kernel: ACPI: IORT: SMMU-v3[33ffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167018 kernel: ACPI: IORT: SMMU-v3[37ffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167025 kernel: ACPI: IORT: SMMU-v3[3bffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167034 kernel: ACPI: IORT: SMMU-v3[3fffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167041 kernel: ACPI: IORT: SMMU-v3[23ffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167048 kernel: ACPI: IORT: SMMU-v3[27ffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167056 kernel: ACPI: IORT: SMMU-v3[2bffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167063 kernel: ACPI: IORT: SMMU-v3[2fffe0000000] Mapped to Proximity domain 0 Apr 30 12:42:12.167070 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x100002600000 (irq = 19, base_baud = 0) is a SBSA Apr 30 12:42:12.167078 kernel: printk: console [ttyAMA0] enabled Apr 30 12:42:12.167085 kernel: ARMH0011:01: ttyAMA1 at MMIO 0x100002620000 (irq = 20, base_baud = 0) is a SBSA Apr 30 12:42:12.167094 kernel: ACPI: PCI Root Bridge [PCI1] (domain 000d [bus 00-ff]) Apr 30 12:42:12.167229 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.167301 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.167364 kernel: acpi PNP0A08:00: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.167425 kernel: acpi PNP0A08:00: MCFG quirk: ECAM at [mem 0x37fff0000000-0x37ffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.167489 kernel: acpi PNP0A08:00: ECAM area [mem 0x37fff0000000-0x37ffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.167551 kernel: acpi PNP0A08:00: ECAM at [mem 0x37fff0000000-0x37ffffffffff] for [bus 00-ff] Apr 30 12:42:12.167563 kernel: PCI host bridge to bus 000d:00 Apr 30 12:42:12.167636 kernel: pci_bus 000d:00: root bus resource [mem 0x50000000-0x5fffffff window] Apr 30 12:42:12.167695 kernel: pci_bus 000d:00: root bus resource [mem 0x340000000000-0x37ffdfffffff window] Apr 30 12:42:12.167754 kernel: pci_bus 000d:00: root bus resource [bus 00-ff] Apr 30 12:42:12.167832 kernel: pci 000d:00:00.0: [1def:e100] type 00 class 0x060000 Apr 30 12:42:12.167909 kernel: pci 000d:00:01.0: [1def:e101] type 01 class 0x060400 Apr 30 12:42:12.167985 kernel: pci 000d:00:01.0: enabling Extended Tags Apr 30 12:42:12.168051 kernel: pci 000d:00:01.0: supports D1 D2 Apr 30 12:42:12.168116 kernel: pci 000d:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.168190 kernel: pci 000d:00:02.0: [1def:e102] type 01 class 0x060400 Apr 30 12:42:12.168257 kernel: pci 000d:00:02.0: supports D1 D2 Apr 30 12:42:12.168321 kernel: pci 000d:00:02.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.168395 kernel: pci 000d:00:03.0: [1def:e103] type 01 class 0x060400 Apr 30 12:42:12.168462 kernel: pci 000d:00:03.0: supports D1 D2 Apr 30 12:42:12.168528 kernel: pci 000d:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.168603 kernel: pci 000d:00:04.0: [1def:e104] type 01 class 0x060400 Apr 30 12:42:12.168668 kernel: pci 000d:00:04.0: supports D1 D2 Apr 30 12:42:12.168733 kernel: pci 000d:00:04.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.168743 kernel: acpiphp: Slot [1] registered Apr 30 12:42:12.168750 kernel: acpiphp: Slot [2] registered Apr 30 12:42:12.168760 kernel: acpiphp: Slot [3] registered Apr 30 12:42:12.168767 kernel: acpiphp: Slot [4] registered Apr 30 12:42:12.168825 kernel: pci_bus 000d:00: on NUMA node 0 Apr 30 12:42:12.168890 kernel: pci 000d:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.168964 kernel: pci 000d:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.169029 kernel: pci 000d:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.169094 kernel: pci 000d:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.169161 kernel: pci 000d:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.169231 kernel: pci 000d:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.169296 kernel: pci 000d:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.169361 kernel: pci 000d:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.169425 kernel: pci 000d:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.169491 kernel: pci 000d:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.169555 kernel: pci 000d:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.169620 kernel: pci 000d:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.169688 kernel: pci 000d:00:01.0: BAR 14: assigned [mem 0x50000000-0x501fffff] Apr 30 12:42:12.169751 kernel: pci 000d:00:01.0: BAR 15: assigned [mem 0x340000000000-0x3400001fffff 64bit pref] Apr 30 12:42:12.169817 kernel: pci 000d:00:02.0: BAR 14: assigned [mem 0x50200000-0x503fffff] Apr 30 12:42:12.169881 kernel: pci 000d:00:02.0: BAR 15: assigned [mem 0x340000200000-0x3400003fffff 64bit pref] Apr 30 12:42:12.169945 kernel: pci 000d:00:03.0: BAR 14: assigned [mem 0x50400000-0x505fffff] Apr 30 12:42:12.170013 kernel: pci 000d:00:03.0: BAR 15: assigned [mem 0x340000400000-0x3400005fffff 64bit pref] Apr 30 12:42:12.170077 kernel: pci 000d:00:04.0: BAR 14: assigned [mem 0x50600000-0x507fffff] Apr 30 12:42:12.170145 kernel: pci 000d:00:04.0: BAR 15: assigned [mem 0x340000600000-0x3400007fffff 64bit pref] Apr 30 12:42:12.170209 kernel: pci 000d:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.170273 kernel: pci 000d:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.170336 kernel: pci 000d:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.170401 kernel: pci 000d:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.170464 kernel: pci 000d:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.170529 kernel: pci 000d:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.170593 kernel: pci 000d:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.170661 kernel: pci 000d:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.170725 kernel: pci 000d:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.170789 kernel: pci 000d:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.170853 kernel: pci 000d:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.170917 kernel: pci 000d:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.170984 kernel: pci 000d:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.171048 kernel: pci 000d:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.171115 kernel: pci 000d:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.171182 kernel: pci 000d:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.171245 kernel: pci 000d:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.171310 kernel: pci 000d:00:01.0: bridge window [mem 0x50000000-0x501fffff] Apr 30 12:42:12.171374 kernel: pci 000d:00:01.0: bridge window [mem 0x340000000000-0x3400001fffff 64bit pref] Apr 30 12:42:12.171439 kernel: pci 000d:00:02.0: PCI bridge to [bus 02] Apr 30 12:42:12.171502 kernel: pci 000d:00:02.0: bridge window [mem 0x50200000-0x503fffff] Apr 30 12:42:12.171566 kernel: pci 000d:00:02.0: bridge window [mem 0x340000200000-0x3400003fffff 64bit pref] Apr 30 12:42:12.171632 kernel: pci 000d:00:03.0: PCI bridge to [bus 03] Apr 30 12:42:12.171697 kernel: pci 000d:00:03.0: bridge window [mem 0x50400000-0x505fffff] Apr 30 12:42:12.171762 kernel: pci 000d:00:03.0: bridge window [mem 0x340000400000-0x3400005fffff 64bit pref] Apr 30 12:42:12.171826 kernel: pci 000d:00:04.0: PCI bridge to [bus 04] Apr 30 12:42:12.171890 kernel: pci 000d:00:04.0: bridge window [mem 0x50600000-0x507fffff] Apr 30 12:42:12.171956 kernel: pci 000d:00:04.0: bridge window [mem 0x340000600000-0x3400007fffff 64bit pref] Apr 30 12:42:12.172020 kernel: pci_bus 000d:00: resource 4 [mem 0x50000000-0x5fffffff window] Apr 30 12:42:12.172077 kernel: pci_bus 000d:00: resource 5 [mem 0x340000000000-0x37ffdfffffff window] Apr 30 12:42:12.172149 kernel: pci_bus 000d:01: resource 1 [mem 0x50000000-0x501fffff] Apr 30 12:42:12.172209 kernel: pci_bus 000d:01: resource 2 [mem 0x340000000000-0x3400001fffff 64bit pref] Apr 30 12:42:12.172278 kernel: pci_bus 000d:02: resource 1 [mem 0x50200000-0x503fffff] Apr 30 12:42:12.172343 kernel: pci_bus 000d:02: resource 2 [mem 0x340000200000-0x3400003fffff 64bit pref] Apr 30 12:42:12.172421 kernel: pci_bus 000d:03: resource 1 [mem 0x50400000-0x505fffff] Apr 30 12:42:12.172482 kernel: pci_bus 000d:03: resource 2 [mem 0x340000400000-0x3400005fffff 64bit pref] Apr 30 12:42:12.172549 kernel: pci_bus 000d:04: resource 1 [mem 0x50600000-0x507fffff] Apr 30 12:42:12.172610 kernel: pci_bus 000d:04: resource 2 [mem 0x340000600000-0x3400007fffff 64bit pref] Apr 30 12:42:12.172619 kernel: ACPI: PCI Root Bridge [PCI3] (domain 0000 [bus 00-ff]) Apr 30 12:42:12.172688 kernel: acpi PNP0A08:01: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.172754 kernel: acpi PNP0A08:01: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.172817 kernel: acpi PNP0A08:01: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.172878 kernel: acpi PNP0A08:01: MCFG quirk: ECAM at [mem 0x3ffff0000000-0x3fffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.172941 kernel: acpi PNP0A08:01: ECAM area [mem 0x3ffff0000000-0x3fffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.173007 kernel: acpi PNP0A08:01: ECAM at [mem 0x3ffff0000000-0x3fffffffffff] for [bus 00-ff] Apr 30 12:42:12.173016 kernel: PCI host bridge to bus 0000:00 Apr 30 12:42:12.173084 kernel: pci_bus 0000:00: root bus resource [mem 0x70000000-0x7fffffff window] Apr 30 12:42:12.173145 kernel: pci_bus 0000:00: root bus resource [mem 0x3c0000000000-0x3fffdfffffff window] Apr 30 12:42:12.173203 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 30 12:42:12.173276 kernel: pci 0000:00:00.0: [1def:e100] type 00 class 0x060000 Apr 30 12:42:12.173347 kernel: pci 0000:00:01.0: [1def:e101] type 01 class 0x060400 Apr 30 12:42:12.173414 kernel: pci 0000:00:01.0: enabling Extended Tags Apr 30 12:42:12.173478 kernel: pci 0000:00:01.0: supports D1 D2 Apr 30 12:42:12.173542 kernel: pci 0000:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.173616 kernel: pci 0000:00:02.0: [1def:e102] type 01 class 0x060400 Apr 30 12:42:12.173681 kernel: pci 0000:00:02.0: supports D1 D2 Apr 30 12:42:12.173745 kernel: pci 0000:00:02.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.173815 kernel: pci 0000:00:03.0: [1def:e103] type 01 class 0x060400 Apr 30 12:42:12.173881 kernel: pci 0000:00:03.0: supports D1 D2 Apr 30 12:42:12.173944 kernel: pci 0000:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.174019 kernel: pci 0000:00:04.0: [1def:e104] type 01 class 0x060400 Apr 30 12:42:12.174086 kernel: pci 0000:00:04.0: supports D1 D2 Apr 30 12:42:12.174150 kernel: pci 0000:00:04.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.174160 kernel: acpiphp: Slot [1-1] registered Apr 30 12:42:12.174167 kernel: acpiphp: Slot [2-1] registered Apr 30 12:42:12.174175 kernel: acpiphp: Slot [3-1] registered Apr 30 12:42:12.174182 kernel: acpiphp: Slot [4-1] registered Apr 30 12:42:12.174238 kernel: pci_bus 0000:00: on NUMA node 0 Apr 30 12:42:12.174303 kernel: pci 0000:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.174370 kernel: pci 0000:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.174434 kernel: pci 0000:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.174500 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.174564 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.174628 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.174692 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.174756 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.174822 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.174887 kernel: pci 0000:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.174953 kernel: pci 0000:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.175018 kernel: pci 0000:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.175083 kernel: pci 0000:00:01.0: BAR 14: assigned [mem 0x70000000-0x701fffff] Apr 30 12:42:12.175147 kernel: pci 0000:00:01.0: BAR 15: assigned [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Apr 30 12:42:12.175212 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x70200000-0x703fffff] Apr 30 12:42:12.175278 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Apr 30 12:42:12.175343 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x70400000-0x705fffff] Apr 30 12:42:12.175407 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Apr 30 12:42:12.175471 kernel: pci 0000:00:04.0: BAR 14: assigned [mem 0x70600000-0x707fffff] Apr 30 12:42:12.175535 kernel: pci 0000:00:04.0: BAR 15: assigned [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Apr 30 12:42:12.175599 kernel: pci 0000:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.175664 kernel: pci 0000:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.175728 kernel: pci 0000:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.175795 kernel: pci 0000:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.175858 kernel: pci 0000:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.175923 kernel: pci 0000:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.175994 kernel: pci 0000:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.176059 kernel: pci 0000:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.176123 kernel: pci 0000:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.176186 kernel: pci 0000:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.176251 kernel: pci 0000:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.176317 kernel: pci 0000:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.176382 kernel: pci 0000:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.176446 kernel: pci 0000:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.176510 kernel: pci 0000:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.176574 kernel: pci 0000:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.176637 kernel: pci 0000:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.176701 kernel: pci 0000:00:01.0: bridge window [mem 0x70000000-0x701fffff] Apr 30 12:42:12.176765 kernel: pci 0000:00:01.0: bridge window [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Apr 30 12:42:12.176832 kernel: pci 0000:00:02.0: PCI bridge to [bus 02] Apr 30 12:42:12.176895 kernel: pci 0000:00:02.0: bridge window [mem 0x70200000-0x703fffff] Apr 30 12:42:12.176963 kernel: pci 0000:00:02.0: bridge window [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Apr 30 12:42:12.177029 kernel: pci 0000:00:03.0: PCI bridge to [bus 03] Apr 30 12:42:12.177095 kernel: pci 0000:00:03.0: bridge window [mem 0x70400000-0x705fffff] Apr 30 12:42:12.177160 kernel: pci 0000:00:03.0: bridge window [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Apr 30 12:42:12.177225 kernel: pci 0000:00:04.0: PCI bridge to [bus 04] Apr 30 12:42:12.177289 kernel: pci 0000:00:04.0: bridge window [mem 0x70600000-0x707fffff] Apr 30 12:42:12.177352 kernel: pci 0000:00:04.0: bridge window [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Apr 30 12:42:12.177412 kernel: pci_bus 0000:00: resource 4 [mem 0x70000000-0x7fffffff window] Apr 30 12:42:12.177472 kernel: pci_bus 0000:00: resource 5 [mem 0x3c0000000000-0x3fffdfffffff window] Apr 30 12:42:12.177543 kernel: pci_bus 0000:01: resource 1 [mem 0x70000000-0x701fffff] Apr 30 12:42:12.177604 kernel: pci_bus 0000:01: resource 2 [mem 0x3c0000000000-0x3c00001fffff 64bit pref] Apr 30 12:42:12.177671 kernel: pci_bus 0000:02: resource 1 [mem 0x70200000-0x703fffff] Apr 30 12:42:12.177732 kernel: pci_bus 0000:02: resource 2 [mem 0x3c0000200000-0x3c00003fffff 64bit pref] Apr 30 12:42:12.177806 kernel: pci_bus 0000:03: resource 1 [mem 0x70400000-0x705fffff] Apr 30 12:42:12.177871 kernel: pci_bus 0000:03: resource 2 [mem 0x3c0000400000-0x3c00005fffff 64bit pref] Apr 30 12:42:12.177938 kernel: pci_bus 0000:04: resource 1 [mem 0x70600000-0x707fffff] Apr 30 12:42:12.178002 kernel: pci_bus 0000:04: resource 2 [mem 0x3c0000600000-0x3c00007fffff 64bit pref] Apr 30 12:42:12.178012 kernel: ACPI: PCI Root Bridge [PCI7] (domain 0005 [bus 00-ff]) Apr 30 12:42:12.178083 kernel: acpi PNP0A08:02: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.178145 kernel: acpi PNP0A08:02: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.178211 kernel: acpi PNP0A08:02: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.178273 kernel: acpi PNP0A08:02: MCFG quirk: ECAM at [mem 0x2ffff0000000-0x2fffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.178335 kernel: acpi PNP0A08:02: ECAM area [mem 0x2ffff0000000-0x2fffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.178398 kernel: acpi PNP0A08:02: ECAM at [mem 0x2ffff0000000-0x2fffffffffff] for [bus 00-ff] Apr 30 12:42:12.178407 kernel: PCI host bridge to bus 0005:00 Apr 30 12:42:12.178471 kernel: pci_bus 0005:00: root bus resource [mem 0x30000000-0x3fffffff window] Apr 30 12:42:12.178530 kernel: pci_bus 0005:00: root bus resource [mem 0x2c0000000000-0x2fffdfffffff window] Apr 30 12:42:12.178589 kernel: pci_bus 0005:00: root bus resource [bus 00-ff] Apr 30 12:42:12.178664 kernel: pci 0005:00:00.0: [1def:e110] type 00 class 0x060000 Apr 30 12:42:12.178736 kernel: pci 0005:00:01.0: [1def:e111] type 01 class 0x060400 Apr 30 12:42:12.178802 kernel: pci 0005:00:01.0: supports D1 D2 Apr 30 12:42:12.178866 kernel: pci 0005:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.178937 kernel: pci 0005:00:03.0: [1def:e113] type 01 class 0x060400 Apr 30 12:42:12.179005 kernel: pci 0005:00:03.0: supports D1 D2 Apr 30 12:42:12.179072 kernel: pci 0005:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.179142 kernel: pci 0005:00:05.0: [1def:e115] type 01 class 0x060400 Apr 30 12:42:12.179206 kernel: pci 0005:00:05.0: supports D1 D2 Apr 30 12:42:12.179271 kernel: pci 0005:00:05.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.179341 kernel: pci 0005:00:07.0: [1def:e117] type 01 class 0x060400 Apr 30 12:42:12.179409 kernel: pci 0005:00:07.0: supports D1 D2 Apr 30 12:42:12.179473 kernel: pci 0005:00:07.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.179485 kernel: acpiphp: Slot [1-2] registered Apr 30 12:42:12.179492 kernel: acpiphp: Slot [2-2] registered Apr 30 12:42:12.179566 kernel: pci 0005:03:00.0: [144d:a808] type 00 class 0x010802 Apr 30 12:42:12.179636 kernel: pci 0005:03:00.0: reg 0x10: [mem 0x30110000-0x30113fff 64bit] Apr 30 12:42:12.179702 kernel: pci 0005:03:00.0: reg 0x30: [mem 0x30100000-0x3010ffff pref] Apr 30 12:42:12.179778 kernel: pci 0005:04:00.0: [144d:a808] type 00 class 0x010802 Apr 30 12:42:12.179847 kernel: pci 0005:04:00.0: reg 0x10: [mem 0x30010000-0x30013fff 64bit] Apr 30 12:42:12.179915 kernel: pci 0005:04:00.0: reg 0x30: [mem 0x30000000-0x3000ffff pref] Apr 30 12:42:12.179978 kernel: pci_bus 0005:00: on NUMA node 0 Apr 30 12:42:12.180043 kernel: pci 0005:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.180109 kernel: pci 0005:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.180173 kernel: pci 0005:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.180239 kernel: pci 0005:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.180307 kernel: pci 0005:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.180377 kernel: pci 0005:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.180441 kernel: pci 0005:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.180507 kernel: pci 0005:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.180572 kernel: pci 0005:00:05.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Apr 30 12:42:12.180638 kernel: pci 0005:00:07.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.180742 kernel: pci 0005:00:07.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.180813 kernel: pci 0005:00:07.0: bridge window [mem 0x00100000-0x001fffff] to [bus 04] add_size 100000 add_align 100000 Apr 30 12:42:12.180879 kernel: pci 0005:00:01.0: BAR 14: assigned [mem 0x30000000-0x301fffff] Apr 30 12:42:12.180965 kernel: pci 0005:00:01.0: BAR 15: assigned [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Apr 30 12:42:12.181033 kernel: pci 0005:00:03.0: BAR 14: assigned [mem 0x30200000-0x303fffff] Apr 30 12:42:12.181096 kernel: pci 0005:00:03.0: BAR 15: assigned [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Apr 30 12:42:12.181161 kernel: pci 0005:00:05.0: BAR 14: assigned [mem 0x30400000-0x305fffff] Apr 30 12:42:12.181225 kernel: pci 0005:00:05.0: BAR 15: assigned [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Apr 30 12:42:12.181289 kernel: pci 0005:00:07.0: BAR 14: assigned [mem 0x30600000-0x307fffff] Apr 30 12:42:12.181356 kernel: pci 0005:00:07.0: BAR 15: assigned [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Apr 30 12:42:12.181420 kernel: pci 0005:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.181485 kernel: pci 0005:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.181549 kernel: pci 0005:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.181616 kernel: pci 0005:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.181680 kernel: pci 0005:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.181743 kernel: pci 0005:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.181807 kernel: pci 0005:00:07.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.181874 kernel: pci 0005:00:07.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.181940 kernel: pci 0005:00:07.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.182006 kernel: pci 0005:00:07.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.182070 kernel: pci 0005:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.182134 kernel: pci 0005:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.182199 kernel: pci 0005:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.182263 kernel: pci 0005:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.182327 kernel: pci 0005:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.182392 kernel: pci 0005:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.182460 kernel: pci 0005:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.182524 kernel: pci 0005:00:01.0: bridge window [mem 0x30000000-0x301fffff] Apr 30 12:42:12.182588 kernel: pci 0005:00:01.0: bridge window [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Apr 30 12:42:12.182653 kernel: pci 0005:00:03.0: PCI bridge to [bus 02] Apr 30 12:42:12.182718 kernel: pci 0005:00:03.0: bridge window [mem 0x30200000-0x303fffff] Apr 30 12:42:12.182782 kernel: pci 0005:00:03.0: bridge window [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Apr 30 12:42:12.182853 kernel: pci 0005:03:00.0: BAR 6: assigned [mem 0x30400000-0x3040ffff pref] Apr 30 12:42:12.182919 kernel: pci 0005:03:00.0: BAR 0: assigned [mem 0x30410000-0x30413fff 64bit] Apr 30 12:42:12.182987 kernel: pci 0005:00:05.0: PCI bridge to [bus 03] Apr 30 12:42:12.183050 kernel: pci 0005:00:05.0: bridge window [mem 0x30400000-0x305fffff] Apr 30 12:42:12.183115 kernel: pci 0005:00:05.0: bridge window [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Apr 30 12:42:12.183183 kernel: pci 0005:04:00.0: BAR 6: assigned [mem 0x30600000-0x3060ffff pref] Apr 30 12:42:12.183249 kernel: pci 0005:04:00.0: BAR 0: assigned [mem 0x30610000-0x30613fff 64bit] Apr 30 12:42:12.183317 kernel: pci 0005:00:07.0: PCI bridge to [bus 04] Apr 30 12:42:12.183383 kernel: pci 0005:00:07.0: bridge window [mem 0x30600000-0x307fffff] Apr 30 12:42:12.183448 kernel: pci 0005:00:07.0: bridge window [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Apr 30 12:42:12.183509 kernel: pci_bus 0005:00: resource 4 [mem 0x30000000-0x3fffffff window] Apr 30 12:42:12.183568 kernel: pci_bus 0005:00: resource 5 [mem 0x2c0000000000-0x2fffdfffffff window] Apr 30 12:42:12.183638 kernel: pci_bus 0005:01: resource 1 [mem 0x30000000-0x301fffff] Apr 30 12:42:12.183700 kernel: pci_bus 0005:01: resource 2 [mem 0x2c0000000000-0x2c00001fffff 64bit pref] Apr 30 12:42:12.183780 kernel: pci_bus 0005:02: resource 1 [mem 0x30200000-0x303fffff] Apr 30 12:42:12.183840 kernel: pci_bus 0005:02: resource 2 [mem 0x2c0000200000-0x2c00003fffff 64bit pref] Apr 30 12:42:12.183908 kernel: pci_bus 0005:03: resource 1 [mem 0x30400000-0x305fffff] Apr 30 12:42:12.183972 kernel: pci_bus 0005:03: resource 2 [mem 0x2c0000400000-0x2c00005fffff 64bit pref] Apr 30 12:42:12.184042 kernel: pci_bus 0005:04: resource 1 [mem 0x30600000-0x307fffff] Apr 30 12:42:12.184104 kernel: pci_bus 0005:04: resource 2 [mem 0x2c0000600000-0x2c00007fffff 64bit pref] Apr 30 12:42:12.184114 kernel: ACPI: PCI Root Bridge [PCI5] (domain 0003 [bus 00-ff]) Apr 30 12:42:12.184184 kernel: acpi PNP0A08:03: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.184247 kernel: acpi PNP0A08:03: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.184309 kernel: acpi PNP0A08:03: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.184372 kernel: acpi PNP0A08:03: MCFG quirk: ECAM at [mem 0x27fff0000000-0x27ffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.184433 kernel: acpi PNP0A08:03: ECAM area [mem 0x27fff0000000-0x27ffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.184498 kernel: acpi PNP0A08:03: ECAM at [mem 0x27fff0000000-0x27ffffffffff] for [bus 00-ff] Apr 30 12:42:12.184508 kernel: PCI host bridge to bus 0003:00 Apr 30 12:42:12.184574 kernel: pci_bus 0003:00: root bus resource [mem 0x10000000-0x1fffffff window] Apr 30 12:42:12.184632 kernel: pci_bus 0003:00: root bus resource [mem 0x240000000000-0x27ffdfffffff window] Apr 30 12:42:12.184690 kernel: pci_bus 0003:00: root bus resource [bus 00-ff] Apr 30 12:42:12.184761 kernel: pci 0003:00:00.0: [1def:e110] type 00 class 0x060000 Apr 30 12:42:12.184835 kernel: pci 0003:00:01.0: [1def:e111] type 01 class 0x060400 Apr 30 12:42:12.184905 kernel: pci 0003:00:01.0: supports D1 D2 Apr 30 12:42:12.185031 kernel: pci 0003:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.185108 kernel: pci 0003:00:03.0: [1def:e113] type 01 class 0x060400 Apr 30 12:42:12.185175 kernel: pci 0003:00:03.0: supports D1 D2 Apr 30 12:42:12.185242 kernel: pci 0003:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.185316 kernel: pci 0003:00:05.0: [1def:e115] type 01 class 0x060400 Apr 30 12:42:12.185385 kernel: pci 0003:00:05.0: supports D1 D2 Apr 30 12:42:12.185447 kernel: pci 0003:00:05.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.185457 kernel: acpiphp: Slot [1-3] registered Apr 30 12:42:12.185464 kernel: acpiphp: Slot [2-3] registered Apr 30 12:42:12.185537 kernel: pci 0003:03:00.0: [8086:1521] type 00 class 0x020000 Apr 30 12:42:12.185602 kernel: pci 0003:03:00.0: reg 0x10: [mem 0x10020000-0x1003ffff] Apr 30 12:42:12.185667 kernel: pci 0003:03:00.0: reg 0x18: [io 0x0020-0x003f] Apr 30 12:42:12.185733 kernel: pci 0003:03:00.0: reg 0x1c: [mem 0x10044000-0x10047fff] Apr 30 12:42:12.185803 kernel: pci 0003:03:00.0: PME# supported from D0 D3hot D3cold Apr 30 12:42:12.185872 kernel: pci 0003:03:00.0: reg 0x184: [mem 0x240000060000-0x240000063fff 64bit pref] Apr 30 12:42:12.185939 kernel: pci 0003:03:00.0: VF(n) BAR0 space: [mem 0x240000060000-0x24000007ffff 64bit pref] (contains BAR0 for 8 VFs) Apr 30 12:42:12.186010 kernel: pci 0003:03:00.0: reg 0x190: [mem 0x240000040000-0x240000043fff 64bit pref] Apr 30 12:42:12.186077 kernel: pci 0003:03:00.0: VF(n) BAR3 space: [mem 0x240000040000-0x24000005ffff 64bit pref] (contains BAR3 for 8 VFs) Apr 30 12:42:12.186142 kernel: pci 0003:03:00.0: 8.000 Gb/s available PCIe bandwidth, limited by 5.0 GT/s PCIe x2 link at 0003:00:05.0 (capable of 16.000 Gb/s with 5.0 GT/s PCIe x4 link) Apr 30 12:42:12.186215 kernel: pci 0003:03:00.1: [8086:1521] type 00 class 0x020000 Apr 30 12:42:12.186284 kernel: pci 0003:03:00.1: reg 0x10: [mem 0x10000000-0x1001ffff] Apr 30 12:42:12.186349 kernel: pci 0003:03:00.1: reg 0x18: [io 0x0000-0x001f] Apr 30 12:42:12.186414 kernel: pci 0003:03:00.1: reg 0x1c: [mem 0x10040000-0x10043fff] Apr 30 12:42:12.186479 kernel: pci 0003:03:00.1: PME# supported from D0 D3hot D3cold Apr 30 12:42:12.186543 kernel: pci 0003:03:00.1: reg 0x184: [mem 0x240000020000-0x240000023fff 64bit pref] Apr 30 12:42:12.186608 kernel: pci 0003:03:00.1: VF(n) BAR0 space: [mem 0x240000020000-0x24000003ffff 64bit pref] (contains BAR0 for 8 VFs) Apr 30 12:42:12.186673 kernel: pci 0003:03:00.1: reg 0x190: [mem 0x240000000000-0x240000003fff 64bit pref] Apr 30 12:42:12.186740 kernel: pci 0003:03:00.1: VF(n) BAR3 space: [mem 0x240000000000-0x24000001ffff 64bit pref] (contains BAR3 for 8 VFs) Apr 30 12:42:12.186798 kernel: pci_bus 0003:00: on NUMA node 0 Apr 30 12:42:12.186863 kernel: pci 0003:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.186926 kernel: pci 0003:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.186993 kernel: pci 0003:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.187057 kernel: pci 0003:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.187120 kernel: pci 0003:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.187185 kernel: pci 0003:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.187249 kernel: pci 0003:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03-04] add_size 300000 add_align 100000 Apr 30 12:42:12.187314 kernel: pci 0003:00:05.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03-04] add_size 100000 add_align 100000 Apr 30 12:42:12.187377 kernel: pci 0003:00:01.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Apr 30 12:42:12.187452 kernel: pci 0003:00:01.0: BAR 15: assigned [mem 0x240000000000-0x2400001fffff 64bit pref] Apr 30 12:42:12.187518 kernel: pci 0003:00:03.0: BAR 14: assigned [mem 0x10200000-0x103fffff] Apr 30 12:42:12.187583 kernel: pci 0003:00:03.0: BAR 15: assigned [mem 0x240000200000-0x2400003fffff 64bit pref] Apr 30 12:42:12.187647 kernel: pci 0003:00:05.0: BAR 14: assigned [mem 0x10400000-0x105fffff] Apr 30 12:42:12.187715 kernel: pci 0003:00:05.0: BAR 15: assigned [mem 0x240000400000-0x2400006fffff 64bit pref] Apr 30 12:42:12.187780 kernel: pci 0003:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.187843 kernel: pci 0003:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.187908 kernel: pci 0003:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.187975 kernel: pci 0003:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.188040 kernel: pci 0003:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.188105 kernel: pci 0003:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.188170 kernel: pci 0003:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.188236 kernel: pci 0003:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.188300 kernel: pci 0003:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.188365 kernel: pci 0003:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.188428 kernel: pci 0003:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.188493 kernel: pci 0003:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.188557 kernel: pci 0003:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.188622 kernel: pci 0003:00:01.0: bridge window [mem 0x10000000-0x101fffff] Apr 30 12:42:12.188687 kernel: pci 0003:00:01.0: bridge window [mem 0x240000000000-0x2400001fffff 64bit pref] Apr 30 12:42:12.188755 kernel: pci 0003:00:03.0: PCI bridge to [bus 02] Apr 30 12:42:12.188822 kernel: pci 0003:00:03.0: bridge window [mem 0x10200000-0x103fffff] Apr 30 12:42:12.188886 kernel: pci 0003:00:03.0: bridge window [mem 0x240000200000-0x2400003fffff 64bit pref] Apr 30 12:42:12.189141 kernel: pci 0003:03:00.0: BAR 0: assigned [mem 0x10400000-0x1041ffff] Apr 30 12:42:12.189224 kernel: pci 0003:03:00.1: BAR 0: assigned [mem 0x10420000-0x1043ffff] Apr 30 12:42:12.189290 kernel: pci 0003:03:00.0: BAR 3: assigned [mem 0x10440000-0x10443fff] Apr 30 12:42:12.189361 kernel: pci 0003:03:00.0: BAR 7: assigned [mem 0x240000400000-0x24000041ffff 64bit pref] Apr 30 12:42:12.189426 kernel: pci 0003:03:00.0: BAR 10: assigned [mem 0x240000420000-0x24000043ffff 64bit pref] Apr 30 12:42:12.189491 kernel: pci 0003:03:00.1: BAR 3: assigned [mem 0x10444000-0x10447fff] Apr 30 12:42:12.189559 kernel: pci 0003:03:00.1: BAR 7: assigned [mem 0x240000440000-0x24000045ffff 64bit pref] Apr 30 12:42:12.189625 kernel: pci 0003:03:00.1: BAR 10: assigned [mem 0x240000460000-0x24000047ffff 64bit pref] Apr 30 12:42:12.189690 kernel: pci 0003:03:00.0: BAR 2: no space for [io size 0x0020] Apr 30 12:42:12.189755 kernel: pci 0003:03:00.0: BAR 2: failed to assign [io size 0x0020] Apr 30 12:42:12.189823 kernel: pci 0003:03:00.1: BAR 2: no space for [io size 0x0020] Apr 30 12:42:12.189888 kernel: pci 0003:03:00.1: BAR 2: failed to assign [io size 0x0020] Apr 30 12:42:12.189957 kernel: pci 0003:03:00.0: BAR 2: no space for [io size 0x0020] Apr 30 12:42:12.190023 kernel: pci 0003:03:00.0: BAR 2: failed to assign [io size 0x0020] Apr 30 12:42:12.190088 kernel: pci 0003:03:00.1: BAR 2: no space for [io size 0x0020] Apr 30 12:42:12.190152 kernel: pci 0003:03:00.1: BAR 2: failed to assign [io size 0x0020] Apr 30 12:42:12.190217 kernel: pci 0003:00:05.0: PCI bridge to [bus 03-04] Apr 30 12:42:12.190280 kernel: pci 0003:00:05.0: bridge window [mem 0x10400000-0x105fffff] Apr 30 12:42:12.190347 kernel: pci 0003:00:05.0: bridge window [mem 0x240000400000-0x2400006fffff 64bit pref] Apr 30 12:42:12.190406 kernel: pci_bus 0003:00: Some PCI device resources are unassigned, try booting with pci=realloc Apr 30 12:42:12.190463 kernel: pci_bus 0003:00: resource 4 [mem 0x10000000-0x1fffffff window] Apr 30 12:42:12.190520 kernel: pci_bus 0003:00: resource 5 [mem 0x240000000000-0x27ffdfffffff window] Apr 30 12:42:12.190598 kernel: pci_bus 0003:01: resource 1 [mem 0x10000000-0x101fffff] Apr 30 12:42:12.190659 kernel: pci_bus 0003:01: resource 2 [mem 0x240000000000-0x2400001fffff 64bit pref] Apr 30 12:42:12.190729 kernel: pci_bus 0003:02: resource 1 [mem 0x10200000-0x103fffff] Apr 30 12:42:12.190790 kernel: pci_bus 0003:02: resource 2 [mem 0x240000200000-0x2400003fffff 64bit pref] Apr 30 12:42:12.190855 kernel: pci_bus 0003:03: resource 1 [mem 0x10400000-0x105fffff] Apr 30 12:42:12.190915 kernel: pci_bus 0003:03: resource 2 [mem 0x240000400000-0x2400006fffff 64bit pref] Apr 30 12:42:12.190925 kernel: ACPI: PCI Root Bridge [PCI0] (domain 000c [bus 00-ff]) Apr 30 12:42:12.190997 kernel: acpi PNP0A08:04: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.191063 kernel: acpi PNP0A08:04: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.191124 kernel: acpi PNP0A08:04: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.191185 kernel: acpi PNP0A08:04: MCFG quirk: ECAM at [mem 0x33fff0000000-0x33ffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.191246 kernel: acpi PNP0A08:04: ECAM area [mem 0x33fff0000000-0x33ffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.191306 kernel: acpi PNP0A08:04: ECAM at [mem 0x33fff0000000-0x33ffffffffff] for [bus 00-ff] Apr 30 12:42:12.191316 kernel: PCI host bridge to bus 000c:00 Apr 30 12:42:12.191380 kernel: pci_bus 000c:00: root bus resource [mem 0x40000000-0x4fffffff window] Apr 30 12:42:12.191439 kernel: pci_bus 000c:00: root bus resource [mem 0x300000000000-0x33ffdfffffff window] Apr 30 12:42:12.191495 kernel: pci_bus 000c:00: root bus resource [bus 00-ff] Apr 30 12:42:12.191567 kernel: pci 000c:00:00.0: [1def:e100] type 00 class 0x060000 Apr 30 12:42:12.191639 kernel: pci 000c:00:01.0: [1def:e101] type 01 class 0x060400 Apr 30 12:42:12.191704 kernel: pci 000c:00:01.0: enabling Extended Tags Apr 30 12:42:12.191767 kernel: pci 000c:00:01.0: supports D1 D2 Apr 30 12:42:12.191831 kernel: pci 000c:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.191904 kernel: pci 000c:00:02.0: [1def:e102] type 01 class 0x060400 Apr 30 12:42:12.191973 kernel: pci 000c:00:02.0: supports D1 D2 Apr 30 12:42:12.192037 kernel: pci 000c:00:02.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.192108 kernel: pci 000c:00:03.0: [1def:e103] type 01 class 0x060400 Apr 30 12:42:12.192173 kernel: pci 000c:00:03.0: supports D1 D2 Apr 30 12:42:12.192235 kernel: pci 000c:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.192306 kernel: pci 000c:00:04.0: [1def:e104] type 01 class 0x060400 Apr 30 12:42:12.192373 kernel: pci 000c:00:04.0: supports D1 D2 Apr 30 12:42:12.192437 kernel: pci 000c:00:04.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.192447 kernel: acpiphp: Slot [1-4] registered Apr 30 12:42:12.192456 kernel: acpiphp: Slot [2-4] registered Apr 30 12:42:12.192464 kernel: acpiphp: Slot [3-2] registered Apr 30 12:42:12.192472 kernel: acpiphp: Slot [4-2] registered Apr 30 12:42:12.192529 kernel: pci_bus 000c:00: on NUMA node 0 Apr 30 12:42:12.192594 kernel: pci 000c:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.192662 kernel: pci 000c:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.192725 kernel: pci 000c:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.192789 kernel: pci 000c:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.192852 kernel: pci 000c:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.192916 kernel: pci 000c:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.192982 kernel: pci 000c:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.193046 kernel: pci 000c:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.193111 kernel: pci 000c:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.193178 kernel: pci 000c:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.193242 kernel: pci 000c:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.193304 kernel: pci 000c:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.193369 kernel: pci 000c:00:01.0: BAR 14: assigned [mem 0x40000000-0x401fffff] Apr 30 12:42:12.193432 kernel: pci 000c:00:01.0: BAR 15: assigned [mem 0x300000000000-0x3000001fffff 64bit pref] Apr 30 12:42:12.193496 kernel: pci 000c:00:02.0: BAR 14: assigned [mem 0x40200000-0x403fffff] Apr 30 12:42:12.193561 kernel: pci 000c:00:02.0: BAR 15: assigned [mem 0x300000200000-0x3000003fffff 64bit pref] Apr 30 12:42:12.193625 kernel: pci 000c:00:03.0: BAR 14: assigned [mem 0x40400000-0x405fffff] Apr 30 12:42:12.193689 kernel: pci 000c:00:03.0: BAR 15: assigned [mem 0x300000400000-0x3000005fffff 64bit pref] Apr 30 12:42:12.193751 kernel: pci 000c:00:04.0: BAR 14: assigned [mem 0x40600000-0x407fffff] Apr 30 12:42:12.193815 kernel: pci 000c:00:04.0: BAR 15: assigned [mem 0x300000600000-0x3000007fffff 64bit pref] Apr 30 12:42:12.193877 kernel: pci 000c:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.193941 kernel: pci 000c:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194007 kernel: pci 000c:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194073 kernel: pci 000c:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194136 kernel: pci 000c:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194198 kernel: pci 000c:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194262 kernel: pci 000c:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194324 kernel: pci 000c:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194387 kernel: pci 000c:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194449 kernel: pci 000c:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194513 kernel: pci 000c:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194575 kernel: pci 000c:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194641 kernel: pci 000c:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194704 kernel: pci 000c:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194766 kernel: pci 000c:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.194830 kernel: pci 000c:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.194893 kernel: pci 000c:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.194961 kernel: pci 000c:00:01.0: bridge window [mem 0x40000000-0x401fffff] Apr 30 12:42:12.195025 kernel: pci 000c:00:01.0: bridge window [mem 0x300000000000-0x3000001fffff 64bit pref] Apr 30 12:42:12.195091 kernel: pci 000c:00:02.0: PCI bridge to [bus 02] Apr 30 12:42:12.195155 kernel: pci 000c:00:02.0: bridge window [mem 0x40200000-0x403fffff] Apr 30 12:42:12.195219 kernel: pci 000c:00:02.0: bridge window [mem 0x300000200000-0x3000003fffff 64bit pref] Apr 30 12:42:12.195282 kernel: pci 000c:00:03.0: PCI bridge to [bus 03] Apr 30 12:42:12.195344 kernel: pci 000c:00:03.0: bridge window [mem 0x40400000-0x405fffff] Apr 30 12:42:12.195408 kernel: pci 000c:00:03.0: bridge window [mem 0x300000400000-0x3000005fffff 64bit pref] Apr 30 12:42:12.195471 kernel: pci 000c:00:04.0: PCI bridge to [bus 04] Apr 30 12:42:12.195537 kernel: pci 000c:00:04.0: bridge window [mem 0x40600000-0x407fffff] Apr 30 12:42:12.195600 kernel: pci 000c:00:04.0: bridge window [mem 0x300000600000-0x3000007fffff 64bit pref] Apr 30 12:42:12.195658 kernel: pci_bus 000c:00: resource 4 [mem 0x40000000-0x4fffffff window] Apr 30 12:42:12.195715 kernel: pci_bus 000c:00: resource 5 [mem 0x300000000000-0x33ffdfffffff window] Apr 30 12:42:12.195784 kernel: pci_bus 000c:01: resource 1 [mem 0x40000000-0x401fffff] Apr 30 12:42:12.195844 kernel: pci_bus 000c:01: resource 2 [mem 0x300000000000-0x3000001fffff 64bit pref] Apr 30 12:42:12.195922 kernel: pci_bus 000c:02: resource 1 [mem 0x40200000-0x403fffff] Apr 30 12:42:12.195985 kernel: pci_bus 000c:02: resource 2 [mem 0x300000200000-0x3000003fffff 64bit pref] Apr 30 12:42:12.196051 kernel: pci_bus 000c:03: resource 1 [mem 0x40400000-0x405fffff] Apr 30 12:42:12.196111 kernel: pci_bus 000c:03: resource 2 [mem 0x300000400000-0x3000005fffff 64bit pref] Apr 30 12:42:12.196177 kernel: pci_bus 000c:04: resource 1 [mem 0x40600000-0x407fffff] Apr 30 12:42:12.196238 kernel: pci_bus 000c:04: resource 2 [mem 0x300000600000-0x3000007fffff 64bit pref] Apr 30 12:42:12.196250 kernel: ACPI: PCI Root Bridge [PCI4] (domain 0002 [bus 00-ff]) Apr 30 12:42:12.196319 kernel: acpi PNP0A08:05: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.196380 kernel: acpi PNP0A08:05: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.196442 kernel: acpi PNP0A08:05: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.196502 kernel: acpi PNP0A08:05: MCFG quirk: ECAM at [mem 0x23fff0000000-0x23ffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.196563 kernel: acpi PNP0A08:05: ECAM area [mem 0x23fff0000000-0x23ffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.196624 kernel: acpi PNP0A08:05: ECAM at [mem 0x23fff0000000-0x23ffffffffff] for [bus 00-ff] Apr 30 12:42:12.196636 kernel: PCI host bridge to bus 0002:00 Apr 30 12:42:12.196703 kernel: pci_bus 0002:00: root bus resource [mem 0x00800000-0x0fffffff window] Apr 30 12:42:12.196761 kernel: pci_bus 0002:00: root bus resource [mem 0x200000000000-0x23ffdfffffff window] Apr 30 12:42:12.196817 kernel: pci_bus 0002:00: root bus resource [bus 00-ff] Apr 30 12:42:12.196887 kernel: pci 0002:00:00.0: [1def:e110] type 00 class 0x060000 Apr 30 12:42:12.196965 kernel: pci 0002:00:01.0: [1def:e111] type 01 class 0x060400 Apr 30 12:42:12.197030 kernel: pci 0002:00:01.0: supports D1 D2 Apr 30 12:42:12.197095 kernel: pci 0002:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.197165 kernel: pci 0002:00:03.0: [1def:e113] type 01 class 0x060400 Apr 30 12:42:12.197229 kernel: pci 0002:00:03.0: supports D1 D2 Apr 30 12:42:12.197293 kernel: pci 0002:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.197363 kernel: pci 0002:00:05.0: [1def:e115] type 01 class 0x060400 Apr 30 12:42:12.197426 kernel: pci 0002:00:05.0: supports D1 D2 Apr 30 12:42:12.197489 kernel: pci 0002:00:05.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.197562 kernel: pci 0002:00:07.0: [1def:e117] type 01 class 0x060400 Apr 30 12:42:12.197627 kernel: pci 0002:00:07.0: supports D1 D2 Apr 30 12:42:12.197690 kernel: pci 0002:00:07.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.197700 kernel: acpiphp: Slot [1-5] registered Apr 30 12:42:12.197708 kernel: acpiphp: Slot [2-5] registered Apr 30 12:42:12.197715 kernel: acpiphp: Slot [3-3] registered Apr 30 12:42:12.197723 kernel: acpiphp: Slot [4-3] registered Apr 30 12:42:12.197779 kernel: pci_bus 0002:00: on NUMA node 0 Apr 30 12:42:12.197846 kernel: pci 0002:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.197909 kernel: pci 0002:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.197977 kernel: pci 0002:00:01.0: bridge window [mem 0x00100000-0x000fffff] to [bus 01] add_size 200000 add_align 100000 Apr 30 12:42:12.198043 kernel: pci 0002:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.198109 kernel: pci 0002:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.198172 kernel: pci 0002:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.198237 kernel: pci 0002:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.198301 kernel: pci 0002:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.198364 kernel: pci 0002:00:05.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.198428 kernel: pci 0002:00:07.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.198491 kernel: pci 0002:00:07.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.198557 kernel: pci 0002:00:07.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.198620 kernel: pci 0002:00:01.0: BAR 14: assigned [mem 0x00800000-0x009fffff] Apr 30 12:42:12.198684 kernel: pci 0002:00:01.0: BAR 15: assigned [mem 0x200000000000-0x2000001fffff 64bit pref] Apr 30 12:42:12.198747 kernel: pci 0002:00:03.0: BAR 14: assigned [mem 0x00a00000-0x00bfffff] Apr 30 12:42:12.198810 kernel: pci 0002:00:03.0: BAR 15: assigned [mem 0x200000200000-0x2000003fffff 64bit pref] Apr 30 12:42:12.198875 kernel: pci 0002:00:05.0: BAR 14: assigned [mem 0x00c00000-0x00dfffff] Apr 30 12:42:12.198937 kernel: pci 0002:00:05.0: BAR 15: assigned [mem 0x200000400000-0x2000005fffff 64bit pref] Apr 30 12:42:12.199005 kernel: pci 0002:00:07.0: BAR 14: assigned [mem 0x00e00000-0x00ffffff] Apr 30 12:42:12.199073 kernel: pci 0002:00:07.0: BAR 15: assigned [mem 0x200000600000-0x2000007fffff 64bit pref] Apr 30 12:42:12.199139 kernel: pci 0002:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199203 kernel: pci 0002:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.199269 kernel: pci 0002:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199333 kernel: pci 0002:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.199397 kernel: pci 0002:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199462 kernel: pci 0002:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.199525 kernel: pci 0002:00:07.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199592 kernel: pci 0002:00:07.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.199655 kernel: pci 0002:00:07.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199720 kernel: pci 0002:00:07.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.199783 kernel: pci 0002:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199846 kernel: pci 0002:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.199910 kernel: pci 0002:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.199975 kernel: pci 0002:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.200039 kernel: pci 0002:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.200102 kernel: pci 0002:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.200168 kernel: pci 0002:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.200232 kernel: pci 0002:00:01.0: bridge window [mem 0x00800000-0x009fffff] Apr 30 12:42:12.200296 kernel: pci 0002:00:01.0: bridge window [mem 0x200000000000-0x2000001fffff 64bit pref] Apr 30 12:42:12.200359 kernel: pci 0002:00:03.0: PCI bridge to [bus 02] Apr 30 12:42:12.200423 kernel: pci 0002:00:03.0: bridge window [mem 0x00a00000-0x00bfffff] Apr 30 12:42:12.200486 kernel: pci 0002:00:03.0: bridge window [mem 0x200000200000-0x2000003fffff 64bit pref] Apr 30 12:42:12.200551 kernel: pci 0002:00:05.0: PCI bridge to [bus 03] Apr 30 12:42:12.200627 kernel: pci 0002:00:05.0: bridge window [mem 0x00c00000-0x00dfffff] Apr 30 12:42:12.200692 kernel: pci 0002:00:05.0: bridge window [mem 0x200000400000-0x2000005fffff 64bit pref] Apr 30 12:42:12.200756 kernel: pci 0002:00:07.0: PCI bridge to [bus 04] Apr 30 12:42:12.200821 kernel: pci 0002:00:07.0: bridge window [mem 0x00e00000-0x00ffffff] Apr 30 12:42:12.200886 kernel: pci 0002:00:07.0: bridge window [mem 0x200000600000-0x2000007fffff 64bit pref] Apr 30 12:42:12.200948 kernel: pci_bus 0002:00: resource 4 [mem 0x00800000-0x0fffffff window] Apr 30 12:42:12.201068 kernel: pci_bus 0002:00: resource 5 [mem 0x200000000000-0x23ffdfffffff window] Apr 30 12:42:12.201139 kernel: pci_bus 0002:01: resource 1 [mem 0x00800000-0x009fffff] Apr 30 12:42:12.201198 kernel: pci_bus 0002:01: resource 2 [mem 0x200000000000-0x2000001fffff 64bit pref] Apr 30 12:42:12.201264 kernel: pci_bus 0002:02: resource 1 [mem 0x00a00000-0x00bfffff] Apr 30 12:42:12.201324 kernel: pci_bus 0002:02: resource 2 [mem 0x200000200000-0x2000003fffff 64bit pref] Apr 30 12:42:12.201399 kernel: pci_bus 0002:03: resource 1 [mem 0x00c00000-0x00dfffff] Apr 30 12:42:12.201461 kernel: pci_bus 0002:03: resource 2 [mem 0x200000400000-0x2000005fffff 64bit pref] Apr 30 12:42:12.201527 kernel: pci_bus 0002:04: resource 1 [mem 0x00e00000-0x00ffffff] Apr 30 12:42:12.201586 kernel: pci_bus 0002:04: resource 2 [mem 0x200000600000-0x2000007fffff 64bit pref] Apr 30 12:42:12.201596 kernel: ACPI: PCI Root Bridge [PCI2] (domain 0001 [bus 00-ff]) Apr 30 12:42:12.201664 kernel: acpi PNP0A08:06: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.201726 kernel: acpi PNP0A08:06: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.201789 kernel: acpi PNP0A08:06: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.201850 kernel: acpi PNP0A08:06: MCFG quirk: ECAM at [mem 0x3bfff0000000-0x3bffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.201912 kernel: acpi PNP0A08:06: ECAM area [mem 0x3bfff0000000-0x3bffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.201976 kernel: acpi PNP0A08:06: ECAM at [mem 0x3bfff0000000-0x3bffffffffff] for [bus 00-ff] Apr 30 12:42:12.201987 kernel: PCI host bridge to bus 0001:00 Apr 30 12:42:12.202050 kernel: pci_bus 0001:00: root bus resource [mem 0x60000000-0x6fffffff window] Apr 30 12:42:12.202109 kernel: pci_bus 0001:00: root bus resource [mem 0x380000000000-0x3bffdfffffff window] Apr 30 12:42:12.202166 kernel: pci_bus 0001:00: root bus resource [bus 00-ff] Apr 30 12:42:12.202235 kernel: pci 0001:00:00.0: [1def:e100] type 00 class 0x060000 Apr 30 12:42:12.202307 kernel: pci 0001:00:01.0: [1def:e101] type 01 class 0x060400 Apr 30 12:42:12.202371 kernel: pci 0001:00:01.0: enabling Extended Tags Apr 30 12:42:12.202434 kernel: pci 0001:00:01.0: supports D1 D2 Apr 30 12:42:12.202497 kernel: pci 0001:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.202572 kernel: pci 0001:00:02.0: [1def:e102] type 01 class 0x060400 Apr 30 12:42:12.202637 kernel: pci 0001:00:02.0: supports D1 D2 Apr 30 12:42:12.202699 kernel: pci 0001:00:02.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.202770 kernel: pci 0001:00:03.0: [1def:e103] type 01 class 0x060400 Apr 30 12:42:12.202835 kernel: pci 0001:00:03.0: supports D1 D2 Apr 30 12:42:12.202899 kernel: pci 0001:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.202982 kernel: pci 0001:00:04.0: [1def:e104] type 01 class 0x060400 Apr 30 12:42:12.203052 kernel: pci 0001:00:04.0: supports D1 D2 Apr 30 12:42:12.203116 kernel: pci 0001:00:04.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.203126 kernel: acpiphp: Slot [1-6] registered Apr 30 12:42:12.203199 kernel: pci 0001:01:00.0: [15b3:1015] type 00 class 0x020000 Apr 30 12:42:12.203267 kernel: pci 0001:01:00.0: reg 0x10: [mem 0x380002000000-0x380003ffffff 64bit pref] Apr 30 12:42:12.203332 kernel: pci 0001:01:00.0: reg 0x30: [mem 0x60100000-0x601fffff pref] Apr 30 12:42:12.203397 kernel: pci 0001:01:00.0: PME# supported from D3cold Apr 30 12:42:12.203465 kernel: pci 0001:01:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Apr 30 12:42:12.203537 kernel: pci 0001:01:00.1: [15b3:1015] type 00 class 0x020000 Apr 30 12:42:12.203606 kernel: pci 0001:01:00.1: reg 0x10: [mem 0x380000000000-0x380001ffffff 64bit pref] Apr 30 12:42:12.203671 kernel: pci 0001:01:00.1: reg 0x30: [mem 0x60000000-0x600fffff pref] Apr 30 12:42:12.203736 kernel: pci 0001:01:00.1: PME# supported from D3cold Apr 30 12:42:12.203746 kernel: acpiphp: Slot [2-6] registered Apr 30 12:42:12.203754 kernel: acpiphp: Slot [3-4] registered Apr 30 12:42:12.203762 kernel: acpiphp: Slot [4-4] registered Apr 30 12:42:12.203820 kernel: pci_bus 0001:00: on NUMA node 0 Apr 30 12:42:12.203884 kernel: pci 0001:00:01.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Apr 30 12:42:12.204019 kernel: pci 0001:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Apr 30 12:42:12.204098 kernel: pci 0001:00:02.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.204162 kernel: pci 0001:00:02.0: bridge window [mem 0x00100000-0x000fffff] to [bus 02] add_size 200000 add_align 100000 Apr 30 12:42:12.204226 kernel: pci 0001:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.204289 kernel: pci 0001:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.204358 kernel: pci 0001:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.204423 kernel: pci 0001:00:04.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.204486 kernel: pci 0001:00:04.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.204549 kernel: pci 0001:00:04.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.204616 kernel: pci 0001:00:01.0: BAR 15: assigned [mem 0x380000000000-0x380003ffffff 64bit pref] Apr 30 12:42:12.204679 kernel: pci 0001:00:01.0: BAR 14: assigned [mem 0x60000000-0x601fffff] Apr 30 12:42:12.204742 kernel: pci 0001:00:02.0: BAR 14: assigned [mem 0x60200000-0x603fffff] Apr 30 12:42:12.204808 kernel: pci 0001:00:02.0: BAR 15: assigned [mem 0x380004000000-0x3800041fffff 64bit pref] Apr 30 12:42:12.204870 kernel: pci 0001:00:03.0: BAR 14: assigned [mem 0x60400000-0x605fffff] Apr 30 12:42:12.204934 kernel: pci 0001:00:03.0: BAR 15: assigned [mem 0x380004200000-0x3800043fffff 64bit pref] Apr 30 12:42:12.205000 kernel: pci 0001:00:04.0: BAR 14: assigned [mem 0x60600000-0x607fffff] Apr 30 12:42:12.205063 kernel: pci 0001:00:04.0: BAR 15: assigned [mem 0x380004400000-0x3800045fffff 64bit pref] Apr 30 12:42:12.205125 kernel: pci 0001:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205188 kernel: pci 0001:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.205251 kernel: pci 0001:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205316 kernel: pci 0001:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.205379 kernel: pci 0001:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205441 kernel: pci 0001:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.205504 kernel: pci 0001:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205566 kernel: pci 0001:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.205629 kernel: pci 0001:00:04.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205691 kernel: pci 0001:00:04.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.205753 kernel: pci 0001:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205819 kernel: pci 0001:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.205881 kernel: pci 0001:00:02.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.205944 kernel: pci 0001:00:02.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.206013 kernel: pci 0001:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.206076 kernel: pci 0001:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.206143 kernel: pci 0001:01:00.0: BAR 0: assigned [mem 0x380000000000-0x380001ffffff 64bit pref] Apr 30 12:42:12.206210 kernel: pci 0001:01:00.1: BAR 0: assigned [mem 0x380002000000-0x380003ffffff 64bit pref] Apr 30 12:42:12.206275 kernel: pci 0001:01:00.0: BAR 6: assigned [mem 0x60000000-0x600fffff pref] Apr 30 12:42:12.206342 kernel: pci 0001:01:00.1: BAR 6: assigned [mem 0x60100000-0x601fffff pref] Apr 30 12:42:12.206406 kernel: pci 0001:00:01.0: PCI bridge to [bus 01] Apr 30 12:42:12.206468 kernel: pci 0001:00:01.0: bridge window [mem 0x60000000-0x601fffff] Apr 30 12:42:12.206532 kernel: pci 0001:00:01.0: bridge window [mem 0x380000000000-0x380003ffffff 64bit pref] Apr 30 12:42:12.206594 kernel: pci 0001:00:02.0: PCI bridge to [bus 02] Apr 30 12:42:12.206657 kernel: pci 0001:00:02.0: bridge window [mem 0x60200000-0x603fffff] Apr 30 12:42:12.206720 kernel: pci 0001:00:02.0: bridge window [mem 0x380004000000-0x3800041fffff 64bit pref] Apr 30 12:42:12.206785 kernel: pci 0001:00:03.0: PCI bridge to [bus 03] Apr 30 12:42:12.206849 kernel: pci 0001:00:03.0: bridge window [mem 0x60400000-0x605fffff] Apr 30 12:42:12.206912 kernel: pci 0001:00:03.0: bridge window [mem 0x380004200000-0x3800043fffff 64bit pref] Apr 30 12:42:12.206979 kernel: pci 0001:00:04.0: PCI bridge to [bus 04] Apr 30 12:42:12.207042 kernel: pci 0001:00:04.0: bridge window [mem 0x60600000-0x607fffff] Apr 30 12:42:12.207105 kernel: pci 0001:00:04.0: bridge window [mem 0x380004400000-0x3800045fffff 64bit pref] Apr 30 12:42:12.207165 kernel: pci_bus 0001:00: resource 4 [mem 0x60000000-0x6fffffff window] Apr 30 12:42:12.207223 kernel: pci_bus 0001:00: resource 5 [mem 0x380000000000-0x3bffdfffffff window] Apr 30 12:42:12.207299 kernel: pci_bus 0001:01: resource 1 [mem 0x60000000-0x601fffff] Apr 30 12:42:12.207359 kernel: pci_bus 0001:01: resource 2 [mem 0x380000000000-0x380003ffffff 64bit pref] Apr 30 12:42:12.207426 kernel: pci_bus 0001:02: resource 1 [mem 0x60200000-0x603fffff] Apr 30 12:42:12.207485 kernel: pci_bus 0001:02: resource 2 [mem 0x380004000000-0x3800041fffff 64bit pref] Apr 30 12:42:12.207553 kernel: pci_bus 0001:03: resource 1 [mem 0x60400000-0x605fffff] Apr 30 12:42:12.207612 kernel: pci_bus 0001:03: resource 2 [mem 0x380004200000-0x3800043fffff 64bit pref] Apr 30 12:42:12.207679 kernel: pci_bus 0001:04: resource 1 [mem 0x60600000-0x607fffff] Apr 30 12:42:12.207738 kernel: pci_bus 0001:04: resource 2 [mem 0x380004400000-0x3800045fffff 64bit pref] Apr 30 12:42:12.207748 kernel: ACPI: PCI Root Bridge [PCI6] (domain 0004 [bus 00-ff]) Apr 30 12:42:12.207818 kernel: acpi PNP0A08:07: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 30 12:42:12.207883 kernel: acpi PNP0A08:07: _OSC: platform does not support [PCIeHotplug PME LTR] Apr 30 12:42:12.207943 kernel: acpi PNP0A08:07: _OSC: OS now controls [AER PCIeCapability] Apr 30 12:42:12.208194 kernel: acpi PNP0A08:07: MCFG quirk: ECAM at [mem 0x2bfff0000000-0x2bffffffffff] for [bus 00-ff] with pci_32b_read_ops Apr 30 12:42:12.208256 kernel: acpi PNP0A08:07: ECAM area [mem 0x2bfff0000000-0x2bffffffffff] reserved by PNP0C02:00 Apr 30 12:42:12.208317 kernel: acpi PNP0A08:07: ECAM at [mem 0x2bfff0000000-0x2bffffffffff] for [bus 00-ff] Apr 30 12:42:12.208327 kernel: PCI host bridge to bus 0004:00 Apr 30 12:42:12.208392 kernel: pci_bus 0004:00: root bus resource [mem 0x20000000-0x2fffffff window] Apr 30 12:42:12.208453 kernel: pci_bus 0004:00: root bus resource [mem 0x280000000000-0x2bffdfffffff window] Apr 30 12:42:12.208509 kernel: pci_bus 0004:00: root bus resource [bus 00-ff] Apr 30 12:42:12.208581 kernel: pci 0004:00:00.0: [1def:e110] type 00 class 0x060000 Apr 30 12:42:12.208653 kernel: pci 0004:00:01.0: [1def:e111] type 01 class 0x060400 Apr 30 12:42:12.208717 kernel: pci 0004:00:01.0: supports D1 D2 Apr 30 12:42:12.208780 kernel: pci 0004:00:01.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.208850 kernel: pci 0004:00:03.0: [1def:e113] type 01 class 0x060400 Apr 30 12:42:12.208919 kernel: pci 0004:00:03.0: supports D1 D2 Apr 30 12:42:12.208987 kernel: pci 0004:00:03.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.209058 kernel: pci 0004:00:05.0: [1def:e115] type 01 class 0x060400 Apr 30 12:42:12.209122 kernel: pci 0004:00:05.0: supports D1 D2 Apr 30 12:42:12.209185 kernel: pci 0004:00:05.0: PME# supported from D0 D1 D3hot Apr 30 12:42:12.209258 kernel: pci 0004:01:00.0: [1a03:1150] type 01 class 0x060400 Apr 30 12:42:12.209324 kernel: pci 0004:01:00.0: enabling Extended Tags Apr 30 12:42:12.209392 kernel: pci 0004:01:00.0: supports D1 D2 Apr 30 12:42:12.209456 kernel: pci 0004:01:00.0: PME# supported from D0 D1 D2 D3hot D3cold Apr 30 12:42:12.209534 kernel: pci_bus 0004:02: extended config space not accessible Apr 30 12:42:12.209611 kernel: pci 0004:02:00.0: [1a03:2000] type 00 class 0x030000 Apr 30 12:42:12.209679 kernel: pci 0004:02:00.0: reg 0x10: [mem 0x20000000-0x21ffffff] Apr 30 12:42:12.209747 kernel: pci 0004:02:00.0: reg 0x14: [mem 0x22000000-0x2201ffff] Apr 30 12:42:12.209813 kernel: pci 0004:02:00.0: reg 0x18: [io 0x0000-0x007f] Apr 30 12:42:12.209883 kernel: pci 0004:02:00.0: BAR 0: assigned to efifb Apr 30 12:42:12.209953 kernel: pci 0004:02:00.0: supports D1 D2 Apr 30 12:42:12.210021 kernel: pci 0004:02:00.0: PME# supported from D0 D1 D2 D3hot D3cold Apr 30 12:42:12.210097 kernel: pci 0004:03:00.0: [1912:0014] type 00 class 0x0c0330 Apr 30 12:42:12.210163 kernel: pci 0004:03:00.0: reg 0x10: [mem 0x22200000-0x22201fff 64bit] Apr 30 12:42:12.210228 kernel: pci 0004:03:00.0: PME# supported from D0 D3hot D3cold Apr 30 12:42:12.210287 kernel: pci_bus 0004:00: on NUMA node 0 Apr 30 12:42:12.210353 kernel: pci 0004:00:01.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 01-02] add_size 200000 add_align 100000 Apr 30 12:42:12.210418 kernel: pci 0004:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Apr 30 12:42:12.210481 kernel: pci 0004:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 03] add_size 200000 add_align 100000 Apr 30 12:42:12.210545 kernel: pci 0004:00:03.0: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Apr 30 12:42:12.210609 kernel: pci 0004:00:05.0: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Apr 30 12:42:12.210673 kernel: pci 0004:00:05.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.210736 kernel: pci 0004:00:05.0: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Apr 30 12:42:12.210802 kernel: pci 0004:00:01.0: BAR 14: assigned [mem 0x20000000-0x22ffffff] Apr 30 12:42:12.210866 kernel: pci 0004:00:01.0: BAR 15: assigned [mem 0x280000000000-0x2800001fffff 64bit pref] Apr 30 12:42:12.210929 kernel: pci 0004:00:03.0: BAR 14: assigned [mem 0x23000000-0x231fffff] Apr 30 12:42:12.210996 kernel: pci 0004:00:03.0: BAR 15: assigned [mem 0x280000200000-0x2800003fffff 64bit pref] Apr 30 12:42:12.211059 kernel: pci 0004:00:05.0: BAR 14: assigned [mem 0x23200000-0x233fffff] Apr 30 12:42:12.211122 kernel: pci 0004:00:05.0: BAR 15: assigned [mem 0x280000400000-0x2800005fffff 64bit pref] Apr 30 12:42:12.211185 kernel: pci 0004:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.211248 kernel: pci 0004:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.211314 kernel: pci 0004:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.211377 kernel: pci 0004:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.211441 kernel: pci 0004:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.211503 kernel: pci 0004:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.211567 kernel: pci 0004:00:01.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.211629 kernel: pci 0004:00:01.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.211694 kernel: pci 0004:00:05.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.211757 kernel: pci 0004:00:05.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.211822 kernel: pci 0004:00:03.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.211885 kernel: pci 0004:00:03.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.211954 kernel: pci 0004:01:00.0: BAR 14: assigned [mem 0x20000000-0x22ffffff] Apr 30 12:42:12.212020 kernel: pci 0004:01:00.0: BAR 13: no space for [io size 0x1000] Apr 30 12:42:12.212085 kernel: pci 0004:01:00.0: BAR 13: failed to assign [io size 0x1000] Apr 30 12:42:12.212154 kernel: pci 0004:02:00.0: BAR 0: assigned [mem 0x20000000-0x21ffffff] Apr 30 12:42:12.212222 kernel: pci 0004:02:00.0: BAR 1: assigned [mem 0x22000000-0x2201ffff] Apr 30 12:42:12.212290 kernel: pci 0004:02:00.0: BAR 2: no space for [io size 0x0080] Apr 30 12:42:12.212360 kernel: pci 0004:02:00.0: BAR 2: failed to assign [io size 0x0080] Apr 30 12:42:12.212426 kernel: pci 0004:01:00.0: PCI bridge to [bus 02] Apr 30 12:42:12.212491 kernel: pci 0004:01:00.0: bridge window [mem 0x20000000-0x22ffffff] Apr 30 12:42:12.212555 kernel: pci 0004:00:01.0: PCI bridge to [bus 01-02] Apr 30 12:42:12.212619 kernel: pci 0004:00:01.0: bridge window [mem 0x20000000-0x22ffffff] Apr 30 12:42:12.212683 kernel: pci 0004:00:01.0: bridge window [mem 0x280000000000-0x2800001fffff 64bit pref] Apr 30 12:42:12.212749 kernel: pci 0004:03:00.0: BAR 0: assigned [mem 0x23000000-0x23001fff 64bit] Apr 30 12:42:12.212813 kernel: pci 0004:00:03.0: PCI bridge to [bus 03] Apr 30 12:42:12.212879 kernel: pci 0004:00:03.0: bridge window [mem 0x23000000-0x231fffff] Apr 30 12:42:12.212943 kernel: pci 0004:00:03.0: bridge window [mem 0x280000200000-0x2800003fffff 64bit pref] Apr 30 12:42:12.213009 kernel: pci 0004:00:05.0: PCI bridge to [bus 04] Apr 30 12:42:12.213072 kernel: pci 0004:00:05.0: bridge window [mem 0x23200000-0x233fffff] Apr 30 12:42:12.213136 kernel: pci 0004:00:05.0: bridge window [mem 0x280000400000-0x2800005fffff 64bit pref] Apr 30 12:42:12.213194 kernel: pci_bus 0004:00: Some PCI device resources are unassigned, try booting with pci=realloc Apr 30 12:42:12.213253 kernel: pci_bus 0004:00: resource 4 [mem 0x20000000-0x2fffffff window] Apr 30 12:42:12.213309 kernel: pci_bus 0004:00: resource 5 [mem 0x280000000000-0x2bffdfffffff window] Apr 30 12:42:12.213378 kernel: pci_bus 0004:01: resource 1 [mem 0x20000000-0x22ffffff] Apr 30 12:42:12.213437 kernel: pci_bus 0004:01: resource 2 [mem 0x280000000000-0x2800001fffff 64bit pref] Apr 30 12:42:12.213500 kernel: pci_bus 0004:02: resource 1 [mem 0x20000000-0x22ffffff] Apr 30 12:42:12.213567 kernel: pci_bus 0004:03: resource 1 [mem 0x23000000-0x231fffff] Apr 30 12:42:12.213626 kernel: pci_bus 0004:03: resource 2 [mem 0x280000200000-0x2800003fffff 64bit pref] Apr 30 12:42:12.213696 kernel: pci_bus 0004:04: resource 1 [mem 0x23200000-0x233fffff] Apr 30 12:42:12.213755 kernel: pci_bus 0004:04: resource 2 [mem 0x280000400000-0x2800005fffff 64bit pref] Apr 30 12:42:12.213765 kernel: iommu: Default domain type: Translated Apr 30 12:42:12.213773 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 30 12:42:12.213781 kernel: efivars: Registered efivars operations Apr 30 12:42:12.213850 kernel: pci 0004:02:00.0: vgaarb: setting as boot VGA device Apr 30 12:42:12.213919 kernel: pci 0004:02:00.0: vgaarb: bridge control possible Apr 30 12:42:12.213989 kernel: pci 0004:02:00.0: vgaarb: VGA device added: decodes=io+mem,owns=none,locks=none Apr 30 12:42:12.214002 kernel: vgaarb: loaded Apr 30 12:42:12.214010 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 30 12:42:12.214018 kernel: VFS: Disk quotas dquot_6.6.0 Apr 30 12:42:12.214026 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 30 12:42:12.214034 kernel: pnp: PnP ACPI init Apr 30 12:42:12.214102 kernel: system 00:00: [mem 0x3bfff0000000-0x3bffffffffff window] could not be reserved Apr 30 12:42:12.214162 kernel: system 00:00: [mem 0x3ffff0000000-0x3fffffffffff window] could not be reserved Apr 30 12:42:12.214223 kernel: system 00:00: [mem 0x23fff0000000-0x23ffffffffff window] could not be reserved Apr 30 12:42:12.214280 kernel: system 00:00: [mem 0x27fff0000000-0x27ffffffffff window] could not be reserved Apr 30 12:42:12.214338 kernel: system 00:00: [mem 0x2bfff0000000-0x2bffffffffff window] could not be reserved Apr 30 12:42:12.214396 kernel: system 00:00: [mem 0x2ffff0000000-0x2fffffffffff window] could not be reserved Apr 30 12:42:12.214455 kernel: system 00:00: [mem 0x33fff0000000-0x33ffffffffff window] could not be reserved Apr 30 12:42:12.214513 kernel: system 00:00: [mem 0x37fff0000000-0x37ffffffffff window] could not be reserved Apr 30 12:42:12.214523 kernel: pnp: PnP ACPI: found 1 devices Apr 30 12:42:12.214533 kernel: NET: Registered PF_INET protocol family Apr 30 12:42:12.214541 kernel: IP idents hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.214549 kernel: tcp_listen_portaddr_hash hash table entries: 65536 (order: 8, 1048576 bytes, linear) Apr 30 12:42:12.214557 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 30 12:42:12.214564 kernel: TCP established hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 30 12:42:12.214572 kernel: TCP bind hash table entries: 65536 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.214580 kernel: TCP: Hash tables configured (established 524288 bind 65536) Apr 30 12:42:12.214588 kernel: UDP hash table entries: 65536 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.214597 kernel: UDP-Lite hash table entries: 65536 (order: 9, 2097152 bytes, linear) Apr 30 12:42:12.214605 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 30 12:42:12.214671 kernel: pci 0001:01:00.0: CLS mismatch (64 != 32), using 64 bytes Apr 30 12:42:12.214681 kernel: kvm [1]: IPA Size Limit: 48 bits Apr 30 12:42:12.214689 kernel: kvm [1]: GICv3: no GICV resource entry Apr 30 12:42:12.214697 kernel: kvm [1]: disabling GICv2 emulation Apr 30 12:42:12.214704 kernel: kvm [1]: GIC system register CPU interface enabled Apr 30 12:42:12.214712 kernel: kvm [1]: vgic interrupt IRQ9 Apr 30 12:42:12.214720 kernel: kvm [1]: VHE mode initialized successfully Apr 30 12:42:12.214730 kernel: Initialise system trusted keyrings Apr 30 12:42:12.214737 kernel: workingset: timestamp_bits=39 max_order=26 bucket_order=0 Apr 30 12:42:12.214745 kernel: Key type asymmetric registered Apr 30 12:42:12.214752 kernel: Asymmetric key parser 'x509' registered Apr 30 12:42:12.214760 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Apr 30 12:42:12.214768 kernel: io scheduler mq-deadline registered Apr 30 12:42:12.214775 kernel: io scheduler kyber registered Apr 30 12:42:12.214783 kernel: io scheduler bfq registered Apr 30 12:42:12.214791 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Apr 30 12:42:12.214799 kernel: ACPI: button: Power Button [PWRB] Apr 30 12:42:12.214808 kernel: ACPI GTDT: found 1 SBSA generic Watchdog(s). Apr 30 12:42:12.214816 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 30 12:42:12.214887 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: option mask 0x0 Apr 30 12:42:12.214948 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.215013 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.215073 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.215132 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 131072 entries for evtq Apr 30 12:42:12.215195 kernel: arm-smmu-v3 arm-smmu-v3.0.auto: allocated 262144 entries for priq Apr 30 12:42:12.215262 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: option mask 0x0 Apr 30 12:42:12.215322 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.215381 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.215440 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.215499 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 131072 entries for evtq Apr 30 12:42:12.215561 kernel: arm-smmu-v3 arm-smmu-v3.1.auto: allocated 262144 entries for priq Apr 30 12:42:12.215627 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: option mask 0x0 Apr 30 12:42:12.215687 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.215747 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.215807 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.215866 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 131072 entries for evtq Apr 30 12:42:12.215925 kernel: arm-smmu-v3 arm-smmu-v3.2.auto: allocated 262144 entries for priq Apr 30 12:42:12.215997 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: option mask 0x0 Apr 30 12:42:12.216058 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.216117 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.216176 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.216235 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 131072 entries for evtq Apr 30 12:42:12.216294 kernel: arm-smmu-v3 arm-smmu-v3.3.auto: allocated 262144 entries for priq Apr 30 12:42:12.216366 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: option mask 0x0 Apr 30 12:42:12.216429 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.216487 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.216547 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.216606 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 131072 entries for evtq Apr 30 12:42:12.216664 kernel: arm-smmu-v3 arm-smmu-v3.4.auto: allocated 262144 entries for priq Apr 30 12:42:12.216732 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: option mask 0x0 Apr 30 12:42:12.216793 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.216853 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.216911 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.216975 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 131072 entries for evtq Apr 30 12:42:12.217034 kernel: arm-smmu-v3 arm-smmu-v3.5.auto: allocated 262144 entries for priq Apr 30 12:42:12.217101 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: option mask 0x0 Apr 30 12:42:12.217163 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.217222 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.217282 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.217341 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 131072 entries for evtq Apr 30 12:42:12.217400 kernel: arm-smmu-v3 arm-smmu-v3.6.auto: allocated 262144 entries for priq Apr 30 12:42:12.217466 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: option mask 0x0 Apr 30 12:42:12.217529 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: IDR0.COHACC overridden by FW configuration (false) Apr 30 12:42:12.217589 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: ias 48-bit, oas 48-bit (features 0x000c1eff) Apr 30 12:42:12.217647 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 262144 entries for cmdq Apr 30 12:42:12.217709 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 131072 entries for evtq Apr 30 12:42:12.217767 kernel: arm-smmu-v3 arm-smmu-v3.7.auto: allocated 262144 entries for priq Apr 30 12:42:12.217777 kernel: thunder_xcv, ver 1.0 Apr 30 12:42:12.217785 kernel: thunder_bgx, ver 1.0 Apr 30 12:42:12.217793 kernel: nicpf, ver 1.0 Apr 30 12:42:12.217802 kernel: nicvf, ver 1.0 Apr 30 12:42:12.217868 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 30 12:42:12.217928 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-04-30T12:42:10 UTC (1746016930) Apr 30 12:42:12.217938 kernel: efifb: probing for efifb Apr 30 12:42:12.217946 kernel: efifb: framebuffer at 0x20000000, using 1876k, total 1875k Apr 30 12:42:12.217957 kernel: efifb: mode is 800x600x32, linelength=3200, pages=1 Apr 30 12:42:12.217965 kernel: efifb: scrolling: redraw Apr 30 12:42:12.217972 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 30 12:42:12.217982 kernel: Console: switching to colour frame buffer device 100x37 Apr 30 12:42:12.217990 kernel: fb0: EFI VGA frame buffer device Apr 30 12:42:12.217998 kernel: SMCCC: SOC_ID: ID = jep106:0a16:0001 Revision = 0x000000a1 Apr 30 12:42:12.218005 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 30 12:42:12.218013 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Apr 30 12:42:12.218021 kernel: watchdog: Delayed init of the lockup detector failed: -19 Apr 30 12:42:12.218029 kernel: watchdog: Hard watchdog permanently disabled Apr 30 12:42:12.218037 kernel: NET: Registered PF_INET6 protocol family Apr 30 12:42:12.218044 kernel: Segment Routing with IPv6 Apr 30 12:42:12.218053 kernel: In-situ OAM (IOAM) with IPv6 Apr 30 12:42:12.218061 kernel: NET: Registered PF_PACKET protocol family Apr 30 12:42:12.218069 kernel: Key type dns_resolver registered Apr 30 12:42:12.218076 kernel: registered taskstats version 1 Apr 30 12:42:12.218084 kernel: Loading compiled-in X.509 certificates Apr 30 12:42:12.218092 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.88-flatcar: 4e3d8be893bce81adbd52ab54fa98214a1a14a2e' Apr 30 12:42:12.218099 kernel: Key type .fscrypt registered Apr 30 12:42:12.218107 kernel: Key type fscrypt-provisioning registered Apr 30 12:42:12.218116 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 30 12:42:12.218125 kernel: ima: Allocated hash algorithm: sha1 Apr 30 12:42:12.218133 kernel: ima: No architecture policies found Apr 30 12:42:12.218140 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 30 12:42:12.218208 kernel: pcieport 000d:00:01.0: Adding to iommu group 0 Apr 30 12:42:12.218273 kernel: pcieport 000d:00:01.0: AER: enabled with IRQ 91 Apr 30 12:42:12.218338 kernel: pcieport 000d:00:02.0: Adding to iommu group 1 Apr 30 12:42:12.218402 kernel: pcieport 000d:00:02.0: AER: enabled with IRQ 91 Apr 30 12:42:12.218467 kernel: pcieport 000d:00:03.0: Adding to iommu group 2 Apr 30 12:42:12.218531 kernel: pcieport 000d:00:03.0: AER: enabled with IRQ 91 Apr 30 12:42:12.218598 kernel: pcieport 000d:00:04.0: Adding to iommu group 3 Apr 30 12:42:12.218662 kernel: pcieport 000d:00:04.0: AER: enabled with IRQ 91 Apr 30 12:42:12.218727 kernel: pcieport 0000:00:01.0: Adding to iommu group 4 Apr 30 12:42:12.218792 kernel: pcieport 0000:00:01.0: AER: enabled with IRQ 92 Apr 30 12:42:12.218858 kernel: pcieport 0000:00:02.0: Adding to iommu group 5 Apr 30 12:42:12.218921 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 92 Apr 30 12:42:12.218989 kernel: pcieport 0000:00:03.0: Adding to iommu group 6 Apr 30 12:42:12.219052 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 92 Apr 30 12:42:12.219120 kernel: pcieport 0000:00:04.0: Adding to iommu group 7 Apr 30 12:42:12.219184 kernel: pcieport 0000:00:04.0: AER: enabled with IRQ 92 Apr 30 12:42:12.219248 kernel: pcieport 0005:00:01.0: Adding to iommu group 8 Apr 30 12:42:12.219312 kernel: pcieport 0005:00:01.0: AER: enabled with IRQ 93 Apr 30 12:42:12.219375 kernel: pcieport 0005:00:03.0: Adding to iommu group 9 Apr 30 12:42:12.219440 kernel: pcieport 0005:00:03.0: AER: enabled with IRQ 93 Apr 30 12:42:12.219503 kernel: pcieport 0005:00:05.0: Adding to iommu group 10 Apr 30 12:42:12.219567 kernel: pcieport 0005:00:05.0: AER: enabled with IRQ 93 Apr 30 12:42:12.219634 kernel: pcieport 0005:00:07.0: Adding to iommu group 11 Apr 30 12:42:12.219698 kernel: pcieport 0005:00:07.0: AER: enabled with IRQ 93 Apr 30 12:42:12.219763 kernel: pcieport 0003:00:01.0: Adding to iommu group 12 Apr 30 12:42:12.219827 kernel: pcieport 0003:00:01.0: AER: enabled with IRQ 94 Apr 30 12:42:12.219892 kernel: pcieport 0003:00:03.0: Adding to iommu group 13 Apr 30 12:42:12.219958 kernel: pcieport 0003:00:03.0: AER: enabled with IRQ 94 Apr 30 12:42:12.220024 kernel: pcieport 0003:00:05.0: Adding to iommu group 14 Apr 30 12:42:12.220087 kernel: pcieport 0003:00:05.0: AER: enabled with IRQ 94 Apr 30 12:42:12.220152 kernel: pcieport 000c:00:01.0: Adding to iommu group 15 Apr 30 12:42:12.220219 kernel: pcieport 000c:00:01.0: AER: enabled with IRQ 95 Apr 30 12:42:12.220284 kernel: pcieport 000c:00:02.0: Adding to iommu group 16 Apr 30 12:42:12.220349 kernel: pcieport 000c:00:02.0: AER: enabled with IRQ 95 Apr 30 12:42:12.220412 kernel: pcieport 000c:00:03.0: Adding to iommu group 17 Apr 30 12:42:12.220477 kernel: pcieport 000c:00:03.0: AER: enabled with IRQ 95 Apr 30 12:42:12.220542 kernel: pcieport 000c:00:04.0: Adding to iommu group 18 Apr 30 12:42:12.220607 kernel: pcieport 000c:00:04.0: AER: enabled with IRQ 95 Apr 30 12:42:12.220672 kernel: pcieport 0002:00:01.0: Adding to iommu group 19 Apr 30 12:42:12.220739 kernel: pcieport 0002:00:01.0: AER: enabled with IRQ 96 Apr 30 12:42:12.220804 kernel: pcieport 0002:00:03.0: Adding to iommu group 20 Apr 30 12:42:12.220870 kernel: pcieport 0002:00:03.0: AER: enabled with IRQ 96 Apr 30 12:42:12.220937 kernel: pcieport 0002:00:05.0: Adding to iommu group 21 Apr 30 12:42:12.221004 kernel: pcieport 0002:00:05.0: AER: enabled with IRQ 96 Apr 30 12:42:12.221071 kernel: pcieport 0002:00:07.0: Adding to iommu group 22 Apr 30 12:42:12.221135 kernel: pcieport 0002:00:07.0: AER: enabled with IRQ 96 Apr 30 12:42:12.221202 kernel: pcieport 0001:00:01.0: Adding to iommu group 23 Apr 30 12:42:12.221268 kernel: pcieport 0001:00:01.0: AER: enabled with IRQ 97 Apr 30 12:42:12.221336 kernel: pcieport 0001:00:02.0: Adding to iommu group 24 Apr 30 12:42:12.221402 kernel: pcieport 0001:00:02.0: AER: enabled with IRQ 97 Apr 30 12:42:12.221467 kernel: pcieport 0001:00:03.0: Adding to iommu group 25 Apr 30 12:42:12.221533 kernel: pcieport 0001:00:03.0: AER: enabled with IRQ 97 Apr 30 12:42:12.221599 kernel: pcieport 0001:00:04.0: Adding to iommu group 26 Apr 30 12:42:12.221665 kernel: pcieport 0001:00:04.0: AER: enabled with IRQ 97 Apr 30 12:42:12.221730 kernel: pcieport 0004:00:01.0: Adding to iommu group 27 Apr 30 12:42:12.221798 kernel: pcieport 0004:00:01.0: AER: enabled with IRQ 98 Apr 30 12:42:12.221864 kernel: pcieport 0004:00:03.0: Adding to iommu group 28 Apr 30 12:42:12.221931 kernel: pcieport 0004:00:03.0: AER: enabled with IRQ 98 Apr 30 12:42:12.222000 kernel: pcieport 0004:00:05.0: Adding to iommu group 29 Apr 30 12:42:12.222065 kernel: pcieport 0004:00:05.0: AER: enabled with IRQ 98 Apr 30 12:42:12.222133 kernel: pcieport 0004:01:00.0: Adding to iommu group 30 Apr 30 12:42:12.222144 kernel: clk: Disabling unused clocks Apr 30 12:42:12.222152 kernel: Freeing unused kernel memory: 38336K Apr 30 12:42:12.222162 kernel: Run /init as init process Apr 30 12:42:12.222169 kernel: with arguments: Apr 30 12:42:12.222177 kernel: /init Apr 30 12:42:12.222184 kernel: with environment: Apr 30 12:42:12.222192 kernel: HOME=/ Apr 30 12:42:12.222199 kernel: TERM=linux Apr 30 12:42:12.222207 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Apr 30 12:42:12.222215 systemd[1]: Successfully made /usr/ read-only. Apr 30 12:42:12.222226 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 30 12:42:12.222236 systemd[1]: Detected architecture arm64. Apr 30 12:42:12.222244 systemd[1]: Running in initrd. Apr 30 12:42:12.222252 systemd[1]: No hostname configured, using default hostname. Apr 30 12:42:12.222260 systemd[1]: Hostname set to . Apr 30 12:42:12.222268 systemd[1]: Initializing machine ID from random generator. Apr 30 12:42:12.222276 systemd[1]: Queued start job for default target initrd.target. Apr 30 12:42:12.222284 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 12:42:12.222294 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 12:42:12.222303 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Apr 30 12:42:12.222311 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 12:42:12.222319 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Apr 30 12:42:12.222328 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Apr 30 12:42:12.222337 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Apr 30 12:42:12.222346 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Apr 30 12:42:12.222355 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 12:42:12.222364 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 12:42:12.222372 systemd[1]: Reached target paths.target - Path Units. Apr 30 12:42:12.222380 systemd[1]: Reached target slices.target - Slice Units. Apr 30 12:42:12.222388 systemd[1]: Reached target swap.target - Swaps. Apr 30 12:42:12.222396 systemd[1]: Reached target timers.target - Timer Units. Apr 30 12:42:12.222404 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 12:42:12.222412 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 12:42:12.222422 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Apr 30 12:42:12.222430 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Apr 30 12:42:12.222438 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 12:42:12.222447 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 12:42:12.222455 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 12:42:12.222463 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 12:42:12.222471 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Apr 30 12:42:12.222479 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 12:42:12.222487 systemd[1]: Finished network-cleanup.service - Network Cleanup. Apr 30 12:42:12.222497 systemd[1]: Starting systemd-fsck-usr.service... Apr 30 12:42:12.222505 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 12:42:12.222513 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 12:42:12.222544 systemd-journald[899]: Collecting audit messages is disabled. Apr 30 12:42:12.222565 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 12:42:12.222573 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Apr 30 12:42:12.222581 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 30 12:42:12.222589 kernel: Bridge firewalling registered Apr 30 12:42:12.222598 systemd-journald[899]: Journal started Apr 30 12:42:12.222616 systemd-journald[899]: Runtime Journal (/run/log/journal/e9d88d68e07c44fcb1cab6cf4f3f5ea2) is 8M, max 4G, 3.9G free. Apr 30 12:42:12.181472 systemd-modules-load[903]: Inserted module 'overlay' Apr 30 12:42:12.256308 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 12:42:12.203767 systemd-modules-load[903]: Inserted module 'br_netfilter' Apr 30 12:42:12.261984 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 12:42:12.272813 systemd[1]: Finished systemd-fsck-usr.service. Apr 30 12:42:12.283663 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 12:42:12.294448 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 12:42:12.316085 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 12:42:12.333660 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 12:42:12.340271 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Apr 30 12:42:12.351839 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 12:42:12.368466 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 12:42:12.384346 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 12:42:12.401044 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Apr 30 12:42:12.412431 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 12:42:12.444108 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Apr 30 12:42:12.457321 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 12:42:12.465760 dracut-cmdline[943]: dracut-dracut-053 Apr 30 12:42:12.476879 dracut-cmdline[943]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty0 console=ttyS1,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=packet flatcar.autologin verity.usrhash=984055eb0c340c9cf0fb51b368030ed72e75b7f2e065edc13766888ef0b42074 Apr 30 12:42:12.471044 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 12:42:12.484452 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 12:42:12.496938 systemd-resolved[949]: Positive Trust Anchors: Apr 30 12:42:12.496947 systemd-resolved[949]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 12:42:12.496982 systemd-resolved[949]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 12:42:12.511815 systemd-resolved[949]: Defaulting to hostname 'linux'. Apr 30 12:42:12.522384 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 12:42:12.541758 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 12:42:12.655744 kernel: SCSI subsystem initialized Apr 30 12:42:12.655771 kernel: Loading iSCSI transport class v2.0-870. Apr 30 12:42:12.674957 kernel: iscsi: registered transport (tcp) Apr 30 12:42:12.702392 kernel: iscsi: registered transport (qla4xxx) Apr 30 12:42:12.702417 kernel: QLogic iSCSI HBA Driver Apr 30 12:42:12.745913 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Apr 30 12:42:12.768111 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Apr 30 12:42:12.812920 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 30 12:42:12.812962 kernel: device-mapper: uevent: version 1.0.3 Apr 30 12:42:12.822590 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Apr 30 12:42:12.887961 kernel: raid6: neonx8 gen() 15839 MB/s Apr 30 12:42:12.912959 kernel: raid6: neonx4 gen() 15875 MB/s Apr 30 12:42:12.937959 kernel: raid6: neonx2 gen() 13255 MB/s Apr 30 12:42:12.962959 kernel: raid6: neonx1 gen() 10570 MB/s Apr 30 12:42:12.987960 kernel: raid6: int64x8 gen() 6818 MB/s Apr 30 12:42:13.012959 kernel: raid6: int64x4 gen() 7384 MB/s Apr 30 12:42:13.037959 kernel: raid6: int64x2 gen() 6134 MB/s Apr 30 12:42:13.065941 kernel: raid6: int64x1 gen() 5077 MB/s Apr 30 12:42:13.065976 kernel: raid6: using algorithm neonx4 gen() 15875 MB/s Apr 30 12:42:13.100362 kernel: raid6: .... xor() 12394 MB/s, rmw enabled Apr 30 12:42:13.100385 kernel: raid6: using neon recovery algorithm Apr 30 12:42:13.123412 kernel: xor: measuring software checksum speed Apr 30 12:42:13.123435 kernel: 8regs : 21624 MB/sec Apr 30 12:42:13.131342 kernel: 32regs : 21670 MB/sec Apr 30 12:42:13.139103 kernel: arm64_neon : 28109 MB/sec Apr 30 12:42:13.146743 kernel: xor: using function: arm64_neon (28109 MB/sec) Apr 30 12:42:13.207959 kernel: Btrfs loaded, zoned=no, fsverity=no Apr 30 12:42:13.217239 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Apr 30 12:42:13.234148 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 12:42:13.247669 systemd-udevd[1137]: Using default interface naming scheme 'v255'. Apr 30 12:42:13.251246 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 12:42:13.270103 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Apr 30 12:42:13.284187 dracut-pre-trigger[1148]: rd.md=0: removing MD RAID activation Apr 30 12:42:13.309904 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 12:42:13.329064 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 12:42:13.438413 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 12:42:13.458957 kernel: pps_core: LinuxPPS API ver. 1 registered Apr 30 12:42:13.458975 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Apr 30 12:42:13.484954 kernel: PTP clock support registered Apr 30 12:42:13.498778 kernel: ACPI: bus type USB registered Apr 30 12:42:13.498807 kernel: usbcore: registered new interface driver usbfs Apr 30 12:42:13.508549 kernel: usbcore: registered new interface driver hub Apr 30 12:42:13.518220 kernel: usbcore: registered new device driver usb Apr 30 12:42:13.526074 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Apr 30 12:42:13.566148 kernel: igb: Intel(R) Gigabit Ethernet Network Driver Apr 30 12:42:13.566167 kernel: igb: Copyright (c) 2007-2014 Intel Corporation. Apr 30 12:42:13.566177 kernel: igb 0003:03:00.0: Adding to iommu group 31 Apr 30 12:42:13.748153 kernel: xhci_hcd 0004:03:00.0: Adding to iommu group 32 Apr 30 12:42:14.059576 kernel: nvme 0005:03:00.0: Adding to iommu group 33 Apr 30 12:42:14.265102 kernel: xhci_hcd 0004:03:00.0: xHCI Host Controller Apr 30 12:42:14.265271 kernel: xhci_hcd 0004:03:00.0: new USB bus registered, assigned bus number 1 Apr 30 12:42:14.265356 kernel: mlx5_core 0001:01:00.0: Adding to iommu group 34 Apr 30 12:42:14.704488 kernel: xhci_hcd 0004:03:00.0: Zeroing 64bit base registers, expecting fault Apr 30 12:42:14.704676 kernel: nvme 0005:04:00.0: Adding to iommu group 35 Apr 30 12:42:14.704767 kernel: igb 0003:03:00.0: added PHC on eth0 Apr 30 12:42:14.704852 kernel: igb 0003:03:00.0: Intel(R) Gigabit Ethernet Network Connection Apr 30 12:42:14.704929 kernel: igb 0003:03:00.0: eth0: (PCIe:5.0Gb/s:Width x2) 18:c0:4d:0c:6f:98 Apr 30 12:42:14.705027 kernel: igb 0003:03:00.0: eth0: PBA No: 106300-000 Apr 30 12:42:14.705104 kernel: igb 0003:03:00.0: Using MSI-X interrupts. 8 rx queue(s), 8 tx queue(s) Apr 30 12:42:14.705181 kernel: igb 0003:03:00.1: Adding to iommu group 36 Apr 30 12:42:14.705264 kernel: xhci_hcd 0004:03:00.0: hcc params 0x014051cf hci version 0x100 quirks 0x0000001100000010 Apr 30 12:42:14.705345 kernel: xhci_hcd 0004:03:00.0: xHCI Host Controller Apr 30 12:42:14.705424 kernel: xhci_hcd 0004:03:00.0: new USB bus registered, assigned bus number 2 Apr 30 12:42:14.705500 kernel: xhci_hcd 0004:03:00.0: Host supports USB 3.0 SuperSpeed Apr 30 12:42:14.705576 kernel: nvme nvme0: pci function 0005:03:00.0 Apr 30 12:42:14.705668 kernel: hub 1-0:1.0: USB hub found Apr 30 12:42:14.705768 kernel: hub 1-0:1.0: 4 ports detected Apr 30 12:42:14.705855 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Apr 30 12:42:14.705991 kernel: nvme nvme0: Shutdown timeout set to 8 seconds Apr 30 12:42:14.706071 kernel: hub 2-0:1.0: USB hub found Apr 30 12:42:14.706167 kernel: mlx5_core 0001:01:00.0: firmware version: 14.31.1014 Apr 30 12:42:14.706248 kernel: hub 2-0:1.0: 4 ports detected Apr 30 12:42:14.706332 kernel: mlx5_core 0001:01:00.0: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Apr 30 12:42:14.706412 kernel: nvme nvme1: pci function 0005:04:00.0 Apr 30 12:42:14.706497 kernel: nvme nvme1: Shutdown timeout set to 8 seconds Apr 30 12:42:14.706569 kernel: nvme nvme0: 32/0/0 default/read/poll queues Apr 30 12:42:14.706641 kernel: nvme nvme1: 32/0/0 default/read/poll queues Apr 30 12:42:14.706713 kernel: igb 0003:03:00.1: added PHC on eth1 Apr 30 12:42:14.706793 kernel: igb 0003:03:00.1: Intel(R) Gigabit Ethernet Network Connection Apr 30 12:42:14.706870 kernel: igb 0003:03:00.1: eth1: (PCIe:5.0Gb/s:Width x2) 18:c0:4d:0c:6f:99 Apr 30 12:42:14.706947 kernel: igb 0003:03:00.1: eth1: PBA No: 106300-000 Apr 30 12:42:14.707031 kernel: igb 0003:03:00.1: Using MSI-X interrupts. 8 rx queue(s), 8 tx queue(s) Apr 30 12:42:14.707109 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 30 12:42:14.707121 kernel: GPT:9289727 != 1875385007 Apr 30 12:42:14.707130 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 30 12:42:14.707140 kernel: GPT:9289727 != 1875385007 Apr 30 12:42:14.707148 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 30 12:42:14.707158 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 12:42:14.707167 kernel: igb 0003:03:00.0 eno1: renamed from eth0 Apr 30 12:42:14.707251 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/nvme0n1p6 scanned by (udev-worker) (1195) Apr 30 12:42:14.707261 kernel: BTRFS: device fsid 8f86a166-b3d6-49f7-a49d-597eaeb9f5e5 devid 1 transid 37 /dev/nvme0n1p3 scanned by (udev-worker) (1193) Apr 30 12:42:14.707271 kernel: igb 0003:03:00.1 eno2: renamed from eth1 Apr 30 12:42:14.707347 kernel: usb 1-3: new high-speed USB device number 2 using xhci_hcd Apr 30 12:42:14.707476 kernel: mlx5_core 0001:01:00.0: Port module event: module 0, Cable plugged Apr 30 12:42:14.707558 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 12:42:14.707568 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 12:42:14.707577 kernel: hub 1-3:1.0: USB hub found Apr 30 12:42:14.707674 kernel: hub 1-3:1.0: 4 ports detected Apr 30 12:42:14.707761 kernel: usb 2-3: new SuperSpeed USB device number 2 using xhci_hcd Apr 30 12:42:14.707885 kernel: hub 2-3:1.0: USB hub found Apr 30 12:42:14.707983 kernel: hub 2-3:1.0: 4 ports detected Apr 30 12:42:14.708070 kernel: mlx5_core 0001:01:00.0: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Apr 30 12:42:14.708152 kernel: mlx5_core 0001:01:00.1: Adding to iommu group 37 Apr 30 12:42:15.389224 kernel: mlx5_core 0001:01:00.1: firmware version: 14.31.1014 Apr 30 12:42:15.389405 kernel: mlx5_core 0001:01:00.1: 31.504 Gb/s available PCIe bandwidth, limited by 8.0 GT/s PCIe x4 link at 0001:00:01.0 (capable of 63.008 Gb/s with 8.0 GT/s PCIe x8 link) Apr 30 12:42:15.389516 kernel: mlx5_core 0001:01:00.1: Port module event: module 1, Cable plugged Apr 30 12:42:15.389593 kernel: mlx5_core 0001:01:00.1: MLX5E: StrdRq(0) RqSz(1024) StrdSz(256) RxCqeCmprss(0 basic) Apr 30 12:42:13.536737 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 12:42:15.405258 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: renamed from eth1 Apr 30 12:42:13.536797 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 12:42:15.434828 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: renamed from eth0 Apr 30 12:42:15.434969 kernel: nvme0n1: p1 p2 p3 p4 p6 p7 p9 Apr 30 12:42:13.657291 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 12:42:13.662751 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 12:42:13.662799 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 12:42:13.668374 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 12:42:15.462627 disk-uuid[1290]: Primary Header is updated. Apr 30 12:42:15.462627 disk-uuid[1290]: Secondary Entries is updated. Apr 30 12:42:15.462627 disk-uuid[1290]: Secondary Header is updated. Apr 30 12:42:13.684048 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 12:42:15.483751 disk-uuid[1291]: The operation has completed successfully. Apr 30 12:42:13.689830 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Apr 30 12:42:13.695904 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 12:42:13.701254 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 12:42:13.706467 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 12:42:13.711586 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 12:42:13.739062 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Apr 30 12:42:13.744760 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Apr 30 12:42:13.811640 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Apr 30 12:42:14.087946 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 12:42:14.301235 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - SAMSUNG MZ1LB960HAJQ-00007 EFI-SYSTEM. Apr 30 12:42:14.352874 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - SAMSUNG MZ1LB960HAJQ-00007 ROOT. Apr 30 12:42:14.364725 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - SAMSUNG MZ1LB960HAJQ-00007 OEM. Apr 30 12:42:14.375363 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - SAMSUNG MZ1LB960HAJQ-00007 USR-A. Apr 30 12:42:14.390170 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - SAMSUNG MZ1LB960HAJQ-00007 USR-A. Apr 30 12:42:14.408100 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Apr 30 12:42:15.576353 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 30 12:42:15.576435 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Apr 30 12:42:15.625055 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Apr 30 12:42:15.657229 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 30 12:42:15.657316 sh[1476]: Success Apr 30 12:42:15.679792 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Apr 30 12:42:15.700100 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Apr 30 12:42:15.712080 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Apr 30 12:42:15.814293 kernel: BTRFS info (device dm-0): first mount of filesystem 8f86a166-b3d6-49f7-a49d-597eaeb9f5e5 Apr 30 12:42:15.814310 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Apr 30 12:42:15.814320 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Apr 30 12:42:15.814330 kernel: BTRFS info (device dm-0): disabling log replay at mount time Apr 30 12:42:15.814340 kernel: BTRFS info (device dm-0): using free space tree Apr 30 12:42:15.814349 kernel: BTRFS info (device dm-0): enabling ssd optimizations Apr 30 12:42:15.725547 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Apr 30 12:42:15.821077 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Apr 30 12:42:15.829051 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Apr 30 12:42:15.844279 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Apr 30 12:42:15.960415 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 8d8cccbd-965f-4336-afa9-06a510e76633 Apr 30 12:42:15.960434 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Apr 30 12:42:15.960444 kernel: BTRFS info (device nvme0n1p6): using free space tree Apr 30 12:42:15.960453 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Apr 30 12:42:15.960463 kernel: BTRFS info (device nvme0n1p6): auto enabling async discard Apr 30 12:42:15.960472 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 8d8cccbd-965f-4336-afa9-06a510e76633 Apr 30 12:42:15.951134 systemd[1]: Finished ignition-setup.service - Ignition (setup). Apr 30 12:42:15.966251 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 12:42:16.001057 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Apr 30 12:42:16.013317 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 12:42:16.043843 systemd-networkd[1671]: lo: Link UP Apr 30 12:42:16.043849 systemd-networkd[1671]: lo: Gained carrier Apr 30 12:42:16.047725 systemd-networkd[1671]: Enumeration completed Apr 30 12:42:16.048083 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 12:42:16.049218 systemd-networkd[1671]: eno1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 12:42:16.055470 systemd[1]: Reached target network.target - Network. Apr 30 12:42:16.094760 ignition[1665]: Ignition 2.20.0 Apr 30 12:42:16.094768 ignition[1665]: Stage: fetch-offline Apr 30 12:42:16.100282 systemd-networkd[1671]: eno2: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 12:42:16.094813 ignition[1665]: no configs at "/usr/lib/ignition/base.d" Apr 30 12:42:16.105923 unknown[1665]: fetched base config from "system" Apr 30 12:42:16.094822 ignition[1665]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Apr 30 12:42:16.105958 unknown[1665]: fetched user config from "system" Apr 30 12:42:16.094985 ignition[1665]: parsed url from cmdline: "" Apr 30 12:42:16.108542 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 12:42:16.094988 ignition[1665]: no config URL provided Apr 30 12:42:16.121758 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Apr 30 12:42:16.094993 ignition[1665]: reading system config file "/usr/lib/ignition/user.ign" Apr 30 12:42:16.131102 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Apr 30 12:42:16.095042 ignition[1665]: parsing config with SHA512: 5cb90fa7c84b42a30b61419bf86792f7a31f8be5c398b57151fce01cbb04bdda59fdc27ed0e6f2fe97535644cfe8e439e39e49e7104627ce4d7fb0e64625db12 Apr 30 12:42:16.142452 systemd-networkd[1671]: enP1p1s0f0np0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 12:42:16.106430 ignition[1665]: fetch-offline: fetch-offline passed Apr 30 12:42:16.106436 ignition[1665]: POST message to Packet Timeline Apr 30 12:42:16.106441 ignition[1665]: POST Status error: resource requires networking Apr 30 12:42:16.106505 ignition[1665]: Ignition finished successfully Apr 30 12:42:16.148330 ignition[1706]: Ignition 2.20.0 Apr 30 12:42:16.148336 ignition[1706]: Stage: kargs Apr 30 12:42:16.148705 ignition[1706]: no configs at "/usr/lib/ignition/base.d" Apr 30 12:42:16.148714 ignition[1706]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Apr 30 12:42:16.150440 ignition[1706]: kargs: kargs passed Apr 30 12:42:16.150460 ignition[1706]: POST message to Packet Timeline Apr 30 12:42:16.150706 ignition[1706]: GET https://metadata.packet.net/metadata: attempt #1 Apr 30 12:42:16.153629 ignition[1706]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:55754->[::1]:53: read: connection refused Apr 30 12:42:16.354678 ignition[1706]: GET https://metadata.packet.net/metadata: attempt #2 Apr 30 12:42:16.355162 ignition[1706]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:48622->[::1]:53: read: connection refused Apr 30 12:42:16.721958 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: Link up Apr 30 12:42:16.725141 systemd-networkd[1671]: enP1p1s0f1np1: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 30 12:42:16.755980 ignition[1706]: GET https://metadata.packet.net/metadata: attempt #3 Apr 30 12:42:16.756625 ignition[1706]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:58799->[::1]:53: read: connection refused Apr 30 12:42:17.325965 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: Link up Apr 30 12:42:17.329683 systemd-networkd[1671]: eno1: Link UP Apr 30 12:42:17.329817 systemd-networkd[1671]: eno2: Link UP Apr 30 12:42:17.329936 systemd-networkd[1671]: enP1p1s0f0np0: Link UP Apr 30 12:42:17.330171 systemd-networkd[1671]: enP1p1s0f0np0: Gained carrier Apr 30 12:42:17.336128 systemd-networkd[1671]: enP1p1s0f1np1: Link UP Apr 30 12:42:17.367994 systemd-networkd[1671]: enP1p1s0f0np0: DHCPv4 address 147.75.61.11/31, gateway 147.75.61.10 acquired from 147.28.144.140 Apr 30 12:42:17.557293 ignition[1706]: GET https://metadata.packet.net/metadata: attempt #4 Apr 30 12:42:17.558001 ignition[1706]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:35123->[::1]:53: read: connection refused Apr 30 12:42:17.729290 systemd-networkd[1671]: enP1p1s0f1np1: Gained carrier Apr 30 12:42:18.849202 systemd-networkd[1671]: enP1p1s0f1np1: Gained IPv6LL Apr 30 12:42:19.041231 systemd-networkd[1671]: enP1p1s0f0np0: Gained IPv6LL Apr 30 12:42:19.159713 ignition[1706]: GET https://metadata.packet.net/metadata: attempt #5 Apr 30 12:42:19.160521 ignition[1706]: GET error: Get "https://metadata.packet.net/metadata": dial tcp: lookup metadata.packet.net on [::1]:53: read udp [::1]:50815->[::1]:53: read: connection refused Apr 30 12:42:22.362939 ignition[1706]: GET https://metadata.packet.net/metadata: attempt #6 Apr 30 12:42:22.845639 ignition[1706]: GET result: OK Apr 30 12:42:23.145511 ignition[1706]: Ignition finished successfully Apr 30 12:42:23.148135 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Apr 30 12:42:23.159073 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Apr 30 12:42:23.174069 ignition[1725]: Ignition 2.20.0 Apr 30 12:42:23.174076 ignition[1725]: Stage: disks Apr 30 12:42:23.174229 ignition[1725]: no configs at "/usr/lib/ignition/base.d" Apr 30 12:42:23.174239 ignition[1725]: no config dir at "/usr/lib/ignition/base.platform.d/packet" Apr 30 12:42:23.175750 ignition[1725]: disks: disks passed Apr 30 12:42:23.175755 ignition[1725]: POST message to Packet Timeline Apr 30 12:42:23.175774 ignition[1725]: GET https://metadata.packet.net/metadata: attempt #1 Apr 30 12:42:24.154948 ignition[1725]: GET result: OK Apr 30 12:42:24.512259 ignition[1725]: Ignition finished successfully Apr 30 12:42:24.515358 systemd[1]: Finished ignition-disks.service - Ignition (disks). Apr 30 12:42:24.521039 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Apr 30 12:42:24.528460 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Apr 30 12:42:24.536316 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 12:42:24.544653 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 12:42:24.553430 systemd[1]: Reached target basic.target - Basic System. Apr 30 12:42:24.574051 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Apr 30 12:42:24.589789 systemd-fsck[1745]: ROOT: clean, 14/553520 files, 52654/553472 blocks Apr 30 12:42:24.593583 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Apr 30 12:42:24.613020 systemd[1]: Mounting sysroot.mount - /sysroot... Apr 30 12:42:24.678954 kernel: EXT4-fs (nvme0n1p9): mounted filesystem 597557b0-8ae6-4a5a-8e98-f3f884fcfe65 r/w with ordered data mode. Quota mode: none. Apr 30 12:42:24.679111 systemd[1]: Mounted sysroot.mount - /sysroot. Apr 30 12:42:24.689372 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Apr 30 12:42:24.711028 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 12:42:24.719961 kernel: BTRFS: device label OEM devid 1 transid 18 /dev/nvme0n1p6 scanned by mount (1757) Apr 30 12:42:24.719993 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 8d8cccbd-965f-4336-afa9-06a510e76633 Apr 30 12:42:24.720004 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Apr 30 12:42:24.720013 kernel: BTRFS info (device nvme0n1p6): using free space tree Apr 30 12:42:24.720954 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Apr 30 12:42:24.720965 kernel: BTRFS info (device nvme0n1p6): auto enabling async discard Apr 30 12:42:24.811028 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Apr 30 12:42:24.817582 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Apr 30 12:42:24.828342 systemd[1]: Starting flatcar-static-network.service - Flatcar Static Network Agent... Apr 30 12:42:24.844143 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 30 12:42:24.844183 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 12:42:24.857999 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 12:42:24.888420 coreos-metadata[1779]: Apr 30 12:42:24.872 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Apr 30 12:42:24.907035 coreos-metadata[1776]: Apr 30 12:42:24.872 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Apr 30 12:42:24.871578 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Apr 30 12:42:24.896060 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Apr 30 12:42:24.935362 initrd-setup-root[1799]: cut: /sysroot/etc/passwd: No such file or directory Apr 30 12:42:24.941411 initrd-setup-root[1806]: cut: /sysroot/etc/group: No such file or directory Apr 30 12:42:24.947409 initrd-setup-root[1814]: cut: /sysroot/etc/shadow: No such file or directory Apr 30 12:42:24.953414 initrd-setup-root[1822]: cut: /sysroot/etc/gshadow: No such file or directory Apr 30 12:42:25.021463 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Apr 30 12:42:25.043049 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Apr 30 12:42:25.050956 kernel: BTRFS info (device nvme0n1p6): last unmount of filesystem 8d8cccbd-965f-4336-afa9-06a510e76633 Apr 30 12:42:25.074763 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Apr 30 12:42:25.081718 systemd[1]: sysroot-oem.mount: Deactivated successfully. Apr 30 12:42:25.101932 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Apr 30 12:42:25.110869 ignition[1895]: INFO : Ignition 2.20.0 Apr 30 12:42:25.110869 ignition[1895]: INFO : Stage: mount Apr 30 12:42:25.123511 ignition[1895]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 12:42:25.123511 ignition[1895]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Apr 30 12:42:25.123511 ignition[1895]: INFO : mount: mount passed Apr 30 12:42:25.123511 ignition[1895]: INFO : POST message to Packet Timeline Apr 30 12:42:25.123511 ignition[1895]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Apr 30 12:42:25.287555 coreos-metadata[1779]: Apr 30 12:42:25.287 INFO Fetch successful Apr 30 12:42:25.335574 systemd[1]: flatcar-static-network.service: Deactivated successfully. Apr 30 12:42:25.335722 systemd[1]: Finished flatcar-static-network.service - Flatcar Static Network Agent. Apr 30 12:42:25.447593 coreos-metadata[1776]: Apr 30 12:42:25.447 INFO Fetch successful Apr 30 12:42:25.492831 coreos-metadata[1776]: Apr 30 12:42:25.492 INFO wrote hostname ci-4230.1.1-a-0ae8677ef9 to /sysroot/etc/hostname Apr 30 12:42:25.496238 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 30 12:42:25.875882 ignition[1895]: INFO : GET result: OK Apr 30 12:42:26.383627 ignition[1895]: INFO : Ignition finished successfully Apr 30 12:42:26.385919 systemd[1]: Finished ignition-mount.service - Ignition (mount). Apr 30 12:42:26.405007 systemd[1]: Starting ignition-files.service - Ignition (files)... Apr 30 12:42:26.413313 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Apr 30 12:42:26.450639 kernel: BTRFS: device label OEM devid 1 transid 19 /dev/nvme0n1p6 scanned by mount (1923) Apr 30 12:42:26.450681 kernel: BTRFS info (device nvme0n1p6): first mount of filesystem 8d8cccbd-965f-4336-afa9-06a510e76633 Apr 30 12:42:26.465025 kernel: BTRFS info (device nvme0n1p6): using crc32c (crc32c-generic) checksum algorithm Apr 30 12:42:26.478035 kernel: BTRFS info (device nvme0n1p6): using free space tree Apr 30 12:42:26.501011 kernel: BTRFS info (device nvme0n1p6): enabling ssd optimizations Apr 30 12:42:26.501035 kernel: BTRFS info (device nvme0n1p6): auto enabling async discard Apr 30 12:42:26.509051 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Apr 30 12:42:26.541139 ignition[1942]: INFO : Ignition 2.20.0 Apr 30 12:42:26.541139 ignition[1942]: INFO : Stage: files Apr 30 12:42:26.550811 ignition[1942]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 12:42:26.550811 ignition[1942]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Apr 30 12:42:26.550811 ignition[1942]: DEBUG : files: compiled without relabeling support, skipping Apr 30 12:42:26.550811 ignition[1942]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 30 12:42:26.550811 ignition[1942]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 30 12:42:26.550811 ignition[1942]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 30 12:42:26.550811 ignition[1942]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 30 12:42:26.550811 ignition[1942]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 30 12:42:26.550811 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 30 12:42:26.550811 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Apr 30 12:42:26.546525 unknown[1942]: wrote ssh authorized keys file for user: core Apr 30 12:42:26.834539 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 30 12:42:27.876192 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 12:42:27.887139 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.31.0-arm64.raw: attempt #1 Apr 30 12:42:28.070202 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 30 12:42:28.212345 ignition[1942]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.0-arm64.raw" Apr 30 12:42:28.212345 ignition[1942]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 30 12:42:28.237143 ignition[1942]: INFO : files: files passed Apr 30 12:42:28.237143 ignition[1942]: INFO : POST message to Packet Timeline Apr 30 12:42:28.237143 ignition[1942]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Apr 30 12:42:28.685740 ignition[1942]: INFO : GET result: OK Apr 30 12:42:29.009722 ignition[1942]: INFO : Ignition finished successfully Apr 30 12:42:29.012938 systemd[1]: Finished ignition-files.service - Ignition (files). Apr 30 12:42:29.031110 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Apr 30 12:42:29.043696 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Apr 30 12:42:29.062320 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 30 12:42:29.062499 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Apr 30 12:42:29.080493 initrd-setup-root-after-ignition[1984]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 12:42:29.080493 initrd-setup-root-after-ignition[1984]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Apr 30 12:42:29.074934 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 12:42:29.132844 initrd-setup-root-after-ignition[1988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 30 12:42:29.087894 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Apr 30 12:42:29.113182 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Apr 30 12:42:29.146960 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 30 12:42:29.147052 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Apr 30 12:42:29.157107 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Apr 30 12:42:29.173218 systemd[1]: Reached target initrd.target - Initrd Default Target. Apr 30 12:42:29.184717 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Apr 30 12:42:29.199097 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Apr 30 12:42:29.220855 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 12:42:29.241190 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Apr 30 12:42:29.258176 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Apr 30 12:42:29.267327 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 12:42:29.278895 systemd[1]: Stopped target timers.target - Timer Units. Apr 30 12:42:29.290492 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 30 12:42:29.290606 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Apr 30 12:42:29.302193 systemd[1]: Stopped target initrd.target - Initrd Default Target. Apr 30 12:42:29.313524 systemd[1]: Stopped target basic.target - Basic System. Apr 30 12:42:29.325059 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Apr 30 12:42:29.336515 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Apr 30 12:42:29.347791 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Apr 30 12:42:29.359098 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Apr 30 12:42:29.370350 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Apr 30 12:42:29.381662 systemd[1]: Stopped target sysinit.target - System Initialization. Apr 30 12:42:29.393071 systemd[1]: Stopped target local-fs.target - Local File Systems. Apr 30 12:42:29.409916 systemd[1]: Stopped target swap.target - Swaps. Apr 30 12:42:29.421286 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 30 12:42:29.421378 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Apr 30 12:42:29.432784 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Apr 30 12:42:29.443865 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 12:42:29.455203 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Apr 30 12:42:29.458978 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 12:42:29.466553 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 30 12:42:29.466661 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Apr 30 12:42:29.477994 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 30 12:42:29.478097 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Apr 30 12:42:29.489304 systemd[1]: Stopped target paths.target - Path Units. Apr 30 12:42:29.500512 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 30 12:42:29.500617 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 12:42:29.517619 systemd[1]: Stopped target slices.target - Slice Units. Apr 30 12:42:29.529125 systemd[1]: Stopped target sockets.target - Socket Units. Apr 30 12:42:29.540610 systemd[1]: iscsid.socket: Deactivated successfully. Apr 30 12:42:29.540701 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Apr 30 12:42:29.640176 ignition[2010]: INFO : Ignition 2.20.0 Apr 30 12:42:29.640176 ignition[2010]: INFO : Stage: umount Apr 30 12:42:29.640176 ignition[2010]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 30 12:42:29.640176 ignition[2010]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/packet" Apr 30 12:42:29.640176 ignition[2010]: INFO : umount: umount passed Apr 30 12:42:29.640176 ignition[2010]: INFO : POST message to Packet Timeline Apr 30 12:42:29.640176 ignition[2010]: INFO : GET https://metadata.packet.net/metadata: attempt #1 Apr 30 12:42:29.552240 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 30 12:42:29.552314 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Apr 30 12:42:29.563967 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 30 12:42:29.564056 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Apr 30 12:42:29.575597 systemd[1]: ignition-files.service: Deactivated successfully. Apr 30 12:42:29.575696 systemd[1]: Stopped ignition-files.service - Ignition (files). Apr 30 12:42:29.587214 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 30 12:42:29.587309 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Apr 30 12:42:29.615149 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Apr 30 12:42:29.622265 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 30 12:42:29.622367 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 12:42:29.634929 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Apr 30 12:42:29.646281 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 30 12:42:29.646389 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 12:42:29.658073 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 30 12:42:29.658175 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Apr 30 12:42:29.671912 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 30 12:42:29.672798 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 30 12:42:29.672876 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Apr 30 12:42:29.682928 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 30 12:42:29.683006 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Apr 30 12:42:30.108442 ignition[2010]: INFO : GET result: OK Apr 30 12:42:30.365422 ignition[2010]: INFO : Ignition finished successfully Apr 30 12:42:30.367708 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 30 12:42:30.369037 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Apr 30 12:42:30.375774 systemd[1]: Stopped target network.target - Network. Apr 30 12:42:30.385103 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 30 12:42:30.385176 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Apr 30 12:42:30.394911 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 30 12:42:30.394966 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Apr 30 12:42:30.404491 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 30 12:42:30.404537 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Apr 30 12:42:30.414073 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Apr 30 12:42:30.414103 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Apr 30 12:42:30.423808 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 30 12:42:30.423877 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Apr 30 12:42:30.433711 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Apr 30 12:42:30.443350 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Apr 30 12:42:30.453304 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 30 12:42:30.453423 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Apr 30 12:42:30.467592 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Apr 30 12:42:30.467936 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 30 12:42:30.468029 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Apr 30 12:42:30.478577 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Apr 30 12:42:30.480565 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 30 12:42:30.480758 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Apr 30 12:42:30.498045 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Apr 30 12:42:30.507336 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 30 12:42:30.507386 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Apr 30 12:42:30.517230 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 30 12:42:30.517266 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Apr 30 12:42:30.527322 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 30 12:42:30.527378 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Apr 30 12:42:30.542486 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Apr 30 12:42:30.542518 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 12:42:30.553032 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 12:42:30.564877 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Apr 30 12:42:30.564938 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Apr 30 12:42:30.574216 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 30 12:42:30.575971 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 12:42:30.587039 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 30 12:42:30.587193 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Apr 30 12:42:30.596233 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 30 12:42:30.596282 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 12:42:30.606978 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 30 12:42:30.607017 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Apr 30 12:42:30.623544 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 30 12:42:30.623600 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Apr 30 12:42:30.634524 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 30 12:42:30.634575 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Apr 30 12:42:30.652086 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Apr 30 12:42:30.662621 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 30 12:42:30.662680 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 12:42:30.673957 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 30 12:42:30.674020 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 12:42:30.686713 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Apr 30 12:42:30.686787 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Apr 30 12:42:30.687099 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 30 12:42:30.687186 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Apr 30 12:42:31.238193 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 30 12:42:31.239063 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Apr 30 12:42:31.250133 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Apr 30 12:42:31.271097 systemd[1]: Starting initrd-switch-root.service - Switch Root... Apr 30 12:42:31.285095 systemd[1]: Switching root. Apr 30 12:42:31.341687 systemd-journald[899]: Journal stopped Apr 30 12:42:33.411833 systemd-journald[899]: Received SIGTERM from PID 1 (systemd). Apr 30 12:42:33.411861 kernel: SELinux: policy capability network_peer_controls=1 Apr 30 12:42:33.411872 kernel: SELinux: policy capability open_perms=1 Apr 30 12:42:33.411880 kernel: SELinux: policy capability extended_socket_class=1 Apr 30 12:42:33.411888 kernel: SELinux: policy capability always_check_network=0 Apr 30 12:42:33.411895 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 30 12:42:33.411904 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 30 12:42:33.411913 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 30 12:42:33.411921 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 30 12:42:33.411929 kernel: audit: type=1403 audit(1746016951.508:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 30 12:42:33.411938 systemd[1]: Successfully loaded SELinux policy in 116.022ms. Apr 30 12:42:33.411947 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 9.985ms. Apr 30 12:42:33.411959 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Apr 30 12:42:33.411968 systemd[1]: Detected architecture arm64. Apr 30 12:42:33.411979 systemd[1]: Detected first boot. Apr 30 12:42:33.411988 systemd[1]: Hostname set to . Apr 30 12:42:33.411997 systemd[1]: Initializing machine ID from random generator. Apr 30 12:42:33.412006 zram_generator::config[2078]: No configuration found. Apr 30 12:42:33.412017 systemd[1]: Populated /etc with preset unit settings. Apr 30 12:42:33.412026 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Apr 30 12:42:33.412035 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 30 12:42:33.412043 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Apr 30 12:42:33.412052 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 30 12:42:33.412061 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Apr 30 12:42:33.412070 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Apr 30 12:42:33.412081 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Apr 30 12:42:33.412092 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Apr 30 12:42:33.412102 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Apr 30 12:42:33.412111 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Apr 30 12:42:33.412120 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Apr 30 12:42:33.412129 systemd[1]: Created slice user.slice - User and Session Slice. Apr 30 12:42:33.412138 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Apr 30 12:42:33.412147 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Apr 30 12:42:33.412157 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Apr 30 12:42:33.412166 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Apr 30 12:42:33.412175 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Apr 30 12:42:33.412184 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Apr 30 12:42:33.412193 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Apr 30 12:42:33.412202 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Apr 30 12:42:33.412211 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Apr 30 12:42:33.412222 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Apr 30 12:42:33.412231 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Apr 30 12:42:33.412242 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Apr 30 12:42:33.412251 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Apr 30 12:42:33.412260 systemd[1]: Reached target remote-fs.target - Remote File Systems. Apr 30 12:42:33.412269 systemd[1]: Reached target slices.target - Slice Units. Apr 30 12:42:33.412278 systemd[1]: Reached target swap.target - Swaps. Apr 30 12:42:33.412287 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Apr 30 12:42:33.412296 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Apr 30 12:42:33.412307 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Apr 30 12:42:33.412316 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Apr 30 12:42:33.412326 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Apr 30 12:42:33.412335 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Apr 30 12:42:33.412344 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Apr 30 12:42:33.412354 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Apr 30 12:42:33.412364 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Apr 30 12:42:33.412373 systemd[1]: Mounting media.mount - External Media Directory... Apr 30 12:42:33.412382 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Apr 30 12:42:33.412391 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Apr 30 12:42:33.412400 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Apr 30 12:42:33.412409 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 30 12:42:33.412419 systemd[1]: Reached target machines.target - Containers. Apr 30 12:42:33.412429 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Apr 30 12:42:33.412439 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 12:42:33.412448 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Apr 30 12:42:33.412457 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Apr 30 12:42:33.412466 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 12:42:33.412475 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 30 12:42:33.412485 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 12:42:33.412494 kernel: ACPI: bus type drm_connector registered Apr 30 12:42:33.412503 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Apr 30 12:42:33.412513 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 12:42:33.412522 kernel: fuse: init (API version 7.39) Apr 30 12:42:33.412530 kernel: loop: module loaded Apr 30 12:42:33.412539 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 30 12:42:33.412548 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 30 12:42:33.412557 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Apr 30 12:42:33.412566 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 30 12:42:33.412575 systemd[1]: Stopped systemd-fsck-usr.service. Apr 30 12:42:33.412586 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 30 12:42:33.412596 systemd[1]: Starting systemd-journald.service - Journal Service... Apr 30 12:42:33.412605 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Apr 30 12:42:33.412633 systemd-journald[2185]: Collecting audit messages is disabled. Apr 30 12:42:33.412658 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Apr 30 12:42:33.412669 systemd-journald[2185]: Journal started Apr 30 12:42:33.412687 systemd-journald[2185]: Runtime Journal (/run/log/journal/dee5a3add1eb4485bc05e2278bef1e6d) is 8M, max 4G, 3.9G free. Apr 30 12:42:32.066563 systemd[1]: Queued start job for default target multi-user.target. Apr 30 12:42:32.080367 systemd[1]: Unnecessary job was removed for dev-nvme0n1p6.device - /dev/nvme0n1p6. Apr 30 12:42:32.080698 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 30 12:42:32.082142 systemd[1]: systemd-journald.service: Consumed 3.400s CPU time. Apr 30 12:42:33.463965 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Apr 30 12:42:33.490963 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Apr 30 12:42:33.511967 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Apr 30 12:42:33.535079 systemd[1]: verity-setup.service: Deactivated successfully. Apr 30 12:42:33.535114 systemd[1]: Stopped verity-setup.service. Apr 30 12:42:33.560968 systemd[1]: Started systemd-journald.service - Journal Service. Apr 30 12:42:33.566371 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Apr 30 12:42:33.572019 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Apr 30 12:42:33.577469 systemd[1]: Mounted media.mount - External Media Directory. Apr 30 12:42:33.582880 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Apr 30 12:42:33.588375 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Apr 30 12:42:33.593788 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Apr 30 12:42:33.599305 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Apr 30 12:42:33.604840 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Apr 30 12:42:33.610516 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 30 12:42:33.610685 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Apr 30 12:42:33.616032 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 12:42:33.616193 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 12:42:33.623521 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 30 12:42:33.624981 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 30 12:42:33.630425 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 12:42:33.630584 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 12:42:33.635971 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 30 12:42:33.636131 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Apr 30 12:42:33.641386 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 12:42:33.643006 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 12:42:33.648198 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Apr 30 12:42:33.655336 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Apr 30 12:42:33.661972 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Apr 30 12:42:33.667195 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Apr 30 12:42:33.673034 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Apr 30 12:42:33.688728 systemd[1]: Reached target network-pre.target - Preparation for Network. Apr 30 12:42:33.708025 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Apr 30 12:42:33.713908 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Apr 30 12:42:33.718694 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 30 12:42:33.718723 systemd[1]: Reached target local-fs.target - Local File Systems. Apr 30 12:42:33.724211 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Apr 30 12:42:33.729846 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Apr 30 12:42:33.735617 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Apr 30 12:42:33.740411 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 12:42:33.741859 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Apr 30 12:42:33.747502 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Apr 30 12:42:33.752215 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 30 12:42:33.753348 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Apr 30 12:42:33.758039 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 30 12:42:33.758171 systemd-journald[2185]: Time spent on flushing to /var/log/journal/dee5a3add1eb4485bc05e2278bef1e6d is 24.165ms for 2357 entries. Apr 30 12:42:33.758171 systemd-journald[2185]: System Journal (/var/log/journal/dee5a3add1eb4485bc05e2278bef1e6d) is 8M, max 195.6M, 187.6M free. Apr 30 12:42:33.799474 systemd-journald[2185]: Received client request to flush runtime journal. Apr 30 12:42:33.799521 kernel: loop0: detected capacity change from 0 to 123192 Apr 30 12:42:33.759152 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Apr 30 12:42:33.776193 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Apr 30 12:42:33.781960 systemd[1]: Starting systemd-sysusers.service - Create System Users... Apr 30 12:42:33.787609 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Apr 30 12:42:33.804627 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Apr 30 12:42:33.813955 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Apr 30 12:42:33.818010 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Apr 30 12:42:33.823584 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Apr 30 12:42:33.829074 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Apr 30 12:42:33.833802 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Apr 30 12:42:33.838608 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Apr 30 12:42:33.843529 systemd[1]: Finished systemd-sysusers.service - Create System Users. Apr 30 12:42:33.854012 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Apr 30 12:42:33.871965 kernel: loop1: detected capacity change from 0 to 8 Apr 30 12:42:33.873234 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Apr 30 12:42:33.879239 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Apr 30 12:42:33.885282 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 30 12:42:33.886727 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Apr 30 12:42:33.895212 udevadm[2238]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 30 12:42:33.900724 systemd-tmpfiles[2257]: ACLs are not supported, ignoring. Apr 30 12:42:33.900737 systemd-tmpfiles[2257]: ACLs are not supported, ignoring. Apr 30 12:42:33.904628 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Apr 30 12:42:33.934984 kernel: loop2: detected capacity change from 0 to 113512 Apr 30 12:42:33.984962 kernel: loop3: detected capacity change from 0 to 189592 Apr 30 12:42:34.014424 ldconfig[2222]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 30 12:42:34.016045 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Apr 30 12:42:34.031957 kernel: loop4: detected capacity change from 0 to 123192 Apr 30 12:42:34.046965 kernel: loop5: detected capacity change from 0 to 8 Apr 30 12:42:34.058962 kernel: loop6: detected capacity change from 0 to 113512 Apr 30 12:42:34.074962 kernel: loop7: detected capacity change from 0 to 189592 Apr 30 12:42:34.080665 (sd-merge)[2271]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-packet'. Apr 30 12:42:34.081121 (sd-merge)[2271]: Merged extensions into '/usr'. Apr 30 12:42:34.083988 systemd[1]: Reload requested from client PID 2234 ('systemd-sysext') (unit systemd-sysext.service)... Apr 30 12:42:34.083999 systemd[1]: Reloading... Apr 30 12:42:34.136956 zram_generator::config[2301]: No configuration found. Apr 30 12:42:34.230806 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 12:42:34.292285 systemd[1]: Reloading finished in 207 ms. Apr 30 12:42:34.310455 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Apr 30 12:42:34.316276 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Apr 30 12:42:34.334347 systemd[1]: Starting ensure-sysext.service... Apr 30 12:42:34.340089 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Apr 30 12:42:34.346555 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Apr 30 12:42:34.357407 systemd[1]: Reload requested from client PID 2351 ('systemctl') (unit ensure-sysext.service)... Apr 30 12:42:34.357420 systemd[1]: Reloading... Apr 30 12:42:34.359480 systemd-tmpfiles[2352]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 30 12:42:34.359679 systemd-tmpfiles[2352]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Apr 30 12:42:34.360338 systemd-tmpfiles[2352]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 30 12:42:34.360535 systemd-tmpfiles[2352]: ACLs are not supported, ignoring. Apr 30 12:42:34.360582 systemd-tmpfiles[2352]: ACLs are not supported, ignoring. Apr 30 12:42:34.364672 systemd-tmpfiles[2352]: Detected autofs mount point /boot during canonicalization of boot. Apr 30 12:42:34.364680 systemd-tmpfiles[2352]: Skipping /boot Apr 30 12:42:34.373134 systemd-tmpfiles[2352]: Detected autofs mount point /boot during canonicalization of boot. Apr 30 12:42:34.373142 systemd-tmpfiles[2352]: Skipping /boot Apr 30 12:42:34.374287 systemd-udevd[2353]: Using default interface naming scheme 'v255'. Apr 30 12:42:34.410960 zram_generator::config[2406]: No configuration found. Apr 30 12:42:34.439971 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (2437) Apr 30 12:42:34.460971 kernel: IPMI message handler: version 39.2 Apr 30 12:42:34.470960 kernel: ipmi device interface Apr 30 12:42:34.487837 kernel: ipmi_si: IPMI System Interface driver Apr 30 12:42:34.487939 kernel: ipmi_ssif: IPMI SSIF Interface driver Apr 30 12:42:34.487985 kernel: ipmi_si: Unable to find any System Interface(s) Apr 30 12:42:34.522459 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 12:42:34.603077 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - SAMSUNG MZ1LB960HAJQ-00007 OEM. Apr 30 12:42:34.607774 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Apr 30 12:42:34.607993 systemd[1]: Reloading finished in 250 ms. Apr 30 12:42:34.621323 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Apr 30 12:42:34.645940 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Apr 30 12:42:34.664154 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Apr 30 12:42:34.674525 systemd[1]: Finished ensure-sysext.service. Apr 30 12:42:34.708273 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 30 12:42:34.714214 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Apr 30 12:42:34.719198 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Apr 30 12:42:34.720267 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Apr 30 12:42:34.726013 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Apr 30 12:42:34.731813 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Apr 30 12:42:34.737392 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Apr 30 12:42:34.737992 lvm[2577]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 30 12:42:34.743019 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Apr 30 12:42:34.747872 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Apr 30 12:42:34.748675 augenrules[2597]: No rules Apr 30 12:42:34.748768 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Apr 30 12:42:34.753509 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Apr 30 12:42:34.754741 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Apr 30 12:42:34.761109 systemd[1]: Starting systemd-networkd.service - Network Configuration... Apr 30 12:42:34.767734 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Apr 30 12:42:34.773748 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Apr 30 12:42:34.779272 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Apr 30 12:42:34.784761 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Apr 30 12:42:34.790200 systemd[1]: audit-rules.service: Deactivated successfully. Apr 30 12:42:34.790412 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 30 12:42:34.795425 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Apr 30 12:42:34.801725 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Apr 30 12:42:34.806743 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 30 12:42:34.807432 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Apr 30 12:42:34.812217 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 30 12:42:34.812365 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Apr 30 12:42:34.817383 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 30 12:42:34.818060 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Apr 30 12:42:34.823055 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 30 12:42:34.823203 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Apr 30 12:42:34.827899 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Apr 30 12:42:34.832971 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Apr 30 12:42:34.837960 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Apr 30 12:42:34.851314 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Apr 30 12:42:34.874136 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Apr 30 12:42:34.877994 lvm[2629]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 30 12:42:34.878669 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 30 12:42:34.878734 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Apr 30 12:42:34.879879 systemd[1]: Starting systemd-update-done.service - Update is Completed... Apr 30 12:42:34.886292 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Apr 30 12:42:34.890917 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 30 12:42:34.891987 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Apr 30 12:42:34.896856 systemd[1]: Finished systemd-update-done.service - Update is Completed. Apr 30 12:42:34.912444 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Apr 30 12:42:34.922840 systemd[1]: Started systemd-userdbd.service - User Database Manager. Apr 30 12:42:34.987343 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Apr 30 12:42:34.992396 systemd[1]: Reached target time-set.target - System Time Set. Apr 30 12:42:34.994553 systemd-resolved[2607]: Positive Trust Anchors: Apr 30 12:42:34.994565 systemd-resolved[2607]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 30 12:42:34.994595 systemd-resolved[2607]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Apr 30 12:42:34.998133 systemd-resolved[2607]: Using system hostname 'ci-4230.1.1-a-0ae8677ef9'. Apr 30 12:42:35.000244 systemd-networkd[2606]: lo: Link UP Apr 30 12:42:35.000250 systemd-networkd[2606]: lo: Gained carrier Apr 30 12:42:35.000270 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Apr 30 12:42:35.004121 systemd-networkd[2606]: bond0: netdev ready Apr 30 12:42:35.004768 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Apr 30 12:42:35.009074 systemd[1]: Reached target sysinit.target - System Initialization. Apr 30 12:42:35.012934 systemd-networkd[2606]: Enumeration completed Apr 30 12:42:35.013465 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Apr 30 12:42:35.017827 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Apr 30 12:42:35.021409 systemd-networkd[2606]: enP1p1s0f0np0: Configuring with /etc/systemd/network/10-0c:42:a1:49:ce:00.network. Apr 30 12:42:35.022540 systemd[1]: Started logrotate.timer - Daily rotation of log files. Apr 30 12:42:35.026889 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Apr 30 12:42:35.031227 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Apr 30 12:42:35.035587 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 30 12:42:35.035609 systemd[1]: Reached target paths.target - Path Units. Apr 30 12:42:35.039953 systemd[1]: Reached target timers.target - Timer Units. Apr 30 12:42:35.045010 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Apr 30 12:42:35.050817 systemd[1]: Starting docker.socket - Docker Socket for the API... Apr 30 12:42:35.057223 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Apr 30 12:42:35.064147 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Apr 30 12:42:35.069067 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Apr 30 12:42:35.074073 systemd[1]: Started systemd-networkd.service - Network Configuration. Apr 30 12:42:35.078811 systemd[1]: Listening on docker.socket - Docker Socket for the API. Apr 30 12:42:35.083369 systemd[1]: Reached target network.target - Network. Apr 30 12:42:35.087744 systemd[1]: Reached target sockets.target - Socket Units. Apr 30 12:42:35.092050 systemd[1]: Reached target basic.target - Basic System. Apr 30 12:42:35.096327 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Apr 30 12:42:35.096348 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Apr 30 12:42:35.112039 systemd[1]: Starting containerd.service - containerd container runtime... Apr 30 12:42:35.117661 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Apr 30 12:42:35.123267 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Apr 30 12:42:35.128856 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Apr 30 12:42:35.134445 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Apr 30 12:42:35.138900 jq[2664]: false Apr 30 12:42:35.138925 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Apr 30 12:42:35.139210 coreos-metadata[2660]: Apr 30 12:42:35.139 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Apr 30 12:42:35.139989 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Apr 30 12:42:35.141920 coreos-metadata[2660]: Apr 30 12:42:35.141 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Apr 30 12:42:35.144417 dbus-daemon[2661]: [system] SELinux support is enabled Apr 30 12:42:35.145494 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Apr 30 12:42:35.151056 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Apr 30 12:42:35.154652 extend-filesystems[2665]: Found loop4 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found loop5 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found loop6 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found loop7 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme1n1 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p1 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p2 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p3 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found usr Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p4 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p6 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p7 Apr 30 12:42:35.160338 extend-filesystems[2665]: Found nvme0n1p9 Apr 30 12:42:35.160338 extend-filesystems[2665]: Checking size of /dev/nvme0n1p9 Apr 30 12:42:35.304503 kernel: EXT4-fs (nvme0n1p9): resizing filesystem from 553472 to 233815889 blocks Apr 30 12:42:35.304526 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (2445) Apr 30 12:42:35.156680 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Apr 30 12:42:35.304595 extend-filesystems[2665]: Resized partition /dev/nvme0n1p9 Apr 30 12:42:35.298226 dbus-daemon[2661]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 30 12:42:35.168833 systemd[1]: Starting systemd-logind.service - User Login Management... Apr 30 12:42:35.309244 extend-filesystems[2686]: resize2fs 1.47.1 (20-May-2024) Apr 30 12:42:35.175028 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Apr 30 12:42:35.215417 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Apr 30 12:42:35.224774 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 30 12:42:35.225373 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 30 12:42:35.314405 update_engine[2695]: I20250430 12:42:35.276556 2695 main.cc:92] Flatcar Update Engine starting Apr 30 12:42:35.314405 update_engine[2695]: I20250430 12:42:35.279963 2695 update_check_scheduler.cc:74] Next update check in 5m40s Apr 30 12:42:35.226003 systemd[1]: Starting update-engine.service - Update Engine... Apr 30 12:42:35.314723 jq[2696]: true Apr 30 12:42:35.234261 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Apr 30 12:42:35.243051 systemd[1]: Started dbus.service - D-Bus System Message Bus. Apr 30 12:42:35.315027 tar[2698]: linux-arm64/helm Apr 30 12:42:35.251584 systemd-logind[2684]: Watching system buttons on /dev/input/event0 (Power Button) Apr 30 12:42:35.315396 jq[2699]: true Apr 30 12:42:35.252065 systemd-logind[2684]: New seat seat0. Apr 30 12:42:35.258554 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 30 12:42:35.258745 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Apr 30 12:42:35.259130 systemd[1]: motdgen.service: Deactivated successfully. Apr 30 12:42:35.259308 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Apr 30 12:42:35.267424 systemd[1]: Started systemd-logind.service - User Login Management. Apr 30 12:42:35.277000 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 30 12:42:35.277182 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Apr 30 12:42:35.295846 (ntainerd)[2700]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Apr 30 12:42:35.313746 systemd[1]: Started update-engine.service - Update Engine. Apr 30 12:42:35.323382 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 30 12:42:35.323548 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Apr 30 12:42:35.328520 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 30 12:42:35.328632 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Apr 30 12:42:35.339003 bash[2724]: Updated "/home/core/.ssh/authorized_keys" Apr 30 12:42:35.347153 systemd[1]: Started locksmithd.service - Cluster reboot manager. Apr 30 12:42:35.355391 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Apr 30 12:42:35.362595 systemd[1]: Starting sshkeys.service... Apr 30 12:42:35.374452 locksmithd[2725]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 30 12:42:35.375646 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Apr 30 12:42:35.381594 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Apr 30 12:42:35.401186 coreos-metadata[2748]: Apr 30 12:42:35.401 INFO Fetching https://metadata.packet.net/metadata: Attempt #1 Apr 30 12:42:35.402302 coreos-metadata[2748]: Apr 30 12:42:35.402 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Apr 30 12:42:35.434994 containerd[2700]: time="2025-04-30T12:42:35.434911920Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Apr 30 12:42:35.457277 containerd[2700]: time="2025-04-30T12:42:35.457236960Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.458550 containerd[2700]: time="2025-04-30T12:42:35.458525080Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.88-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 30 12:42:35.458569 containerd[2700]: time="2025-04-30T12:42:35.458550720Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 30 12:42:35.458585 containerd[2700]: time="2025-04-30T12:42:35.458565760Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 30 12:42:35.458736 containerd[2700]: time="2025-04-30T12:42:35.458720240Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Apr 30 12:42:35.458759 containerd[2700]: time="2025-04-30T12:42:35.458738560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.458810 containerd[2700]: time="2025-04-30T12:42:35.458795000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 12:42:35.458834 containerd[2700]: time="2025-04-30T12:42:35.458809920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459017 containerd[2700]: time="2025-04-30T12:42:35.459001000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459037 containerd[2700]: time="2025-04-30T12:42:35.459015560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459037 containerd[2700]: time="2025-04-30T12:42:35.459028160Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459069 containerd[2700]: time="2025-04-30T12:42:35.459038040Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459128 containerd[2700]: time="2025-04-30T12:42:35.459114760Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459316 containerd[2700]: time="2025-04-30T12:42:35.459300480Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459434 containerd[2700]: time="2025-04-30T12:42:35.459419840Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 30 12:42:35.459455 containerd[2700]: time="2025-04-30T12:42:35.459433360Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 30 12:42:35.459522 containerd[2700]: time="2025-04-30T12:42:35.459508000Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 30 12:42:35.459559 containerd[2700]: time="2025-04-30T12:42:35.459547040Z" level=info msg="metadata content store policy set" policy=shared Apr 30 12:42:35.465992 containerd[2700]: time="2025-04-30T12:42:35.465970080Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 30 12:42:35.466031 containerd[2700]: time="2025-04-30T12:42:35.466007040Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 30 12:42:35.466031 containerd[2700]: time="2025-04-30T12:42:35.466021840Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Apr 30 12:42:35.466105 containerd[2700]: time="2025-04-30T12:42:35.466039320Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Apr 30 12:42:35.466105 containerd[2700]: time="2025-04-30T12:42:35.466052680Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 30 12:42:35.466191 containerd[2700]: time="2025-04-30T12:42:35.466178040Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 30 12:42:35.466407 containerd[2700]: time="2025-04-30T12:42:35.466393920Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 30 12:42:35.466936 containerd[2700]: time="2025-04-30T12:42:35.466640080Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Apr 30 12:42:35.467095 containerd[2700]: time="2025-04-30T12:42:35.467080480Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Apr 30 12:42:35.467116 containerd[2700]: time="2025-04-30T12:42:35.467101760Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Apr 30 12:42:35.467133 containerd[2700]: time="2025-04-30T12:42:35.467119240Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467158 containerd[2700]: time="2025-04-30T12:42:35.467134080Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467158 containerd[2700]: time="2025-04-30T12:42:35.467147560Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467191 containerd[2700]: time="2025-04-30T12:42:35.467160800Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467191 containerd[2700]: time="2025-04-30T12:42:35.467174520Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467191 containerd[2700]: time="2025-04-30T12:42:35.467187040Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467236 containerd[2700]: time="2025-04-30T12:42:35.467199600Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467236 containerd[2700]: time="2025-04-30T12:42:35.467210760Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 30 12:42:35.467236 containerd[2700]: time="2025-04-30T12:42:35.467230360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467285 containerd[2700]: time="2025-04-30T12:42:35.467244560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467285 containerd[2700]: time="2025-04-30T12:42:35.467257480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467285 containerd[2700]: time="2025-04-30T12:42:35.467270240Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467285 containerd[2700]: time="2025-04-30T12:42:35.467281640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467354 containerd[2700]: time="2025-04-30T12:42:35.467293440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467354 containerd[2700]: time="2025-04-30T12:42:35.467304560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467354 containerd[2700]: time="2025-04-30T12:42:35.467316840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467354 containerd[2700]: time="2025-04-30T12:42:35.467328920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467354 containerd[2700]: time="2025-04-30T12:42:35.467344080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467435 containerd[2700]: time="2025-04-30T12:42:35.467355640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467435 containerd[2700]: time="2025-04-30T12:42:35.467369760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467435 containerd[2700]: time="2025-04-30T12:42:35.467382160Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467435 containerd[2700]: time="2025-04-30T12:42:35.467396880Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Apr 30 12:42:35.467500 containerd[2700]: time="2025-04-30T12:42:35.467453760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467500 containerd[2700]: time="2025-04-30T12:42:35.467467760Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467500 containerd[2700]: time="2025-04-30T12:42:35.467478240Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 30 12:42:35.467658 containerd[2700]: time="2025-04-30T12:42:35.467646040Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 30 12:42:35.467679 containerd[2700]: time="2025-04-30T12:42:35.467662920Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Apr 30 12:42:35.467679 containerd[2700]: time="2025-04-30T12:42:35.467672880Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 30 12:42:35.467713 containerd[2700]: time="2025-04-30T12:42:35.467684840Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Apr 30 12:42:35.467713 containerd[2700]: time="2025-04-30T12:42:35.467694800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.467713 containerd[2700]: time="2025-04-30T12:42:35.467706880Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Apr 30 12:42:35.467767 containerd[2700]: time="2025-04-30T12:42:35.467717320Z" level=info msg="NRI interface is disabled by configuration." Apr 30 12:42:35.467767 containerd[2700]: time="2025-04-30T12:42:35.467727640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 30 12:42:35.468115 containerd[2700]: time="2025-04-30T12:42:35.468054040Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 30 12:42:35.468211 containerd[2700]: time="2025-04-30T12:42:35.468123480Z" level=info msg="Connect containerd service" Apr 30 12:42:35.468211 containerd[2700]: time="2025-04-30T12:42:35.468151880Z" level=info msg="using legacy CRI server" Apr 30 12:42:35.468211 containerd[2700]: time="2025-04-30T12:42:35.468157480Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Apr 30 12:42:35.468397 containerd[2700]: time="2025-04-30T12:42:35.468384080Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 30 12:42:35.469018 containerd[2700]: time="2025-04-30T12:42:35.468996160Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 30 12:42:35.469219 containerd[2700]: time="2025-04-30T12:42:35.469183280Z" level=info msg="Start subscribing containerd event" Apr 30 12:42:35.469246 containerd[2700]: time="2025-04-30T12:42:35.469235320Z" level=info msg="Start recovering state" Apr 30 12:42:35.469317 containerd[2700]: time="2025-04-30T12:42:35.469305480Z" level=info msg="Start event monitor" Apr 30 12:42:35.469341 containerd[2700]: time="2025-04-30T12:42:35.469319680Z" level=info msg="Start snapshots syncer" Apr 30 12:42:35.469341 containerd[2700]: time="2025-04-30T12:42:35.469331040Z" level=info msg="Start cni network conf syncer for default" Apr 30 12:42:35.469341 containerd[2700]: time="2025-04-30T12:42:35.469338280Z" level=info msg="Start streaming server" Apr 30 12:42:35.469593 containerd[2700]: time="2025-04-30T12:42:35.469580000Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 30 12:42:35.469628 containerd[2700]: time="2025-04-30T12:42:35.469620160Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 30 12:42:35.469682 containerd[2700]: time="2025-04-30T12:42:35.469673200Z" level=info msg="containerd successfully booted in 0.035654s" Apr 30 12:42:35.469722 systemd[1]: Started containerd.service - containerd container runtime. Apr 30 12:42:35.597942 tar[2698]: linux-arm64/LICENSE Apr 30 12:42:35.598046 tar[2698]: linux-arm64/README.md Apr 30 12:42:35.609427 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Apr 30 12:42:35.712963 kernel: EXT4-fs (nvme0n1p9): resized filesystem to 233815889 Apr 30 12:42:35.728953 extend-filesystems[2686]: Filesystem at /dev/nvme0n1p9 is mounted on /; on-line resizing required Apr 30 12:42:35.728953 extend-filesystems[2686]: old_desc_blocks = 1, new_desc_blocks = 112 Apr 30 12:42:35.728953 extend-filesystems[2686]: The filesystem on /dev/nvme0n1p9 is now 233815889 (4k) blocks long. Apr 30 12:42:35.757402 extend-filesystems[2665]: Resized filesystem in /dev/nvme0n1p9 Apr 30 12:42:35.731392 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 30 12:42:35.731693 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Apr 30 12:42:35.743984 systemd[1]: extend-filesystems.service: Consumed 205ms CPU time, 68.9M memory peak. Apr 30 12:42:36.093642 sshd_keygen[2689]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 30 12:42:36.111749 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Apr 30 12:42:36.131244 systemd[1]: Starting issuegen.service - Generate /run/issue... Apr 30 12:42:36.140252 systemd[1]: issuegen.service: Deactivated successfully. Apr 30 12:42:36.141063 systemd[1]: Finished issuegen.service - Generate /run/issue. Apr 30 12:42:36.142042 coreos-metadata[2660]: Apr 30 12:42:36.142 INFO Fetching https://metadata.packet.net/metadata: Attempt #2 Apr 30 12:42:36.142432 coreos-metadata[2660]: Apr 30 12:42:36.142 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Apr 30 12:42:36.147695 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Apr 30 12:42:36.161023 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Apr 30 12:42:36.167386 systemd[1]: Started getty@tty1.service - Getty on tty1. Apr 30 12:42:36.173375 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Apr 30 12:42:36.178533 systemd[1]: Reached target getty.target - Login Prompts. Apr 30 12:42:36.294965 kernel: mlx5_core 0001:01:00.0 enP1p1s0f0np0: Link up Apr 30 12:42:36.310957 kernel: bond0: (slave enP1p1s0f0np0): Enslaving as a backup interface with an up link Apr 30 12:42:36.316521 systemd-networkd[2606]: enP1p1s0f1np1: Configuring with /etc/systemd/network/10-0c:42:a1:49:ce:01.network. Apr 30 12:42:36.402463 coreos-metadata[2748]: Apr 30 12:42:36.402 INFO Fetching https://metadata.packet.net/metadata: Attempt #2 Apr 30 12:42:36.402969 coreos-metadata[2748]: Apr 30 12:42:36.402 INFO Failed to fetch: error sending request for url (https://metadata.packet.net/metadata) Apr 30 12:42:36.924963 kernel: mlx5_core 0001:01:00.1 enP1p1s0f1np1: Link up Apr 30 12:42:36.941958 kernel: bond0: (slave enP1p1s0f1np1): Enslaving as a backup interface with an up link Apr 30 12:42:36.942174 systemd-networkd[2606]: bond0: Configuring with /etc/systemd/network/05-bond0.network. Apr 30 12:42:36.943350 systemd-networkd[2606]: enP1p1s0f0np0: Link UP Apr 30 12:42:36.943597 systemd-networkd[2606]: enP1p1s0f0np0: Gained carrier Apr 30 12:42:36.944402 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Apr 30 12:42:36.961969 kernel: bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond Apr 30 12:42:36.973387 systemd-networkd[2606]: enP1p1s0f1np1: Reconfiguring with /etc/systemd/network/10-0c:42:a1:49:ce:00.network. Apr 30 12:42:36.973681 systemd-networkd[2606]: enP1p1s0f1np1: Link UP Apr 30 12:42:36.973886 systemd-networkd[2606]: enP1p1s0f1np1: Gained carrier Apr 30 12:42:36.994241 systemd-networkd[2606]: bond0: Link UP Apr 30 12:42:36.994504 systemd-networkd[2606]: bond0: Gained carrier Apr 30 12:42:36.994676 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:36.995299 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:36.995551 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:36.995691 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:37.068443 kernel: bond0: (slave enP1p1s0f0np0): link status definitely up, 25000 Mbps full duplex Apr 30 12:42:37.068476 kernel: bond0: active interface up! Apr 30 12:42:37.191961 kernel: bond0: (slave enP1p1s0f1np1): link status definitely up, 25000 Mbps full duplex Apr 30 12:42:38.049011 systemd-networkd[2606]: bond0: Gained IPv6LL Apr 30 12:42:38.049448 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:38.142540 coreos-metadata[2660]: Apr 30 12:42:38.142 INFO Fetching https://metadata.packet.net/metadata: Attempt #3 Apr 30 12:42:38.403050 coreos-metadata[2748]: Apr 30 12:42:38.403 INFO Fetching https://metadata.packet.net/metadata: Attempt #3 Apr 30 12:42:38.625356 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:38.625475 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:38.628061 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Apr 30 12:42:38.634115 systemd[1]: Reached target network-online.target - Network is Online. Apr 30 12:42:38.653158 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:42:38.659843 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Apr 30 12:42:38.683778 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Apr 30 12:42:39.235128 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:42:39.241283 (kubelet)[2818]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 12:42:39.677121 kubelet[2818]: E0430 12:42:39.677092 2818 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 12:42:39.679308 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 12:42:39.679453 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 12:42:39.679769 systemd[1]: kubelet.service: Consumed 694ms CPU time, 245.9M memory peak. Apr 30 12:42:40.483858 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Apr 30 12:42:40.490217 systemd[1]: Started sshd@0-147.75.61.11:22-139.178.68.195:56146.service - OpenSSH per-connection server daemon (139.178.68.195:56146). Apr 30 12:42:40.719269 kernel: mlx5_core 0001:01:00.0: lag map: port 1:1 port 2:2 Apr 30 12:42:40.719451 kernel: mlx5_core 0001:01:00.0: shared_fdb:0 mode:queue_affinity Apr 30 12:42:40.748498 coreos-metadata[2660]: Apr 30 12:42:40.748 INFO Fetch successful Apr 30 12:42:40.817026 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Apr 30 12:42:40.824083 systemd[1]: Starting packet-phone-home.service - Report Success to Packet... Apr 30 12:42:40.902593 sshd[2839]: Accepted publickey for core from 139.178.68.195 port 56146 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:40.904145 sshd-session[2839]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:40.913668 systemd-logind[2684]: New session 1 of user core. Apr 30 12:42:40.915136 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Apr 30 12:42:40.932287 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Apr 30 12:42:40.941616 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Apr 30 12:42:40.948844 systemd[1]: Starting user@500.service - User Manager for UID 500... Apr 30 12:42:40.957228 (systemd)[2852]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 30 12:42:40.959316 systemd-logind[2684]: New session c1 of user core. Apr 30 12:42:41.076204 systemd[2852]: Queued start job for default target default.target. Apr 30 12:42:41.086123 systemd[2852]: Created slice app.slice - User Application Slice. Apr 30 12:42:41.086150 systemd[2852]: Reached target paths.target - Paths. Apr 30 12:42:41.086182 systemd[2852]: Reached target timers.target - Timers. Apr 30 12:42:41.087496 systemd[2852]: Starting dbus.socket - D-Bus User Message Bus Socket... Apr 30 12:42:41.095884 systemd[2852]: Listening on dbus.socket - D-Bus User Message Bus Socket. Apr 30 12:42:41.095936 systemd[2852]: Reached target sockets.target - Sockets. Apr 30 12:42:41.096012 systemd[2852]: Reached target basic.target - Basic System. Apr 30 12:42:41.096041 systemd[2852]: Reached target default.target - Main User Target. Apr 30 12:42:41.096062 systemd[2852]: Startup finished in 132ms. Apr 30 12:42:41.096283 systemd[1]: Started user@500.service - User Manager for UID 500. Apr 30 12:42:41.102599 systemd[1]: Started session-1.scope - Session 1 of User core. Apr 30 12:42:41.212764 login[2792]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:41.213702 login[2793]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:41.215876 systemd-logind[2684]: New session 3 of user core. Apr 30 12:42:41.217333 systemd[1]: Started session-3.scope - Session 3 of User core. Apr 30 12:42:41.219265 systemd-logind[2684]: New session 2 of user core. Apr 30 12:42:41.220574 systemd[1]: Started session-2.scope - Session 2 of User core. Apr 30 12:42:41.405866 systemd[1]: Started sshd@1-147.75.61.11:22-139.178.68.195:56162.service - OpenSSH per-connection server daemon (139.178.68.195:56162). Apr 30 12:42:41.517805 coreos-metadata[2748]: Apr 30 12:42:41.517 INFO Fetch successful Apr 30 12:42:41.574554 unknown[2748]: wrote ssh authorized keys file for user: core Apr 30 12:42:41.604021 update-ssh-keys[2892]: Updated "/home/core/.ssh/authorized_keys" Apr 30 12:42:41.605205 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Apr 30 12:42:41.606806 systemd[1]: Finished sshkeys.service. Apr 30 12:42:41.739036 systemd[1]: Finished packet-phone-home.service - Report Success to Packet. Apr 30 12:42:41.739458 systemd[1]: Reached target multi-user.target - Multi-User System. Apr 30 12:42:41.742993 systemd[1]: Startup finished in 3.224s (kernel) + 20.038s (initrd) + 10.349s (userspace) = 33.613s. Apr 30 12:42:41.816065 sshd[2889]: Accepted publickey for core from 139.178.68.195 port 56162 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:41.817167 sshd-session[2889]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:41.820127 systemd-logind[2684]: New session 4 of user core. Apr 30 12:42:41.832050 systemd[1]: Started session-4.scope - Session 4 of User core. Apr 30 12:42:42.109918 sshd[2896]: Connection closed by 139.178.68.195 port 56162 Apr 30 12:42:42.110526 sshd-session[2889]: pam_unix(sshd:session): session closed for user core Apr 30 12:42:42.114049 systemd[1]: sshd@1-147.75.61.11:22-139.178.68.195:56162.service: Deactivated successfully. Apr 30 12:42:42.115806 systemd[1]: session-4.scope: Deactivated successfully. Apr 30 12:42:42.116392 systemd-logind[2684]: Session 4 logged out. Waiting for processes to exit. Apr 30 12:42:42.116928 systemd-logind[2684]: Removed session 4. Apr 30 12:42:42.180693 systemd[1]: Started sshd@2-147.75.61.11:22-139.178.68.195:56178.service - OpenSSH per-connection server daemon (139.178.68.195:56178). Apr 30 12:42:42.590202 sshd[2902]: Accepted publickey for core from 139.178.68.195 port 56178 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:42.591292 sshd-session[2902]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:42.594072 systemd-logind[2684]: New session 5 of user core. Apr 30 12:42:42.614119 systemd[1]: Started session-5.scope - Session 5 of User core. Apr 30 12:42:42.697569 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:42.882644 sshd[2904]: Connection closed by 139.178.68.195 port 56178 Apr 30 12:42:42.883268 sshd-session[2902]: pam_unix(sshd:session): session closed for user core Apr 30 12:42:42.886721 systemd[1]: sshd@2-147.75.61.11:22-139.178.68.195:56178.service: Deactivated successfully. Apr 30 12:42:42.888403 systemd[1]: session-5.scope: Deactivated successfully. Apr 30 12:42:42.888917 systemd-logind[2684]: Session 5 logged out. Waiting for processes to exit. Apr 30 12:42:42.889459 systemd-logind[2684]: Removed session 5. Apr 30 12:42:42.954434 systemd[1]: Started sshd@3-147.75.61.11:22-139.178.68.195:56192.service - OpenSSH per-connection server daemon (139.178.68.195:56192). Apr 30 12:42:43.363655 sshd[2910]: Accepted publickey for core from 139.178.68.195 port 56192 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:43.364678 sshd-session[2910]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:43.367691 systemd-logind[2684]: New session 6 of user core. Apr 30 12:42:43.383062 systemd[1]: Started session-6.scope - Session 6 of User core. Apr 30 12:42:43.658677 sshd[2912]: Connection closed by 139.178.68.195 port 56192 Apr 30 12:42:43.659123 sshd-session[2910]: pam_unix(sshd:session): session closed for user core Apr 30 12:42:43.662582 systemd[1]: sshd@3-147.75.61.11:22-139.178.68.195:56192.service: Deactivated successfully. Apr 30 12:42:43.664279 systemd[1]: session-6.scope: Deactivated successfully. Apr 30 12:42:43.664841 systemd-logind[2684]: Session 6 logged out. Waiting for processes to exit. Apr 30 12:42:43.665408 systemd-logind[2684]: Removed session 6. Apr 30 12:42:43.731652 systemd[1]: Started sshd@4-147.75.61.11:22-139.178.68.195:56206.service - OpenSSH per-connection server daemon (139.178.68.195:56206). Apr 30 12:42:44.139249 sshd[2918]: Accepted publickey for core from 139.178.68.195 port 56206 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:44.140673 sshd-session[2918]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:44.143800 systemd-logind[2684]: New session 7 of user core. Apr 30 12:42:44.161061 systemd[1]: Started session-7.scope - Session 7 of User core. Apr 30 12:42:44.390557 sudo[2921]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Apr 30 12:42:44.390827 sudo[2921]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 12:42:44.405809 sudo[2921]: pam_unix(sudo:session): session closed for user root Apr 30 12:42:44.468556 sshd[2920]: Connection closed by 139.178.68.195 port 56206 Apr 30 12:42:44.469090 sshd-session[2918]: pam_unix(sshd:session): session closed for user core Apr 30 12:42:44.472863 systemd[1]: sshd@4-147.75.61.11:22-139.178.68.195:56206.service: Deactivated successfully. Apr 30 12:42:44.476367 systemd[1]: session-7.scope: Deactivated successfully. Apr 30 12:42:44.476908 systemd-logind[2684]: Session 7 logged out. Waiting for processes to exit. Apr 30 12:42:44.477500 systemd-logind[2684]: Removed session 7. Apr 30 12:42:44.540789 systemd[1]: Started sshd@5-147.75.61.11:22-139.178.68.195:56210.service - OpenSSH per-connection server daemon (139.178.68.195:56210). Apr 30 12:42:44.950993 sshd[2928]: Accepted publickey for core from 139.178.68.195 port 56210 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:44.952115 sshd-session[2928]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:44.955069 systemd-logind[2684]: New session 8 of user core. Apr 30 12:42:44.965124 systemd[1]: Started session-8.scope - Session 8 of User core. Apr 30 12:42:45.185230 sudo[2932]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Apr 30 12:42:45.185503 sudo[2932]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 12:42:45.188062 sudo[2932]: pam_unix(sudo:session): session closed for user root Apr 30 12:42:45.192323 sudo[2931]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Apr 30 12:42:45.192576 sudo[2931]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 12:42:45.214313 systemd[1]: Starting audit-rules.service - Load Audit Rules... Apr 30 12:42:45.236419 augenrules[2954]: No rules Apr 30 12:42:45.237552 systemd[1]: audit-rules.service: Deactivated successfully. Apr 30 12:42:45.237785 systemd[1]: Finished audit-rules.service - Load Audit Rules. Apr 30 12:42:45.238514 sudo[2931]: pam_unix(sudo:session): session closed for user root Apr 30 12:42:45.300937 sshd[2930]: Connection closed by 139.178.68.195 port 56210 Apr 30 12:42:45.301350 sshd-session[2928]: pam_unix(sshd:session): session closed for user core Apr 30 12:42:45.304204 systemd[1]: sshd@5-147.75.61.11:22-139.178.68.195:56210.service: Deactivated successfully. Apr 30 12:42:45.306219 systemd[1]: session-8.scope: Deactivated successfully. Apr 30 12:42:45.306759 systemd-logind[2684]: Session 8 logged out. Waiting for processes to exit. Apr 30 12:42:45.307322 systemd-logind[2684]: Removed session 8. Apr 30 12:42:45.376604 systemd[1]: Started sshd@6-147.75.61.11:22-139.178.68.195:56222.service - OpenSSH per-connection server daemon (139.178.68.195:56222). Apr 30 12:42:45.786440 sshd[2963]: Accepted publickey for core from 139.178.68.195 port 56222 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:42:45.787464 sshd-session[2963]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:42:45.790316 systemd-logind[2684]: New session 9 of user core. Apr 30 12:42:45.802056 systemd[1]: Started session-9.scope - Session 9 of User core. Apr 30 12:42:46.020704 sudo[2966]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 30 12:42:46.020989 sudo[2966]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Apr 30 12:42:46.303166 systemd[1]: Starting docker.service - Docker Application Container Engine... Apr 30 12:42:46.303353 (dockerd)[2997]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Apr 30 12:42:46.504695 dockerd[2997]: time="2025-04-30T12:42:46.504649920Z" level=info msg="Starting up" Apr 30 12:42:46.572546 dockerd[2997]: time="2025-04-30T12:42:46.572484080Z" level=info msg="Loading containers: start." Apr 30 12:42:46.708972 kernel: Initializing XFRM netlink socket Apr 30 12:42:46.727060 systemd-timesyncd[2608]: Network configuration changed, trying to establish connection. Apr 30 12:42:46.776352 systemd-networkd[2606]: docker0: Link UP Apr 30 12:42:46.825156 dockerd[2997]: time="2025-04-30T12:42:46.825092720Z" level=info msg="Loading containers: done." Apr 30 12:42:46.834884 dockerd[2997]: time="2025-04-30T12:42:46.834844360Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 30 12:42:46.834973 dockerd[2997]: time="2025-04-30T12:42:46.834922560Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Apr 30 12:42:46.835115 dockerd[2997]: time="2025-04-30T12:42:46.835098000Z" level=info msg="Daemon has completed initialization" Apr 30 12:42:46.854131 dockerd[2997]: time="2025-04-30T12:42:46.854085600Z" level=info msg="API listen on /run/docker.sock" Apr 30 12:42:46.854232 systemd[1]: Started docker.service - Docker Application Container Engine. Apr 30 12:42:46.296051 systemd-resolved[2607]: Clock change detected. Flushing caches. Apr 30 12:42:46.303840 systemd-journald[2185]: Time jumped backwards, rotating. Apr 30 12:42:46.296205 systemd-timesyncd[2608]: Contacted time server [2606:6680:8:1::d14e:69b8]:123 (2.flatcar.pool.ntp.org). Apr 30 12:42:46.296252 systemd-timesyncd[2608]: Initial clock synchronization to Wed 2025-04-30 12:42:46.296000 UTC. Apr 30 12:42:46.883382 containerd[2700]: time="2025-04-30T12:42:46.883351617Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.8\"" Apr 30 12:42:47.544515 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount459574965.mount: Deactivated successfully. Apr 30 12:42:49.279938 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 30 12:42:49.288958 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:42:49.390880 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:42:49.394269 (kubelet)[3305]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 12:42:49.400386 containerd[2700]: time="2025-04-30T12:42:49.400352257Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:49.400550 containerd[2700]: time="2025-04-30T12:42:49.400376217Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.8: active requests=0, bytes read=25554608" Apr 30 12:42:49.401475 containerd[2700]: time="2025-04-30T12:42:49.401451057Z" level=info msg="ImageCreate event name:\"sha256:ef8fb1ea7c9599dbedea6f9d5589975ebc5bf4ec72f6be6acaaec59a723a09b3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:49.404356 containerd[2700]: time="2025-04-30T12:42:49.404332897Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:30090db6a7d53799163ce82dae9e8ddb645fd47db93f2ec9da0cc787fd825625\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:49.405443 containerd[2700]: time="2025-04-30T12:42:49.405419177Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.8\" with image id \"sha256:ef8fb1ea7c9599dbedea6f9d5589975ebc5bf4ec72f6be6acaaec59a723a09b3\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.8\", repo digest \"registry.k8s.io/kube-apiserver@sha256:30090db6a7d53799163ce82dae9e8ddb645fd47db93f2ec9da0cc787fd825625\", size \"25551408\" in 2.52202996s" Apr 30 12:42:49.405466 containerd[2700]: time="2025-04-30T12:42:49.405453297Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.8\" returns image reference \"sha256:ef8fb1ea7c9599dbedea6f9d5589975ebc5bf4ec72f6be6acaaec59a723a09b3\"" Apr 30 12:42:49.405961 containerd[2700]: time="2025-04-30T12:42:49.405940497Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.8\"" Apr 30 12:42:49.427298 kubelet[3305]: E0430 12:42:49.427268 3305 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 12:42:49.430119 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 12:42:49.430268 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 12:42:49.430609 systemd[1]: kubelet.service: Consumed 132ms CPU time, 115.8M memory peak. Apr 30 12:42:50.552863 containerd[2700]: time="2025-04-30T12:42:50.552827337Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:50.553108 containerd[2700]: time="2025-04-30T12:42:50.552894737Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.8: active requests=0, bytes read=22458978" Apr 30 12:42:50.553883 containerd[2700]: time="2025-04-30T12:42:50.553860257Z" level=info msg="ImageCreate event name:\"sha256:ea6e6085feca75547d0422ab0536fe0d18c9ff5831de7a9d6a707c968027bb6a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:50.556515 containerd[2700]: time="2025-04-30T12:42:50.556489737Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:29eaddc64792a689df48506e78bbc641d063ac8bb92d2e66ae2ad05977420747\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:50.557688 containerd[2700]: time="2025-04-30T12:42:50.557657217Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.8\" with image id \"sha256:ea6e6085feca75547d0422ab0536fe0d18c9ff5831de7a9d6a707c968027bb6a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.8\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:29eaddc64792a689df48506e78bbc641d063ac8bb92d2e66ae2ad05977420747\", size \"23900539\" in 1.15166944s" Apr 30 12:42:50.557723 containerd[2700]: time="2025-04-30T12:42:50.557694137Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.8\" returns image reference \"sha256:ea6e6085feca75547d0422ab0536fe0d18c9ff5831de7a9d6a707c968027bb6a\"" Apr 30 12:42:50.558060 containerd[2700]: time="2025-04-30T12:42:50.558042217Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.8\"" Apr 30 12:42:51.731804 containerd[2700]: time="2025-04-30T12:42:51.731756137Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.8: active requests=0, bytes read=17125813" Apr 30 12:42:51.731804 containerd[2700]: time="2025-04-30T12:42:51.731764297Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:51.732843 containerd[2700]: time="2025-04-30T12:42:51.732816777Z" level=info msg="ImageCreate event name:\"sha256:1d2db6ef0dd2f3e08bdfcd46afde7b755b05192841f563d8df54b807daaa7d8d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:51.737035 containerd[2700]: time="2025-04-30T12:42:51.737009697Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:22994a2632e81059720480b9f6bdeb133b08d58492d0b36dfd6e9768b159b22a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:51.738318 containerd[2700]: time="2025-04-30T12:42:51.738290417Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.8\" with image id \"sha256:1d2db6ef0dd2f3e08bdfcd46afde7b755b05192841f563d8df54b807daaa7d8d\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.8\", repo digest \"registry.k8s.io/kube-scheduler@sha256:22994a2632e81059720480b9f6bdeb133b08d58492d0b36dfd6e9768b159b22a\", size \"18567392\" in 1.18021664s" Apr 30 12:42:51.738341 containerd[2700]: time="2025-04-30T12:42:51.738325377Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.8\" returns image reference \"sha256:1d2db6ef0dd2f3e08bdfcd46afde7b755b05192841f563d8df54b807daaa7d8d\"" Apr 30 12:42:51.738670 containerd[2700]: time="2025-04-30T12:42:51.738649577Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.8\"" Apr 30 12:42:52.527389 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2200932357.mount: Deactivated successfully. Apr 30 12:42:52.881977 containerd[2700]: time="2025-04-30T12:42:52.881907697Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:52.882212 containerd[2700]: time="2025-04-30T12:42:52.881968697Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.8: active requests=0, bytes read=26871917" Apr 30 12:42:52.882662 containerd[2700]: time="2025-04-30T12:42:52.882637657Z" level=info msg="ImageCreate event name:\"sha256:c5361ece77e80334cd5fb082c0b678cb3244f5834ecacea1719ae6b38b465581\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:52.884358 containerd[2700]: time="2025-04-30T12:42:52.884334617Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:dd0c9a37670f209947b1ed880f06a2e93e1d41da78c037f52f94b13858769838\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:52.885153 containerd[2700]: time="2025-04-30T12:42:52.885127017Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.8\" with image id \"sha256:c5361ece77e80334cd5fb082c0b678cb3244f5834ecacea1719ae6b38b465581\", repo tag \"registry.k8s.io/kube-proxy:v1.31.8\", repo digest \"registry.k8s.io/kube-proxy@sha256:dd0c9a37670f209947b1ed880f06a2e93e1d41da78c037f52f94b13858769838\", size \"26870936\" in 1.14644532s" Apr 30 12:42:52.885186 containerd[2700]: time="2025-04-30T12:42:52.885157977Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.8\" returns image reference \"sha256:c5361ece77e80334cd5fb082c0b678cb3244f5834ecacea1719ae6b38b465581\"" Apr 30 12:42:52.885514 containerd[2700]: time="2025-04-30T12:42:52.885491697Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Apr 30 12:42:53.364576 systemd[1]: Started sshd@7-147.75.61.11:22-218.92.0.158:25852.service - OpenSSH per-connection server daemon (218.92.0.158:25852). Apr 30 12:42:53.407877 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1748969162.mount: Deactivated successfully. Apr 30 12:42:54.267075 containerd[2700]: time="2025-04-30T12:42:54.267022057Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485381" Apr 30 12:42:54.267075 containerd[2700]: time="2025-04-30T12:42:54.267027817Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:54.268109 containerd[2700]: time="2025-04-30T12:42:54.268091057Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:54.270980 containerd[2700]: time="2025-04-30T12:42:54.270953337Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:54.272057 containerd[2700]: time="2025-04-30T12:42:54.272026817Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.386501s" Apr 30 12:42:54.272086 containerd[2700]: time="2025-04-30T12:42:54.272062777Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Apr 30 12:42:54.272411 containerd[2700]: time="2025-04-30T12:42:54.272394417Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Apr 30 12:42:54.486270 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount209523749.mount: Deactivated successfully. Apr 30 12:42:54.486668 containerd[2700]: time="2025-04-30T12:42:54.486643857Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:54.486707 containerd[2700]: time="2025-04-30T12:42:54.486677817Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Apr 30 12:42:54.487473 containerd[2700]: time="2025-04-30T12:42:54.487454137Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:54.489427 containerd[2700]: time="2025-04-30T12:42:54.489406017Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:54.490195 containerd[2700]: time="2025-04-30T12:42:54.490177697Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 217.75848ms" Apr 30 12:42:54.490228 containerd[2700]: time="2025-04-30T12:42:54.490199337Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Apr 30 12:42:54.490478 containerd[2700]: time="2025-04-30T12:42:54.490458697Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Apr 30 12:42:54.706815 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount107004191.mount: Deactivated successfully. Apr 30 12:42:57.297067 containerd[2700]: time="2025-04-30T12:42:57.297022297Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:57.297385 containerd[2700]: time="2025-04-30T12:42:57.297040697Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406465" Apr 30 12:42:57.298321 containerd[2700]: time="2025-04-30T12:42:57.298295617Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:57.301647 containerd[2700]: time="2025-04-30T12:42:57.301601697Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:42:57.303002 containerd[2700]: time="2025-04-30T12:42:57.302975377Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 2.81248772s" Apr 30 12:42:57.303025 containerd[2700]: time="2025-04-30T12:42:57.303009537Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Apr 30 12:42:57.813075 systemd[1]: Started sshd@8-147.75.61.11:22-80.94.95.115:53046.service - OpenSSH per-connection server daemon (80.94.95.115:53046). Apr 30 12:42:59.529928 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 30 12:42:59.539084 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:42:59.629605 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:42:59.632967 (kubelet)[3539]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Apr 30 12:42:59.664016 kubelet[3539]: E0430 12:42:59.663981 3539 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 30 12:42:59.666106 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 30 12:42:59.666246 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 30 12:42:59.666573 systemd[1]: kubelet.service: Consumed 127ms CPU time, 110.1M memory peak. Apr 30 12:43:00.806082 sshd[3524]: Invalid user app from 80.94.95.115 port 53046 Apr 30 12:43:01.051982 sshd[3524]: Connection closed by invalid user app 80.94.95.115 port 53046 [preauth] Apr 30 12:43:01.053688 systemd[1]: sshd@8-147.75.61.11:22-80.94.95.115:53046.service: Deactivated successfully. Apr 30 12:43:02.479631 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:43:02.479865 systemd[1]: kubelet.service: Consumed 127ms CPU time, 110.1M memory peak. Apr 30 12:43:02.494091 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:43:02.513247 systemd[1]: Reload requested from client PID 3577 ('systemctl') (unit session-9.scope)... Apr 30 12:43:02.513259 systemd[1]: Reloading... Apr 30 12:43:02.591749 zram_generator::config[3631]: No configuration found. Apr 30 12:43:02.682394 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 12:43:02.775411 systemd[1]: Reloading finished in 261 ms. Apr 30 12:43:02.820022 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:43:02.822706 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:43:02.823459 systemd[1]: kubelet.service: Deactivated successfully. Apr 30 12:43:02.823668 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:43:02.823704 systemd[1]: kubelet.service: Consumed 74ms CPU time, 82.5M memory peak. Apr 30 12:43:02.826952 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:43:02.918733 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:43:02.922205 (kubelet)[3696]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 30 12:43:02.957768 kubelet[3696]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 12:43:02.957768 kubelet[3696]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 30 12:43:02.957768 kubelet[3696]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 12:43:02.957958 kubelet[3696]: I0430 12:43:02.957886 3696 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 30 12:43:03.740594 kubelet[3696]: I0430 12:43:03.740560 3696 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Apr 30 12:43:03.740594 kubelet[3696]: I0430 12:43:03.740586 3696 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 30 12:43:03.740810 kubelet[3696]: I0430 12:43:03.740791 3696 server.go:929] "Client rotation is on, will bootstrap in background" Apr 30 12:43:03.764305 kubelet[3696]: E0430 12:43:03.764280 3696 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://147.75.61.11:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 147.75.61.11:6443: connect: connection refused" logger="UnhandledError" Apr 30 12:43:03.764877 kubelet[3696]: I0430 12:43:03.764854 3696 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 30 12:43:03.769919 kubelet[3696]: E0430 12:43:03.769896 3696 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 30 12:43:03.769945 kubelet[3696]: I0430 12:43:03.769921 3696 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 30 12:43:03.791537 kubelet[3696]: I0430 12:43:03.791510 3696 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 30 12:43:03.791820 kubelet[3696]: I0430 12:43:03.791804 3696 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Apr 30 12:43:03.791939 kubelet[3696]: I0430 12:43:03.791915 3696 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 30 12:43:03.792083 kubelet[3696]: I0430 12:43:03.791939 3696 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230.1.1-a-0ae8677ef9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 30 12:43:03.792225 kubelet[3696]: I0430 12:43:03.792214 3696 topology_manager.go:138] "Creating topology manager with none policy" Apr 30 12:43:03.792225 kubelet[3696]: I0430 12:43:03.792224 3696 container_manager_linux.go:300] "Creating device plugin manager" Apr 30 12:43:03.792404 kubelet[3696]: I0430 12:43:03.792393 3696 state_mem.go:36] "Initialized new in-memory state store" Apr 30 12:43:03.794233 kubelet[3696]: I0430 12:43:03.794217 3696 kubelet.go:408] "Attempting to sync node with API server" Apr 30 12:43:03.794252 kubelet[3696]: I0430 12:43:03.794241 3696 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 30 12:43:03.794268 kubelet[3696]: I0430 12:43:03.794260 3696 kubelet.go:314] "Adding apiserver pod source" Apr 30 12:43:03.794292 kubelet[3696]: I0430 12:43:03.794269 3696 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 30 12:43:03.795128 kubelet[3696]: W0430 12:43:03.795084 3696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://147.75.61.11:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 147.75.61.11:6443: connect: connection refused Apr 30 12:43:03.795164 kubelet[3696]: E0430 12:43:03.795149 3696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://147.75.61.11:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 147.75.61.11:6443: connect: connection refused" logger="UnhandledError" Apr 30 12:43:03.796025 kubelet[3696]: I0430 12:43:03.796015 3696 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Apr 30 12:43:03.796769 kubelet[3696]: W0430 12:43:03.796730 3696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://147.75.61.11:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.1.1-a-0ae8677ef9&limit=500&resourceVersion=0": dial tcp 147.75.61.11:6443: connect: connection refused Apr 30 12:43:03.796806 kubelet[3696]: E0430 12:43:03.796789 3696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://147.75.61.11:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.1.1-a-0ae8677ef9&limit=500&resourceVersion=0\": dial tcp 147.75.61.11:6443: connect: connection refused" logger="UnhandledError" Apr 30 12:43:03.797822 kubelet[3696]: I0430 12:43:03.797810 3696 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Apr 30 12:43:03.798551 kubelet[3696]: W0430 12:43:03.798539 3696 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 30 12:43:03.799196 kubelet[3696]: I0430 12:43:03.799183 3696 server.go:1269] "Started kubelet" Apr 30 12:43:03.799396 kubelet[3696]: I0430 12:43:03.799361 3696 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Apr 30 12:43:03.802547 kubelet[3696]: I0430 12:43:03.802421 3696 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 30 12:43:03.802837 kubelet[3696]: I0430 12:43:03.802815 3696 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 30 12:43:03.803873 kubelet[3696]: I0430 12:43:03.803854 3696 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 30 12:43:03.803899 kubelet[3696]: I0430 12:43:03.803860 3696 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 30 12:43:03.804070 kubelet[3696]: I0430 12:43:03.804044 3696 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 30 12:43:03.804126 kubelet[3696]: I0430 12:43:03.804093 3696 volume_manager.go:289] "Starting Kubelet Volume Manager" Apr 30 12:43:03.804352 kubelet[3696]: E0430 12:43:03.804315 3696 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230.1.1-a-0ae8677ef9\" not found" Apr 30 12:43:03.806516 kubelet[3696]: I0430 12:43:03.806505 3696 reconciler.go:26] "Reconciler: start to sync state" Apr 30 12:43:03.806762 kubelet[3696]: E0430 12:43:03.806723 3696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.75.61.11:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.1.1-a-0ae8677ef9?timeout=10s\": dial tcp 147.75.61.11:6443: connect: connection refused" interval="200ms" Apr 30 12:43:03.806889 kubelet[3696]: I0430 12:43:03.806872 3696 factory.go:221] Registration of the systemd container factory successfully Apr 30 12:43:03.806977 kubelet[3696]: W0430 12:43:03.806937 3696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://147.75.61.11:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 147.75.61.11:6443: connect: connection refused Apr 30 12:43:03.807006 kubelet[3696]: E0430 12:43:03.806992 3696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://147.75.61.11:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 147.75.61.11:6443: connect: connection refused" logger="UnhandledError" Apr 30 12:43:03.807006 kubelet[3696]: I0430 12:43:03.806964 3696 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 30 12:43:03.807331 kubelet[3696]: I0430 12:43:03.807313 3696 server.go:460] "Adding debug handlers to kubelet server" Apr 30 12:43:03.807516 kubelet[3696]: E0430 12:43:03.807498 3696 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 30 12:43:03.808223 kubelet[3696]: I0430 12:43:03.808207 3696 factory.go:221] Registration of the containerd container factory successfully Apr 30 12:43:03.808395 kubelet[3696]: E0430 12:43:03.807316 3696 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://147.75.61.11:6443/api/v1/namespaces/default/events\": dial tcp 147.75.61.11:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4230.1.1-a-0ae8677ef9.183b19322b25e0b9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230.1.1-a-0ae8677ef9,UID:ci-4230.1.1-a-0ae8677ef9,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230.1.1-a-0ae8677ef9,},FirstTimestamp:2025-04-30 12:43:03.799161017 +0000 UTC m=+0.874212641,LastTimestamp:2025-04-30 12:43:03.799161017 +0000 UTC m=+0.874212641,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230.1.1-a-0ae8677ef9,}" Apr 30 12:43:03.823848 kubelet[3696]: I0430 12:43:03.823829 3696 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 30 12:43:03.823848 kubelet[3696]: I0430 12:43:03.823842 3696 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 30 12:43:03.823923 kubelet[3696]: I0430 12:43:03.823859 3696 state_mem.go:36] "Initialized new in-memory state store" Apr 30 12:43:03.826258 kubelet[3696]: I0430 12:43:03.826243 3696 policy_none.go:49] "None policy: Start" Apr 30 12:43:03.826693 kubelet[3696]: I0430 12:43:03.826678 3696 memory_manager.go:170] "Starting memorymanager" policy="None" Apr 30 12:43:03.826717 kubelet[3696]: I0430 12:43:03.826701 3696 state_mem.go:35] "Initializing new in-memory state store" Apr 30 12:43:03.830453 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Apr 30 12:43:03.831325 kubelet[3696]: I0430 12:43:03.831286 3696 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 30 12:43:03.832831 kubelet[3696]: I0430 12:43:03.832806 3696 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 30 12:43:03.832854 kubelet[3696]: I0430 12:43:03.832838 3696 status_manager.go:217] "Starting to sync pod status with apiserver" Apr 30 12:43:03.832873 kubelet[3696]: I0430 12:43:03.832860 3696 kubelet.go:2321] "Starting kubelet main sync loop" Apr 30 12:43:03.832923 kubelet[3696]: E0430 12:43:03.832907 3696 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 30 12:43:03.833800 kubelet[3696]: W0430 12:43:03.833758 3696 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://147.75.61.11:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 147.75.61.11:6443: connect: connection refused Apr 30 12:43:03.833833 kubelet[3696]: E0430 12:43:03.833816 3696 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://147.75.61.11:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 147.75.61.11:6443: connect: connection refused" logger="UnhandledError" Apr 30 12:43:03.844088 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Apr 30 12:43:03.862007 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Apr 30 12:43:03.862816 kubelet[3696]: I0430 12:43:03.862794 3696 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 30 12:43:03.863011 kubelet[3696]: I0430 12:43:03.862997 3696 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 30 12:43:03.863048 kubelet[3696]: I0430 12:43:03.863011 3696 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 30 12:43:03.863172 kubelet[3696]: I0430 12:43:03.863156 3696 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 30 12:43:03.863801 kubelet[3696]: E0430 12:43:03.863785 3696 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4230.1.1-a-0ae8677ef9\" not found" Apr 30 12:43:03.939696 systemd[1]: Created slice kubepods-burstable-podfbb3dc52e7c4b56400cba46276b9d9f1.slice - libcontainer container kubepods-burstable-podfbb3dc52e7c4b56400cba46276b9d9f1.slice. Apr 30 12:43:03.962409 systemd[1]: Created slice kubepods-burstable-podab493c0aca086d6f4229e402c66068f4.slice - libcontainer container kubepods-burstable-podab493c0aca086d6f4229e402c66068f4.slice. Apr 30 12:43:03.965312 kubelet[3696]: I0430 12:43:03.965289 3696 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:03.965662 kubelet[3696]: E0430 12:43:03.965639 3696 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://147.75.61.11:6443/api/v1/nodes\": dial tcp 147.75.61.11:6443: connect: connection refused" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:03.978861 systemd[1]: Created slice kubepods-burstable-pod8696336ba6a673aa44d6b4010b4a1ffc.slice - libcontainer container kubepods-burstable-pod8696336ba6a673aa44d6b4010b4a1ffc.slice. Apr 30 12:43:04.007257 kubelet[3696]: I0430 12:43:04.007180 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-ca-certs\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007309 kubelet[3696]: I0430 12:43:04.007254 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8696336ba6a673aa44d6b4010b4a1ffc-ca-certs\") pod \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" (UID: \"8696336ba6a673aa44d6b4010b4a1ffc\") " pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007309 kubelet[3696]: I0430 12:43:04.007289 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8696336ba6a673aa44d6b4010b4a1ffc-k8s-certs\") pod \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" (UID: \"8696336ba6a673aa44d6b4010b4a1ffc\") " pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007388 kubelet[3696]: I0430 12:43:04.007329 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8696336ba6a673aa44d6b4010b4a1ffc-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" (UID: \"8696336ba6a673aa44d6b4010b4a1ffc\") " pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007388 kubelet[3696]: I0430 12:43:04.007360 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-flexvolume-dir\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007459 kubelet[3696]: I0430 12:43:04.007390 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-k8s-certs\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007459 kubelet[3696]: I0430 12:43:04.007419 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-kubeconfig\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007459 kubelet[3696]: E0430 12:43:04.007409 3696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.75.61.11:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.1.1-a-0ae8677ef9?timeout=10s\": dial tcp 147.75.61.11:6443: connect: connection refused" interval="400ms" Apr 30 12:43:04.007514 kubelet[3696]: I0430 12:43:04.007470 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.007514 kubelet[3696]: I0430 12:43:04.007498 3696 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ab493c0aca086d6f4229e402c66068f4-kubeconfig\") pod \"kube-scheduler-ci-4230.1.1-a-0ae8677ef9\" (UID: \"ab493c0aca086d6f4229e402c66068f4\") " pod="kube-system/kube-scheduler-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.167843 kubelet[3696]: I0430 12:43:04.167812 3696 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.168138 kubelet[3696]: E0430 12:43:04.168108 3696 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://147.75.61.11:6443/api/v1/nodes\": dial tcp 147.75.61.11:6443: connect: connection refused" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.261554 containerd[2700]: time="2025-04-30T12:43:04.261448897Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230.1.1-a-0ae8677ef9,Uid:fbb3dc52e7c4b56400cba46276b9d9f1,Namespace:kube-system,Attempt:0,}" Apr 30 12:43:04.278010 containerd[2700]: time="2025-04-30T12:43:04.277977977Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230.1.1-a-0ae8677ef9,Uid:ab493c0aca086d6f4229e402c66068f4,Namespace:kube-system,Attempt:0,}" Apr 30 12:43:04.281516 containerd[2700]: time="2025-04-30T12:43:04.281488217Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230.1.1-a-0ae8677ef9,Uid:8696336ba6a673aa44d6b4010b4a1ffc,Namespace:kube-system,Attempt:0,}" Apr 30 12:43:04.408700 kubelet[3696]: E0430 12:43:04.408666 3696 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://147.75.61.11:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.1.1-a-0ae8677ef9?timeout=10s\": dial tcp 147.75.61.11:6443: connect: connection refused" interval="800ms" Apr 30 12:43:04.515480 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1573623865.mount: Deactivated successfully. Apr 30 12:43:04.515889 containerd[2700]: time="2025-04-30T12:43:04.515861617Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 12:43:04.516651 containerd[2700]: time="2025-04-30T12:43:04.516632777Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 12:43:04.516759 containerd[2700]: time="2025-04-30T12:43:04.516719217Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Apr 30 12:43:04.517058 containerd[2700]: time="2025-04-30T12:43:04.517028897Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 30 12:43:04.517331 containerd[2700]: time="2025-04-30T12:43:04.517314057Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 12:43:04.517450 containerd[2700]: time="2025-04-30T12:43:04.517427257Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Apr 30 12:43:04.519407 containerd[2700]: time="2025-04-30T12:43:04.519382377Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 12:43:04.522057 containerd[2700]: time="2025-04-30T12:43:04.522033777Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 260.47432ms" Apr 30 12:43:04.522490 containerd[2700]: time="2025-04-30T12:43:04.522471657Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Apr 30 12:43:04.523256 containerd[2700]: time="2025-04-30T12:43:04.523237777Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 245.19244ms" Apr 30 12:43:04.525176 containerd[2700]: time="2025-04-30T12:43:04.525147217Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 243.60144ms" Apr 30 12:43:04.569950 kubelet[3696]: I0430 12:43:04.569927 3696 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.570239 kubelet[3696]: E0430 12:43:04.570214 3696 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://147.75.61.11:6443/api/v1/nodes\": dial tcp 147.75.61.11:6443: connect: connection refused" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:04.624613 containerd[2700]: time="2025-04-30T12:43:04.624538257Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:04.624613 containerd[2700]: time="2025-04-30T12:43:04.624605457Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:04.624649 containerd[2700]: time="2025-04-30T12:43:04.624616657Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:04.624710 containerd[2700]: time="2025-04-30T12:43:04.624692097Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:04.624973 containerd[2700]: time="2025-04-30T12:43:04.624923777Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:04.624999 containerd[2700]: time="2025-04-30T12:43:04.624969817Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:04.624999 containerd[2700]: time="2025-04-30T12:43:04.624982697Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:04.625068 containerd[2700]: time="2025-04-30T12:43:04.625049737Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:04.626355 containerd[2700]: time="2025-04-30T12:43:04.626304577Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:04.626375 containerd[2700]: time="2025-04-30T12:43:04.626356137Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:04.626395 containerd[2700]: time="2025-04-30T12:43:04.626369217Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:04.626460 containerd[2700]: time="2025-04-30T12:43:04.626442297Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:04.649944 systemd[1]: Started cri-containerd-4960db181540a83618f66e4f435b6ef83b88751c6315d8a3089704c0cea53199.scope - libcontainer container 4960db181540a83618f66e4f435b6ef83b88751c6315d8a3089704c0cea53199. Apr 30 12:43:04.651310 systemd[1]: Started cri-containerd-9aebc8ab622b5f77c59f6360e7e1493154fb4bff49bda37e4de0d0885828755d.scope - libcontainer container 9aebc8ab622b5f77c59f6360e7e1493154fb4bff49bda37e4de0d0885828755d. Apr 30 12:43:04.652660 systemd[1]: Started cri-containerd-c602eaaa0aa27b96bacf4a6c4a0fb1828c00b23963ef89d0ce72b856521438b3.scope - libcontainer container c602eaaa0aa27b96bacf4a6c4a0fb1828c00b23963ef89d0ce72b856521438b3. Apr 30 12:43:04.674057 containerd[2700]: time="2025-04-30T12:43:04.674024297Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230.1.1-a-0ae8677ef9,Uid:fbb3dc52e7c4b56400cba46276b9d9f1,Namespace:kube-system,Attempt:0,} returns sandbox id \"4960db181540a83618f66e4f435b6ef83b88751c6315d8a3089704c0cea53199\"" Apr 30 12:43:04.674361 containerd[2700]: time="2025-04-30T12:43:04.674338097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230.1.1-a-0ae8677ef9,Uid:ab493c0aca086d6f4229e402c66068f4,Namespace:kube-system,Attempt:0,} returns sandbox id \"9aebc8ab622b5f77c59f6360e7e1493154fb4bff49bda37e4de0d0885828755d\"" Apr 30 12:43:04.675969 containerd[2700]: time="2025-04-30T12:43:04.675946377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230.1.1-a-0ae8677ef9,Uid:8696336ba6a673aa44d6b4010b4a1ffc,Namespace:kube-system,Attempt:0,} returns sandbox id \"c602eaaa0aa27b96bacf4a6c4a0fb1828c00b23963ef89d0ce72b856521438b3\"" Apr 30 12:43:04.676964 containerd[2700]: time="2025-04-30T12:43:04.676944977Z" level=info msg="CreateContainer within sandbox \"9aebc8ab622b5f77c59f6360e7e1493154fb4bff49bda37e4de0d0885828755d\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 30 12:43:04.677036 containerd[2700]: time="2025-04-30T12:43:04.677010057Z" level=info msg="CreateContainer within sandbox \"4960db181540a83618f66e4f435b6ef83b88751c6315d8a3089704c0cea53199\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 30 12:43:04.678869 containerd[2700]: time="2025-04-30T12:43:04.678845777Z" level=info msg="CreateContainer within sandbox \"c602eaaa0aa27b96bacf4a6c4a0fb1828c00b23963ef89d0ce72b856521438b3\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 30 12:43:04.683329 containerd[2700]: time="2025-04-30T12:43:04.683304777Z" level=info msg="CreateContainer within sandbox \"9aebc8ab622b5f77c59f6360e7e1493154fb4bff49bda37e4de0d0885828755d\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"89aebaeee77087c645f45d44396dd0d70fc97e0ff8a44768239ff44784b8859b\"" Apr 30 12:43:04.683365 containerd[2700]: time="2025-04-30T12:43:04.683339137Z" level=info msg="CreateContainer within sandbox \"4960db181540a83618f66e4f435b6ef83b88751c6315d8a3089704c0cea53199\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"b400c4f8335607e499b76975f1c2d4a5cdd74586eb415c8945cf216103ccfbc4\"" Apr 30 12:43:04.683814 containerd[2700]: time="2025-04-30T12:43:04.683795977Z" level=info msg="StartContainer for \"b400c4f8335607e499b76975f1c2d4a5cdd74586eb415c8945cf216103ccfbc4\"" Apr 30 12:43:04.683834 containerd[2700]: time="2025-04-30T12:43:04.683815897Z" level=info msg="StartContainer for \"89aebaeee77087c645f45d44396dd0d70fc97e0ff8a44768239ff44784b8859b\"" Apr 30 12:43:04.684063 containerd[2700]: time="2025-04-30T12:43:04.684041177Z" level=info msg="CreateContainer within sandbox \"c602eaaa0aa27b96bacf4a6c4a0fb1828c00b23963ef89d0ce72b856521438b3\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"17963cab4eef06b9a2e1aa172e51df49e3b70de45ff567fbb75af4dc853ce78e\"" Apr 30 12:43:04.684315 containerd[2700]: time="2025-04-30T12:43:04.684297497Z" level=info msg="StartContainer for \"17963cab4eef06b9a2e1aa172e51df49e3b70de45ff567fbb75af4dc853ce78e\"" Apr 30 12:43:04.723929 systemd[1]: Started cri-containerd-17963cab4eef06b9a2e1aa172e51df49e3b70de45ff567fbb75af4dc853ce78e.scope - libcontainer container 17963cab4eef06b9a2e1aa172e51df49e3b70de45ff567fbb75af4dc853ce78e. Apr 30 12:43:04.725164 systemd[1]: Started cri-containerd-89aebaeee77087c645f45d44396dd0d70fc97e0ff8a44768239ff44784b8859b.scope - libcontainer container 89aebaeee77087c645f45d44396dd0d70fc97e0ff8a44768239ff44784b8859b. Apr 30 12:43:04.726313 systemd[1]: Started cri-containerd-b400c4f8335607e499b76975f1c2d4a5cdd74586eb415c8945cf216103ccfbc4.scope - libcontainer container b400c4f8335607e499b76975f1c2d4a5cdd74586eb415c8945cf216103ccfbc4. Apr 30 12:43:04.750262 containerd[2700]: time="2025-04-30T12:43:04.750218377Z" level=info msg="StartContainer for \"17963cab4eef06b9a2e1aa172e51df49e3b70de45ff567fbb75af4dc853ce78e\" returns successfully" Apr 30 12:43:04.750343 containerd[2700]: time="2025-04-30T12:43:04.750223777Z" level=info msg="StartContainer for \"89aebaeee77087c645f45d44396dd0d70fc97e0ff8a44768239ff44784b8859b\" returns successfully" Apr 30 12:43:04.751705 containerd[2700]: time="2025-04-30T12:43:04.751673017Z" level=info msg="StartContainer for \"b400c4f8335607e499b76975f1c2d4a5cdd74586eb415c8945cf216103ccfbc4\" returns successfully" Apr 30 12:43:05.372365 kubelet[3696]: I0430 12:43:05.372337 3696 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:06.410014 kubelet[3696]: E0430 12:43:06.408267 3696 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4230.1.1-a-0ae8677ef9\" not found" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:06.513282 kubelet[3696]: I0430 12:43:06.512882 3696 kubelet_node_status.go:75] "Successfully registered node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:06.513282 kubelet[3696]: E0430 12:43:06.512916 3696 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4230.1.1-a-0ae8677ef9\": node \"ci-4230.1.1-a-0ae8677ef9\" not found" Apr 30 12:43:06.795689 kubelet[3696]: I0430 12:43:06.795646 3696 apiserver.go:52] "Watching apiserver" Apr 30 12:43:06.804818 kubelet[3696]: I0430 12:43:06.804790 3696 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 30 12:43:08.372806 systemd[1]: Reload requested from client PID 4126 ('systemctl') (unit session-9.scope)... Apr 30 12:43:08.372818 systemd[1]: Reloading... Apr 30 12:43:08.446758 zram_generator::config[4180]: No configuration found. Apr 30 12:43:08.537481 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 30 12:43:08.641439 systemd[1]: Reloading finished in 268 ms. Apr 30 12:43:08.661654 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:43:08.679580 systemd[1]: kubelet.service: Deactivated successfully. Apr 30 12:43:08.680862 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:43:08.680925 systemd[1]: kubelet.service: Consumed 1.320s CPU time, 143.2M memory peak. Apr 30 12:43:08.692926 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Apr 30 12:43:08.793995 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Apr 30 12:43:08.797702 (kubelet)[4238]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Apr 30 12:43:08.826953 kubelet[4238]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 12:43:08.826953 kubelet[4238]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 30 12:43:08.826953 kubelet[4238]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 30 12:43:08.827307 kubelet[4238]: I0430 12:43:08.827002 4238 server.go:206] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 30 12:43:08.831707 kubelet[4238]: I0430 12:43:08.831683 4238 server.go:486] "Kubelet version" kubeletVersion="v1.31.0" Apr 30 12:43:08.831707 kubelet[4238]: I0430 12:43:08.831706 4238 server.go:488] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 30 12:43:08.831926 kubelet[4238]: I0430 12:43:08.831914 4238 server.go:929] "Client rotation is on, will bootstrap in background" Apr 30 12:43:08.834104 kubelet[4238]: I0430 12:43:08.834089 4238 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Apr 30 12:43:08.835908 kubelet[4238]: I0430 12:43:08.835887 4238 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 30 12:43:08.838265 kubelet[4238]: E0430 12:43:08.838237 4238 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Apr 30 12:43:08.838265 kubelet[4238]: I0430 12:43:08.838261 4238 server.go:1403] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Apr 30 12:43:08.856847 kubelet[4238]: I0430 12:43:08.856827 4238 server.go:744] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 30 12:43:08.856942 kubelet[4238]: I0430 12:43:08.856933 4238 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Apr 30 12:43:08.857053 kubelet[4238]: I0430 12:43:08.857031 4238 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 30 12:43:08.857201 kubelet[4238]: I0430 12:43:08.857055 4238 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230.1.1-a-0ae8677ef9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Apr 30 12:43:08.857267 kubelet[4238]: I0430 12:43:08.857213 4238 topology_manager.go:138] "Creating topology manager with none policy" Apr 30 12:43:08.857267 kubelet[4238]: I0430 12:43:08.857221 4238 container_manager_linux.go:300] "Creating device plugin manager" Apr 30 12:43:08.857267 kubelet[4238]: I0430 12:43:08.857253 4238 state_mem.go:36] "Initialized new in-memory state store" Apr 30 12:43:08.857356 kubelet[4238]: I0430 12:43:08.857349 4238 kubelet.go:408] "Attempting to sync node with API server" Apr 30 12:43:08.857377 kubelet[4238]: I0430 12:43:08.857360 4238 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 30 12:43:08.857397 kubelet[4238]: I0430 12:43:08.857380 4238 kubelet.go:314] "Adding apiserver pod source" Apr 30 12:43:08.857397 kubelet[4238]: I0430 12:43:08.857389 4238 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 30 12:43:08.857785 kubelet[4238]: I0430 12:43:08.857773 4238 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Apr 30 12:43:08.858201 kubelet[4238]: I0430 12:43:08.858191 4238 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Apr 30 12:43:08.858560 kubelet[4238]: I0430 12:43:08.858550 4238 server.go:1269] "Started kubelet" Apr 30 12:43:08.858676 kubelet[4238]: I0430 12:43:08.858609 4238 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Apr 30 12:43:08.858700 kubelet[4238]: I0430 12:43:08.858646 4238 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Apr 30 12:43:08.858873 kubelet[4238]: I0430 12:43:08.858862 4238 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 30 12:43:08.859662 kubelet[4238]: I0430 12:43:08.859647 4238 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 30 12:43:08.859686 kubelet[4238]: I0430 12:43:08.859657 4238 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Apr 30 12:43:08.859733 kubelet[4238]: E0430 12:43:08.859714 4238 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230.1.1-a-0ae8677ef9\" not found" Apr 30 12:43:08.859860 kubelet[4238]: I0430 12:43:08.859840 4238 volume_manager.go:289] "Starting Kubelet Volume Manager" Apr 30 12:43:08.860362 kubelet[4238]: I0430 12:43:08.860346 4238 factory.go:221] Registration of the systemd container factory successfully Apr 30 12:43:08.860389 kubelet[4238]: I0430 12:43:08.860359 4238 reconciler.go:26] "Reconciler: start to sync state" Apr 30 12:43:08.860471 kubelet[4238]: I0430 12:43:08.860453 4238 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Apr 30 12:43:08.860791 kubelet[4238]: I0430 12:43:08.860613 4238 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Apr 30 12:43:08.861032 kubelet[4238]: E0430 12:43:08.861007 4238 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 30 12:43:08.862285 kubelet[4238]: I0430 12:43:08.862265 4238 factory.go:221] Registration of the containerd container factory successfully Apr 30 12:43:08.862837 kubelet[4238]: I0430 12:43:08.862709 4238 server.go:460] "Adding debug handlers to kubelet server" Apr 30 12:43:08.867050 kubelet[4238]: I0430 12:43:08.866956 4238 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 30 12:43:08.868102 kubelet[4238]: I0430 12:43:08.868062 4238 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 30 12:43:08.868102 kubelet[4238]: I0430 12:43:08.868091 4238 status_manager.go:217] "Starting to sync pod status with apiserver" Apr 30 12:43:08.868203 kubelet[4238]: I0430 12:43:08.868112 4238 kubelet.go:2321] "Starting kubelet main sync loop" Apr 30 12:43:08.868203 kubelet[4238]: E0430 12:43:08.868164 4238 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 30 12:43:08.893769 kubelet[4238]: I0430 12:43:08.893675 4238 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 30 12:43:08.893769 kubelet[4238]: I0430 12:43:08.893695 4238 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 30 12:43:08.893769 kubelet[4238]: I0430 12:43:08.893716 4238 state_mem.go:36] "Initialized new in-memory state store" Apr 30 12:43:08.893884 kubelet[4238]: I0430 12:43:08.893877 4238 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 30 12:43:08.893903 kubelet[4238]: I0430 12:43:08.893888 4238 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 30 12:43:08.893923 kubelet[4238]: I0430 12:43:08.893907 4238 policy_none.go:49] "None policy: Start" Apr 30 12:43:08.894501 kubelet[4238]: I0430 12:43:08.894479 4238 memory_manager.go:170] "Starting memorymanager" policy="None" Apr 30 12:43:08.894552 kubelet[4238]: I0430 12:43:08.894506 4238 state_mem.go:35] "Initializing new in-memory state store" Apr 30 12:43:08.894682 kubelet[4238]: I0430 12:43:08.894670 4238 state_mem.go:75] "Updated machine memory state" Apr 30 12:43:08.897727 kubelet[4238]: I0430 12:43:08.897710 4238 manager.go:510] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 30 12:43:08.898030 kubelet[4238]: I0430 12:43:08.898018 4238 eviction_manager.go:189] "Eviction manager: starting control loop" Apr 30 12:43:08.898060 kubelet[4238]: I0430 12:43:08.898030 4238 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Apr 30 12:43:08.898301 kubelet[4238]: I0430 12:43:08.898288 4238 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 30 12:43:08.975441 kubelet[4238]: W0430 12:43:08.975414 4238 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 12:43:08.975605 kubelet[4238]: W0430 12:43:08.975587 4238 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 12:43:08.975605 kubelet[4238]: W0430 12:43:08.975595 4238 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 12:43:09.001276 kubelet[4238]: I0430 12:43:09.001254 4238 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.005140 kubelet[4238]: I0430 12:43:09.005116 4238 kubelet_node_status.go:111] "Node was previously registered" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.005197 kubelet[4238]: I0430 12:43:09.005187 4238 kubelet_node_status.go:75] "Successfully registered node" node="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162234 kubelet[4238]: I0430 12:43:09.162165 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/8696336ba6a673aa44d6b4010b4a1ffc-k8s-certs\") pod \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" (UID: \"8696336ba6a673aa44d6b4010b4a1ffc\") " pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162234 kubelet[4238]: I0430 12:43:09.162196 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/8696336ba6a673aa44d6b4010b4a1ffc-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" (UID: \"8696336ba6a673aa44d6b4010b4a1ffc\") " pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162234 kubelet[4238]: I0430 12:43:09.162219 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-ca-certs\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162234 kubelet[4238]: I0430 12:43:09.162235 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-flexvolume-dir\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162393 kubelet[4238]: I0430 12:43:09.162263 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-k8s-certs\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162393 kubelet[4238]: I0430 12:43:09.162307 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-kubeconfig\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162436 kubelet[4238]: I0430 12:43:09.162394 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/ab493c0aca086d6f4229e402c66068f4-kubeconfig\") pod \"kube-scheduler-ci-4230.1.1-a-0ae8677ef9\" (UID: \"ab493c0aca086d6f4229e402c66068f4\") " pod="kube-system/kube-scheduler-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162462 kubelet[4238]: I0430 12:43:09.162447 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/8696336ba6a673aa44d6b4010b4a1ffc-ca-certs\") pod \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" (UID: \"8696336ba6a673aa44d6b4010b4a1ffc\") " pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.162493 kubelet[4238]: I0430 12:43:09.162475 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fbb3dc52e7c4b56400cba46276b9d9f1-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230.1.1-a-0ae8677ef9\" (UID: \"fbb3dc52e7c4b56400cba46276b9d9f1\") " pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.857901 kubelet[4238]: I0430 12:43:09.857865 4238 apiserver.go:52] "Watching apiserver" Apr 30 12:43:09.877393 kubelet[4238]: W0430 12:43:09.877369 4238 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 30 12:43:09.877454 kubelet[4238]: E0430 12:43:09.877426 4238 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4230.1.1-a-0ae8677ef9\" already exists" pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:09.899757 kubelet[4238]: I0430 12:43:09.899702 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4230.1.1-a-0ae8677ef9" podStartSLOduration=1.899685657 podStartE2EDuration="1.899685657s" podCreationTimestamp="2025-04-30 12:43:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 12:43:09.899532257 +0000 UTC m=+1.098892041" watchObservedRunningTime="2025-04-30 12:43:09.899685657 +0000 UTC m=+1.099045401" Apr 30 12:43:09.909381 kubelet[4238]: I0430 12:43:09.909344 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4230.1.1-a-0ae8677ef9" podStartSLOduration=1.9093331770000002 podStartE2EDuration="1.909333177s" podCreationTimestamp="2025-04-30 12:43:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 12:43:09.904870897 +0000 UTC m=+1.104230681" watchObservedRunningTime="2025-04-30 12:43:09.909333177 +0000 UTC m=+1.108692961" Apr 30 12:43:09.914814 kubelet[4238]: I0430 12:43:09.914774 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4230.1.1-a-0ae8677ef9" podStartSLOduration=1.914761777 podStartE2EDuration="1.914761777s" podCreationTimestamp="2025-04-30 12:43:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 12:43:09.909439217 +0000 UTC m=+1.108799001" watchObservedRunningTime="2025-04-30 12:43:09.914761777 +0000 UTC m=+1.114121561" Apr 30 12:43:09.962897 kubelet[4238]: I0430 12:43:09.961888 4238 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Apr 30 12:43:13.167599 sudo[2966]: pam_unix(sudo:session): session closed for user root Apr 30 12:43:13.230114 sshd[2965]: Connection closed by 139.178.68.195 port 56222 Apr 30 12:43:13.230839 sshd-session[2963]: pam_unix(sshd:session): session closed for user core Apr 30 12:43:13.233671 systemd[1]: sshd@6-147.75.61.11:22-139.178.68.195:56222.service: Deactivated successfully. Apr 30 12:43:13.236178 systemd[1]: session-9.scope: Deactivated successfully. Apr 30 12:43:13.236356 systemd[1]: session-9.scope: Consumed 7.146s CPU time, 236.7M memory peak. Apr 30 12:43:13.237497 systemd-logind[2684]: Session 9 logged out. Waiting for processes to exit. Apr 30 12:43:13.238167 systemd-logind[2684]: Removed session 9. Apr 30 12:43:13.517753 kubelet[4238]: I0430 12:43:13.517654 4238 kuberuntime_manager.go:1633] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 30 12:43:13.518025 containerd[2700]: time="2025-04-30T12:43:13.517917377Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 30 12:43:13.518181 kubelet[4238]: I0430 12:43:13.518075 4238 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 30 12:43:14.394585 systemd[1]: Created slice kubepods-besteffort-pod618e221f_764c_47d0_a91d_7c611a770f99.slice - libcontainer container kubepods-besteffort-pod618e221f_764c_47d0_a91d_7c611a770f99.slice. Apr 30 12:43:14.491460 kubelet[4238]: I0430 12:43:14.491435 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/618e221f-764c-47d0-a91d-7c611a770f99-kube-proxy\") pod \"kube-proxy-wcx9x\" (UID: \"618e221f-764c-47d0-a91d-7c611a770f99\") " pod="kube-system/kube-proxy-wcx9x" Apr 30 12:43:14.491556 kubelet[4238]: I0430 12:43:14.491465 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/618e221f-764c-47d0-a91d-7c611a770f99-lib-modules\") pod \"kube-proxy-wcx9x\" (UID: \"618e221f-764c-47d0-a91d-7c611a770f99\") " pod="kube-system/kube-proxy-wcx9x" Apr 30 12:43:14.491556 kubelet[4238]: I0430 12:43:14.491483 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94chm\" (UniqueName: \"kubernetes.io/projected/618e221f-764c-47d0-a91d-7c611a770f99-kube-api-access-94chm\") pod \"kube-proxy-wcx9x\" (UID: \"618e221f-764c-47d0-a91d-7c611a770f99\") " pod="kube-system/kube-proxy-wcx9x" Apr 30 12:43:14.491556 kubelet[4238]: I0430 12:43:14.491512 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/618e221f-764c-47d0-a91d-7c611a770f99-xtables-lock\") pod \"kube-proxy-wcx9x\" (UID: \"618e221f-764c-47d0-a91d-7c611a770f99\") " pod="kube-system/kube-proxy-wcx9x" Apr 30 12:43:14.655483 systemd[1]: Created slice kubepods-besteffort-pod5c6e8e46_c94d_46c5_9cd0_bee6495f6cab.slice - libcontainer container kubepods-besteffort-pod5c6e8e46_c94d_46c5_9cd0_bee6495f6cab.slice. Apr 30 12:43:14.692970 kubelet[4238]: I0430 12:43:14.692896 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/5c6e8e46-c94d-46c5-9cd0-bee6495f6cab-var-lib-calico\") pod \"tigera-operator-6f6897fdc5-8wwvs\" (UID: \"5c6e8e46-c94d-46c5-9cd0-bee6495f6cab\") " pod="tigera-operator/tigera-operator-6f6897fdc5-8wwvs" Apr 30 12:43:14.692970 kubelet[4238]: I0430 12:43:14.692959 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmgj9\" (UniqueName: \"kubernetes.io/projected/5c6e8e46-c94d-46c5-9cd0-bee6495f6cab-kube-api-access-zmgj9\") pod \"tigera-operator-6f6897fdc5-8wwvs\" (UID: \"5c6e8e46-c94d-46c5-9cd0-bee6495f6cab\") " pod="tigera-operator/tigera-operator-6f6897fdc5-8wwvs" Apr 30 12:43:14.713507 containerd[2700]: time="2025-04-30T12:43:14.713479177Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wcx9x,Uid:618e221f-764c-47d0-a91d-7c611a770f99,Namespace:kube-system,Attempt:0,}" Apr 30 12:43:14.726524 containerd[2700]: time="2025-04-30T12:43:14.726135817Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:14.726524 containerd[2700]: time="2025-04-30T12:43:14.726501777Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:14.726524 containerd[2700]: time="2025-04-30T12:43:14.726516137Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:14.726610 containerd[2700]: time="2025-04-30T12:43:14.726592257Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:14.749911 systemd[1]: Started cri-containerd-eaece0be0aa805c05342cd6d5a3c7d238dcc280210cabc844aa2933b098aa09b.scope - libcontainer container eaece0be0aa805c05342cd6d5a3c7d238dcc280210cabc844aa2933b098aa09b. Apr 30 12:43:14.765832 containerd[2700]: time="2025-04-30T12:43:14.765798017Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-wcx9x,Uid:618e221f-764c-47d0-a91d-7c611a770f99,Namespace:kube-system,Attempt:0,} returns sandbox id \"eaece0be0aa805c05342cd6d5a3c7d238dcc280210cabc844aa2933b098aa09b\"" Apr 30 12:43:14.767849 containerd[2700]: time="2025-04-30T12:43:14.767824897Z" level=info msg="CreateContainer within sandbox \"eaece0be0aa805c05342cd6d5a3c7d238dcc280210cabc844aa2933b098aa09b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 30 12:43:14.774578 containerd[2700]: time="2025-04-30T12:43:14.774544977Z" level=info msg="CreateContainer within sandbox \"eaece0be0aa805c05342cd6d5a3c7d238dcc280210cabc844aa2933b098aa09b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"c8617bbe4d607813d8ec36da978f626371964e23e03b0e852937dc833476d9e2\"" Apr 30 12:43:14.775003 containerd[2700]: time="2025-04-30T12:43:14.774981537Z" level=info msg="StartContainer for \"c8617bbe4d607813d8ec36da978f626371964e23e03b0e852937dc833476d9e2\"" Apr 30 12:43:14.803860 systemd[1]: Started cri-containerd-c8617bbe4d607813d8ec36da978f626371964e23e03b0e852937dc833476d9e2.scope - libcontainer container c8617bbe4d607813d8ec36da978f626371964e23e03b0e852937dc833476d9e2. Apr 30 12:43:14.825070 containerd[2700]: time="2025-04-30T12:43:14.825031657Z" level=info msg="StartContainer for \"c8617bbe4d607813d8ec36da978f626371964e23e03b0e852937dc833476d9e2\" returns successfully" Apr 30 12:43:14.895752 kubelet[4238]: I0430 12:43:14.895704 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-wcx9x" podStartSLOduration=0.895688497 podStartE2EDuration="895.688497ms" podCreationTimestamp="2025-04-30 12:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 12:43:14.895593537 +0000 UTC m=+6.094953321" watchObservedRunningTime="2025-04-30 12:43:14.895688497 +0000 UTC m=+6.095048241" Apr 30 12:43:14.958226 containerd[2700]: time="2025-04-30T12:43:14.958156457Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6f6897fdc5-8wwvs,Uid:5c6e8e46-c94d-46c5-9cd0-bee6495f6cab,Namespace:tigera-operator,Attempt:0,}" Apr 30 12:43:14.972068 containerd[2700]: time="2025-04-30T12:43:14.972007537Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:14.972068 containerd[2700]: time="2025-04-30T12:43:14.972061297Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:14.972153 containerd[2700]: time="2025-04-30T12:43:14.972074817Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:14.972174 containerd[2700]: time="2025-04-30T12:43:14.972145577Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:15.005871 systemd[1]: Started cri-containerd-be6908bd4cce3799a459b8a9b7b209d7afd6b7d67fda3e004981318be35408ee.scope - libcontainer container be6908bd4cce3799a459b8a9b7b209d7afd6b7d67fda3e004981318be35408ee. Apr 30 12:43:15.028625 containerd[2700]: time="2025-04-30T12:43:15.028587017Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-6f6897fdc5-8wwvs,Uid:5c6e8e46-c94d-46c5-9cd0-bee6495f6cab,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"be6908bd4cce3799a459b8a9b7b209d7afd6b7d67fda3e004981318be35408ee\"" Apr 30 12:43:15.029941 containerd[2700]: time="2025-04-30T12:43:15.029918177Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\"" Apr 30 12:43:19.569413 update_engine[2695]: I20250430 12:43:19.569345 2695 update_attempter.cc:509] Updating boot flags... Apr 30 12:43:19.604759 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (4810) Apr 30 12:43:19.633749 kernel: BTRFS warning: duplicate device /dev/nvme0n1p3 devid 1 generation 37 scanned by (udev-worker) (4813) Apr 30 12:43:20.919567 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1777920676.mount: Deactivated successfully. Apr 30 12:43:21.480497 containerd[2700]: time="2025-04-30T12:43:21.480442213Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:21.480831 containerd[2700]: time="2025-04-30T12:43:21.480487215Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.7: active requests=0, bytes read=19323084" Apr 30 12:43:21.481347 containerd[2700]: time="2025-04-30T12:43:21.481298132Z" level=info msg="ImageCreate event name:\"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:21.483377 containerd[2700]: time="2025-04-30T12:43:21.483331864Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:21.484233 containerd[2700]: time="2025-04-30T12:43:21.484149461Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.7\" with image id \"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\", repo tag \"quay.io/tigera/operator:v1.36.7\", repo digest \"quay.io/tigera/operator@sha256:a4a44422d8f2a14e0aaea2031ccb5580f2bf68218c9db444450c1888743305e9\", size \"19319079\" in 6.454202004s" Apr 30 12:43:21.484233 containerd[2700]: time="2025-04-30T12:43:21.484183023Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.7\" returns image reference \"sha256:27f7c2cfac802523e44ecd16453a4cc992f6c7d610c13054f2715a7cb4370565\"" Apr 30 12:43:21.485957 containerd[2700]: time="2025-04-30T12:43:21.485858579Z" level=info msg="CreateContainer within sandbox \"be6908bd4cce3799a459b8a9b7b209d7afd6b7d67fda3e004981318be35408ee\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Apr 30 12:43:21.490785 containerd[2700]: time="2025-04-30T12:43:21.490760441Z" level=info msg="CreateContainer within sandbox \"be6908bd4cce3799a459b8a9b7b209d7afd6b7d67fda3e004981318be35408ee\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"f77b2453ca59108ccf74f85e217ea83fe7f36654aef69ecdda9e85a2b36f7e06\"" Apr 30 12:43:21.491130 containerd[2700]: time="2025-04-30T12:43:21.491107537Z" level=info msg="StartContainer for \"f77b2453ca59108ccf74f85e217ea83fe7f36654aef69ecdda9e85a2b36f7e06\"" Apr 30 12:43:21.520854 systemd[1]: Started cri-containerd-f77b2453ca59108ccf74f85e217ea83fe7f36654aef69ecdda9e85a2b36f7e06.scope - libcontainer container f77b2453ca59108ccf74f85e217ea83fe7f36654aef69ecdda9e85a2b36f7e06. Apr 30 12:43:21.538401 containerd[2700]: time="2025-04-30T12:43:21.538375243Z" level=info msg="StartContainer for \"f77b2453ca59108ccf74f85e217ea83fe7f36654aef69ecdda9e85a2b36f7e06\" returns successfully" Apr 30 12:43:21.897345 kubelet[4238]: I0430 12:43:21.896917 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-6f6897fdc5-8wwvs" podStartSLOduration=1.441621249 podStartE2EDuration="7.896903562s" podCreationTimestamp="2025-04-30 12:43:14 +0000 UTC" firstStartedPulling="2025-04-30 12:43:15.029493617 +0000 UTC m=+6.228853401" lastFinishedPulling="2025-04-30 12:43:21.48477593 +0000 UTC m=+12.684135714" observedRunningTime="2025-04-30 12:43:21.896760996 +0000 UTC m=+13.096120780" watchObservedRunningTime="2025-04-30 12:43:21.896903562 +0000 UTC m=+13.096263306" Apr 30 12:43:25.380622 systemd[1]: Created slice kubepods-besteffort-pod57a7274e_0f1d_4cbe_bc2e_62a96303bda0.slice - libcontainer container kubepods-besteffort-pod57a7274e_0f1d_4cbe_bc2e_62a96303bda0.slice. Apr 30 12:43:25.562894 systemd[1]: Created slice kubepods-besteffort-pod46716d0c_25a3_4f9c_962c_9d21d94b9fb1.slice - libcontainer container kubepods-besteffort-pod46716d0c_25a3_4f9c_962c_9d21d94b9fb1.slice. Apr 30 12:43:25.565129 kubelet[4238]: I0430 12:43:25.565096 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/57a7274e-0f1d-4cbe-bc2e-62a96303bda0-typha-certs\") pod \"calico-typha-d6c4d9d66-cntr2\" (UID: \"57a7274e-0f1d-4cbe-bc2e-62a96303bda0\") " pod="calico-system/calico-typha-d6c4d9d66-cntr2" Apr 30 12:43:25.565482 kubelet[4238]: I0430 12:43:25.565134 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/57a7274e-0f1d-4cbe-bc2e-62a96303bda0-tigera-ca-bundle\") pod \"calico-typha-d6c4d9d66-cntr2\" (UID: \"57a7274e-0f1d-4cbe-bc2e-62a96303bda0\") " pod="calico-system/calico-typha-d6c4d9d66-cntr2" Apr 30 12:43:25.565482 kubelet[4238]: I0430 12:43:25.565156 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rqmz\" (UniqueName: \"kubernetes.io/projected/57a7274e-0f1d-4cbe-bc2e-62a96303bda0-kube-api-access-9rqmz\") pod \"calico-typha-d6c4d9d66-cntr2\" (UID: \"57a7274e-0f1d-4cbe-bc2e-62a96303bda0\") " pod="calico-system/calico-typha-d6c4d9d66-cntr2" Apr 30 12:43:25.666385 kubelet[4238]: I0430 12:43:25.666299 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-cni-log-dir\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666385 kubelet[4238]: I0430 12:43:25.666351 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-node-certs\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666479 kubelet[4238]: I0430 12:43:25.666406 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-xtables-lock\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666479 kubelet[4238]: I0430 12:43:25.666435 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-policysync\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666479 kubelet[4238]: I0430 12:43:25.666460 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-tigera-ca-bundle\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666479 kubelet[4238]: I0430 12:43:25.666476 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-var-lib-calico\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666628 kubelet[4238]: I0430 12:43:25.666493 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-lib-modules\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666628 kubelet[4238]: I0430 12:43:25.666512 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvfrm\" (UniqueName: \"kubernetes.io/projected/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-kube-api-access-qvfrm\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666628 kubelet[4238]: I0430 12:43:25.666533 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-var-run-calico\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666628 kubelet[4238]: I0430 12:43:25.666547 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-cni-bin-dir\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666628 kubelet[4238]: I0430 12:43:25.666565 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-cni-net-dir\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.666726 kubelet[4238]: I0430 12:43:25.666632 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/46716d0c-25a3-4f9c-962c-9d21d94b9fb1-flexvol-driver-host\") pod \"calico-node-b4txl\" (UID: \"46716d0c-25a3-4f9c-962c-9d21d94b9fb1\") " pod="calico-system/calico-node-b4txl" Apr 30 12:43:25.684458 containerd[2700]: time="2025-04-30T12:43:25.684415670Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-d6c4d9d66-cntr2,Uid:57a7274e-0f1d-4cbe-bc2e-62a96303bda0,Namespace:calico-system,Attempt:0,}" Apr 30 12:43:25.697657 containerd[2700]: time="2025-04-30T12:43:25.697256360Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:25.697688 containerd[2700]: time="2025-04-30T12:43:25.697655334Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:25.697688 containerd[2700]: time="2025-04-30T12:43:25.697674815Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:25.697778 containerd[2700]: time="2025-04-30T12:43:25.697763058Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:25.722846 systemd[1]: Started cri-containerd-83f0db5bc889c775e53cd0da280625597a76d3c98f17f9771b818d95a00eaee1.scope - libcontainer container 83f0db5bc889c775e53cd0da280625597a76d3c98f17f9771b818d95a00eaee1. Apr 30 12:43:25.745840 containerd[2700]: time="2025-04-30T12:43:25.745811663Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-d6c4d9d66-cntr2,Uid:57a7274e-0f1d-4cbe-bc2e-62a96303bda0,Namespace:calico-system,Attempt:0,} returns sandbox id \"83f0db5bc889c775e53cd0da280625597a76d3c98f17f9771b818d95a00eaee1\"" Apr 30 12:43:25.747490 containerd[2700]: time="2025-04-30T12:43:25.747461801Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\"" Apr 30 12:43:25.768545 kubelet[4238]: E0430 12:43:25.768524 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.768577 kubelet[4238]: W0430 12:43:25.768543 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.768646 kubelet[4238]: E0430 12:43:25.768634 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.769989 kubelet[4238]: E0430 12:43:25.769970 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.770016 kubelet[4238]: W0430 12:43:25.769987 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.770016 kubelet[4238]: E0430 12:43:25.770003 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.771788 kubelet[4238]: E0430 12:43:25.771725 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nckqz" podUID="45ca62a1-734b-4197-a2e5-9df6c9a9ad92" Apr 30 12:43:25.775796 kubelet[4238]: E0430 12:43:25.775780 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.775818 kubelet[4238]: W0430 12:43:25.775794 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.775818 kubelet[4238]: E0430 12:43:25.775808 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.865632 containerd[2700]: time="2025-04-30T12:43:25.865602545Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-b4txl,Uid:46716d0c-25a3-4f9c-962c-9d21d94b9fb1,Namespace:calico-system,Attempt:0,}" Apr 30 12:43:25.868067 kubelet[4238]: E0430 12:43:25.868044 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.868094 kubelet[4238]: W0430 12:43:25.868066 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.868094 kubelet[4238]: E0430 12:43:25.868084 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.868290 kubelet[4238]: E0430 12:43:25.868279 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.868290 kubelet[4238]: W0430 12:43:25.868288 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.868335 kubelet[4238]: E0430 12:43:25.868295 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.868465 kubelet[4238]: E0430 12:43:25.868455 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.868465 kubelet[4238]: W0430 12:43:25.868462 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.868510 kubelet[4238]: E0430 12:43:25.868470 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.868661 kubelet[4238]: E0430 12:43:25.868650 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.868661 kubelet[4238]: W0430 12:43:25.868659 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.868705 kubelet[4238]: E0430 12:43:25.868667 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.868905 kubelet[4238]: E0430 12:43:25.868893 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.868905 kubelet[4238]: W0430 12:43:25.868902 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.868945 kubelet[4238]: E0430 12:43:25.868909 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.869060 kubelet[4238]: E0430 12:43:25.869049 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.869060 kubelet[4238]: W0430 12:43:25.869057 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.869108 kubelet[4238]: E0430 12:43:25.869064 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.869274 kubelet[4238]: E0430 12:43:25.869263 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.869274 kubelet[4238]: W0430 12:43:25.869271 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.869310 kubelet[4238]: E0430 12:43:25.869278 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.869491 kubelet[4238]: E0430 12:43:25.869483 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.869515 kubelet[4238]: W0430 12:43:25.869491 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.869515 kubelet[4238]: E0430 12:43:25.869499 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.869702 kubelet[4238]: E0430 12:43:25.869694 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.869723 kubelet[4238]: W0430 12:43:25.869705 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.869723 kubelet[4238]: E0430 12:43:25.869712 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.869927 kubelet[4238]: E0430 12:43:25.869919 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.869955 kubelet[4238]: W0430 12:43:25.869926 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.869955 kubelet[4238]: E0430 12:43:25.869934 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.870158 kubelet[4238]: E0430 12:43:25.870150 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.870177 kubelet[4238]: W0430 12:43:25.870157 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.870177 kubelet[4238]: E0430 12:43:25.870164 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.870381 kubelet[4238]: E0430 12:43:25.870370 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.870381 kubelet[4238]: W0430 12:43:25.870378 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.870425 kubelet[4238]: E0430 12:43:25.870385 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.870605 kubelet[4238]: E0430 12:43:25.870595 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.870605 kubelet[4238]: W0430 12:43:25.870602 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.870647 kubelet[4238]: E0430 12:43:25.870610 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.870831 kubelet[4238]: E0430 12:43:25.870820 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.870831 kubelet[4238]: W0430 12:43:25.870828 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.870873 kubelet[4238]: E0430 12:43:25.870838 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.871016 kubelet[4238]: E0430 12:43:25.871007 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.871016 kubelet[4238]: W0430 12:43:25.871014 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.871057 kubelet[4238]: E0430 12:43:25.871021 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.871227 kubelet[4238]: E0430 12:43:25.871217 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.871227 kubelet[4238]: W0430 12:43:25.871224 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.871271 kubelet[4238]: E0430 12:43:25.871231 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.871441 kubelet[4238]: E0430 12:43:25.871433 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.871462 kubelet[4238]: W0430 12:43:25.871441 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.871462 kubelet[4238]: E0430 12:43:25.871448 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.871629 kubelet[4238]: E0430 12:43:25.871621 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.871653 kubelet[4238]: W0430 12:43:25.871629 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.871653 kubelet[4238]: E0430 12:43:25.871636 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.871796 kubelet[4238]: E0430 12:43:25.871789 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.871816 kubelet[4238]: W0430 12:43:25.871796 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.871816 kubelet[4238]: E0430 12:43:25.871803 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.871983 kubelet[4238]: E0430 12:43:25.871975 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.872005 kubelet[4238]: W0430 12:43:25.871983 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.872005 kubelet[4238]: E0430 12:43:25.871990 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.878049 containerd[2700]: time="2025-04-30T12:43:25.877703009Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:25.878049 containerd[2700]: time="2025-04-30T12:43:25.878040301Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:25.878090 containerd[2700]: time="2025-04-30T12:43:25.878055861Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:25.878144 containerd[2700]: time="2025-04-30T12:43:25.878128864Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:25.896884 systemd[1]: Started cri-containerd-1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5.scope - libcontainer container 1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5. Apr 30 12:43:25.913712 containerd[2700]: time="2025-04-30T12:43:25.913678431Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-b4txl,Uid:46716d0c-25a3-4f9c-962c-9d21d94b9fb1,Namespace:calico-system,Attempt:0,} returns sandbox id \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\"" Apr 30 12:43:25.968300 kubelet[4238]: E0430 12:43:25.968237 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.968300 kubelet[4238]: W0430 12:43:25.968256 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.968300 kubelet[4238]: E0430 12:43:25.968272 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.968300 kubelet[4238]: I0430 12:43:25.968296 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcxct\" (UniqueName: \"kubernetes.io/projected/45ca62a1-734b-4197-a2e5-9df6c9a9ad92-kube-api-access-lcxct\") pod \"csi-node-driver-nckqz\" (UID: \"45ca62a1-734b-4197-a2e5-9df6c9a9ad92\") " pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:25.968660 kubelet[4238]: E0430 12:43:25.968644 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.968683 kubelet[4238]: W0430 12:43:25.968658 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.968683 kubelet[4238]: E0430 12:43:25.968674 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.968720 kubelet[4238]: I0430 12:43:25.968693 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/45ca62a1-734b-4197-a2e5-9df6c9a9ad92-varrun\") pod \"csi-node-driver-nckqz\" (UID: \"45ca62a1-734b-4197-a2e5-9df6c9a9ad92\") " pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:25.968952 kubelet[4238]: E0430 12:43:25.968933 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.968981 kubelet[4238]: W0430 12:43:25.968951 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.968981 kubelet[4238]: E0430 12:43:25.968971 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.969131 kubelet[4238]: E0430 12:43:25.969120 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.969131 kubelet[4238]: W0430 12:43:25.969128 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.969172 kubelet[4238]: E0430 12:43:25.969138 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.969348 kubelet[4238]: E0430 12:43:25.969337 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.969348 kubelet[4238]: W0430 12:43:25.969345 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.969391 kubelet[4238]: E0430 12:43:25.969355 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.969391 kubelet[4238]: I0430 12:43:25.969376 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/45ca62a1-734b-4197-a2e5-9df6c9a9ad92-kubelet-dir\") pod \"csi-node-driver-nckqz\" (UID: \"45ca62a1-734b-4197-a2e5-9df6c9a9ad92\") " pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:25.969542 kubelet[4238]: E0430 12:43:25.969530 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.969542 kubelet[4238]: W0430 12:43:25.969540 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.969585 kubelet[4238]: E0430 12:43:25.969551 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.969585 kubelet[4238]: I0430 12:43:25.969566 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/45ca62a1-734b-4197-a2e5-9df6c9a9ad92-registration-dir\") pod \"csi-node-driver-nckqz\" (UID: \"45ca62a1-734b-4197-a2e5-9df6c9a9ad92\") " pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:25.969853 kubelet[4238]: E0430 12:43:25.969834 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.969853 kubelet[4238]: W0430 12:43:25.969848 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.969916 kubelet[4238]: E0430 12:43:25.969866 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.970093 kubelet[4238]: E0430 12:43:25.970080 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.970127 kubelet[4238]: W0430 12:43:25.970092 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.970127 kubelet[4238]: E0430 12:43:25.970106 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.970359 kubelet[4238]: E0430 12:43:25.970349 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.970383 kubelet[4238]: W0430 12:43:25.970362 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.970383 kubelet[4238]: E0430 12:43:25.970373 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.970542 kubelet[4238]: E0430 12:43:25.970533 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.970606 kubelet[4238]: W0430 12:43:25.970541 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.970606 kubelet[4238]: E0430 12:43:25.970552 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.970768 kubelet[4238]: E0430 12:43:25.970755 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.970768 kubelet[4238]: W0430 12:43:25.970764 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.970807 kubelet[4238]: E0430 12:43:25.970775 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.970807 kubelet[4238]: I0430 12:43:25.970793 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/45ca62a1-734b-4197-a2e5-9df6c9a9ad92-socket-dir\") pod \"csi-node-driver-nckqz\" (UID: \"45ca62a1-734b-4197-a2e5-9df6c9a9ad92\") " pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:25.971001 kubelet[4238]: E0430 12:43:25.970987 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.971001 kubelet[4238]: W0430 12:43:25.970999 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.971044 kubelet[4238]: E0430 12:43:25.971018 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.971210 kubelet[4238]: E0430 12:43:25.971201 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.971230 kubelet[4238]: W0430 12:43:25.971209 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.971230 kubelet[4238]: E0430 12:43:25.971220 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.971479 kubelet[4238]: E0430 12:43:25.971467 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.971479 kubelet[4238]: W0430 12:43:25.971476 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.971521 kubelet[4238]: E0430 12:43:25.971484 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:25.971709 kubelet[4238]: E0430 12:43:25.971697 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:25.971709 kubelet[4238]: W0430 12:43:25.971706 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:25.971758 kubelet[4238]: E0430 12:43:25.971716 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.071991 kubelet[4238]: E0430 12:43:26.071970 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.071991 kubelet[4238]: W0430 12:43:26.071983 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.072080 kubelet[4238]: E0430 12:43:26.071995 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.072284 kubelet[4238]: E0430 12:43:26.072273 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.072284 kubelet[4238]: W0430 12:43:26.072282 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.072324 kubelet[4238]: E0430 12:43:26.072294 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.072457 kubelet[4238]: E0430 12:43:26.072449 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.072479 kubelet[4238]: W0430 12:43:26.072457 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.072479 kubelet[4238]: E0430 12:43:26.072468 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.072657 kubelet[4238]: E0430 12:43:26.072649 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.072677 kubelet[4238]: W0430 12:43:26.072657 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.072677 kubelet[4238]: E0430 12:43:26.072668 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.072977 kubelet[4238]: E0430 12:43:26.072958 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.073003 kubelet[4238]: W0430 12:43:26.072975 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.073003 kubelet[4238]: E0430 12:43:26.072993 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.073193 kubelet[4238]: E0430 12:43:26.073182 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.073193 kubelet[4238]: W0430 12:43:26.073190 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.073234 kubelet[4238]: E0430 12:43:26.073201 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.073417 kubelet[4238]: E0430 12:43:26.073409 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.073437 kubelet[4238]: W0430 12:43:26.073417 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.073437 kubelet[4238]: E0430 12:43:26.073427 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.073600 kubelet[4238]: E0430 12:43:26.073589 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.073600 kubelet[4238]: W0430 12:43:26.073597 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.073639 kubelet[4238]: E0430 12:43:26.073617 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.073783 kubelet[4238]: E0430 12:43:26.073772 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.073783 kubelet[4238]: W0430 12:43:26.073780 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.073824 kubelet[4238]: E0430 12:43:26.073796 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.073927 kubelet[4238]: E0430 12:43:26.073917 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.073927 kubelet[4238]: W0430 12:43:26.073924 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.073966 kubelet[4238]: E0430 12:43:26.073944 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.074077 kubelet[4238]: E0430 12:43:26.074067 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.074077 kubelet[4238]: W0430 12:43:26.074074 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.074118 kubelet[4238]: E0430 12:43:26.074090 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.074210 kubelet[4238]: E0430 12:43:26.074200 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.074210 kubelet[4238]: W0430 12:43:26.074207 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.074250 kubelet[4238]: E0430 12:43:26.074222 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.074412 kubelet[4238]: E0430 12:43:26.074404 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.074435 kubelet[4238]: W0430 12:43:26.074411 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.074435 kubelet[4238]: E0430 12:43:26.074424 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.074605 kubelet[4238]: E0430 12:43:26.074597 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.074628 kubelet[4238]: W0430 12:43:26.074605 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.074628 kubelet[4238]: E0430 12:43:26.074615 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.074873 kubelet[4238]: E0430 12:43:26.074862 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.074873 kubelet[4238]: W0430 12:43:26.074870 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.074911 kubelet[4238]: E0430 12:43:26.074881 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.075052 kubelet[4238]: E0430 12:43:26.075041 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.075052 kubelet[4238]: W0430 12:43:26.075050 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.075095 kubelet[4238]: E0430 12:43:26.075063 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.075210 kubelet[4238]: E0430 12:43:26.075199 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.075210 kubelet[4238]: W0430 12:43:26.075207 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.075250 kubelet[4238]: E0430 12:43:26.075217 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.075456 kubelet[4238]: E0430 12:43:26.075445 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.075456 kubelet[4238]: W0430 12:43:26.075454 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.075497 kubelet[4238]: E0430 12:43:26.075474 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.075608 kubelet[4238]: E0430 12:43:26.075598 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.075608 kubelet[4238]: W0430 12:43:26.075605 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.075647 kubelet[4238]: E0430 12:43:26.075620 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.075809 kubelet[4238]: E0430 12:43:26.075801 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.075829 kubelet[4238]: W0430 12:43:26.075808 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.075829 kubelet[4238]: E0430 12:43:26.075819 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.075963 kubelet[4238]: E0430 12:43:26.075955 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.075982 kubelet[4238]: W0430 12:43:26.075962 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.075982 kubelet[4238]: E0430 12:43:26.075973 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.076180 kubelet[4238]: E0430 12:43:26.076171 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.076202 kubelet[4238]: W0430 12:43:26.076180 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.076202 kubelet[4238]: E0430 12:43:26.076191 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.076378 kubelet[4238]: E0430 12:43:26.076369 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.076401 kubelet[4238]: W0430 12:43:26.076378 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.076401 kubelet[4238]: E0430 12:43:26.076386 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.076529 kubelet[4238]: E0430 12:43:26.076521 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.076549 kubelet[4238]: W0430 12:43:26.076528 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.076549 kubelet[4238]: E0430 12:43:26.076536 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.076856 kubelet[4238]: E0430 12:43:26.076846 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.076876 kubelet[4238]: W0430 12:43:26.076856 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.076876 kubelet[4238]: E0430 12:43:26.076865 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.082596 kubelet[4238]: E0430 12:43:26.082583 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:26.082616 kubelet[4238]: W0430 12:43:26.082597 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:26.082616 kubelet[4238]: E0430 12:43:26.082609 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:26.986236 containerd[2700]: time="2025-04-30T12:43:26.986198888Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:26.986559 containerd[2700]: time="2025-04-30T12:43:26.986264211Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.3: active requests=0, bytes read=28370571" Apr 30 12:43:26.986952 containerd[2700]: time="2025-04-30T12:43:26.986936153Z" level=info msg="ImageCreate event name:\"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:26.988682 containerd[2700]: time="2025-04-30T12:43:26.988655889Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:26.989361 containerd[2700]: time="2025-04-30T12:43:26.989343832Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.3\" with image id \"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:f5516aa6a78f00931d2625f3012dcf2c69d141ce41483b8d59c6ec6330a18620\", size \"29739745\" in 1.24185203s" Apr 30 12:43:26.989397 containerd[2700]: time="2025-04-30T12:43:26.989367753Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.3\" returns image reference \"sha256:26e730979a07ea7452715da6ac48076016018bc982c06ebd32d5e095f42d3d54\"" Apr 30 12:43:26.990199 containerd[2700]: time="2025-04-30T12:43:26.990181099Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\"" Apr 30 12:43:26.994870 containerd[2700]: time="2025-04-30T12:43:26.994844093Z" level=info msg="CreateContainer within sandbox \"83f0db5bc889c775e53cd0da280625597a76d3c98f17f9771b818d95a00eaee1\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Apr 30 12:43:27.000456 containerd[2700]: time="2025-04-30T12:43:27.000427636Z" level=info msg="CreateContainer within sandbox \"83f0db5bc889c775e53cd0da280625597a76d3c98f17f9771b818d95a00eaee1\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"13c5f4e138a513008d4ab7ed2959188f0d2ec9b4c51cab7c367e826c90933bc8\"" Apr 30 12:43:27.000772 containerd[2700]: time="2025-04-30T12:43:27.000749887Z" level=info msg="StartContainer for \"13c5f4e138a513008d4ab7ed2959188f0d2ec9b4c51cab7c367e826c90933bc8\"" Apr 30 12:43:27.027917 systemd[1]: Started cri-containerd-13c5f4e138a513008d4ab7ed2959188f0d2ec9b4c51cab7c367e826c90933bc8.scope - libcontainer container 13c5f4e138a513008d4ab7ed2959188f0d2ec9b4c51cab7c367e826c90933bc8. Apr 30 12:43:27.052795 containerd[2700]: time="2025-04-30T12:43:27.052764410Z" level=info msg="StartContainer for \"13c5f4e138a513008d4ab7ed2959188f0d2ec9b4c51cab7c367e826c90933bc8\" returns successfully" Apr 30 12:43:27.868646 kubelet[4238]: E0430 12:43:27.868574 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nckqz" podUID="45ca62a1-734b-4197-a2e5-9df6c9a9ad92" Apr 30 12:43:27.908323 kubelet[4238]: I0430 12:43:27.908284 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-d6c4d9d66-cntr2" podStartSLOduration=1.665466242 podStartE2EDuration="2.908273463s" podCreationTimestamp="2025-04-30 12:43:25 +0000 UTC" firstStartedPulling="2025-04-30 12:43:25.747251194 +0000 UTC m=+16.946610978" lastFinishedPulling="2025-04-30 12:43:26.990058415 +0000 UTC m=+18.189418199" observedRunningTime="2025-04-30 12:43:27.90815998 +0000 UTC m=+19.107519764" watchObservedRunningTime="2025-04-30 12:43:27.908273463 +0000 UTC m=+19.107633247" Apr 30 12:43:27.986563 kubelet[4238]: E0430 12:43:27.986546 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.986563 kubelet[4238]: W0430 12:43:27.986563 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.986641 kubelet[4238]: E0430 12:43:27.986578 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.986869 kubelet[4238]: E0430 12:43:27.986861 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.986907 kubelet[4238]: W0430 12:43:27.986869 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.986907 kubelet[4238]: E0430 12:43:27.986877 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.987041 kubelet[4238]: E0430 12:43:27.987032 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.987041 kubelet[4238]: W0430 12:43:27.987040 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.987088 kubelet[4238]: E0430 12:43:27.987049 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.987261 kubelet[4238]: E0430 12:43:27.987252 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.987261 kubelet[4238]: W0430 12:43:27.987261 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.987308 kubelet[4238]: E0430 12:43:27.987270 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.987546 kubelet[4238]: E0430 12:43:27.987537 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.987546 kubelet[4238]: W0430 12:43:27.987546 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.987595 kubelet[4238]: E0430 12:43:27.987554 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.987763 kubelet[4238]: E0430 12:43:27.987755 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.987763 kubelet[4238]: W0430 12:43:27.987763 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.987824 kubelet[4238]: E0430 12:43:27.987771 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.987931 kubelet[4238]: E0430 12:43:27.987921 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.987931 kubelet[4238]: W0430 12:43:27.987930 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.987982 kubelet[4238]: E0430 12:43:27.987938 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.988088 kubelet[4238]: E0430 12:43:27.988078 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.988088 kubelet[4238]: W0430 12:43:27.988085 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.988137 kubelet[4238]: E0430 12:43:27.988093 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.988263 kubelet[4238]: E0430 12:43:27.988255 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.988263 kubelet[4238]: W0430 12:43:27.988263 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.988317 kubelet[4238]: E0430 12:43:27.988271 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.988484 kubelet[4238]: E0430 12:43:27.988476 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.988484 kubelet[4238]: W0430 12:43:27.988483 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.988532 kubelet[4238]: E0430 12:43:27.988491 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.988685 kubelet[4238]: E0430 12:43:27.988672 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.988685 kubelet[4238]: W0430 12:43:27.988680 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.988734 kubelet[4238]: E0430 12:43:27.988688 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.988851 kubelet[4238]: E0430 12:43:27.988841 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.988851 kubelet[4238]: W0430 12:43:27.988850 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.988900 kubelet[4238]: E0430 12:43:27.988858 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.989014 kubelet[4238]: E0430 12:43:27.989005 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.989014 kubelet[4238]: W0430 12:43:27.989014 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.989075 kubelet[4238]: E0430 12:43:27.989022 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.989197 kubelet[4238]: E0430 12:43:27.989188 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.989197 kubelet[4238]: W0430 12:43:27.989196 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.989247 kubelet[4238]: E0430 12:43:27.989204 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.989419 kubelet[4238]: E0430 12:43:27.989411 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.989419 kubelet[4238]: W0430 12:43:27.989419 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.989477 kubelet[4238]: E0430 12:43:27.989426 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.989702 kubelet[4238]: E0430 12:43:27.989693 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.989702 kubelet[4238]: W0430 12:43:27.989702 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.989762 kubelet[4238]: E0430 12:43:27.989710 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.989900 kubelet[4238]: E0430 12:43:27.989891 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.989900 kubelet[4238]: W0430 12:43:27.989899 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.989950 kubelet[4238]: E0430 12:43:27.989911 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.990115 kubelet[4238]: E0430 12:43:27.990102 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.990139 kubelet[4238]: W0430 12:43:27.990116 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.990139 kubelet[4238]: E0430 12:43:27.990132 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.990292 kubelet[4238]: E0430 12:43:27.990283 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.990292 kubelet[4238]: W0430 12:43:27.990291 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.990340 kubelet[4238]: E0430 12:43:27.990301 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.990547 kubelet[4238]: E0430 12:43:27.990539 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.990568 kubelet[4238]: W0430 12:43:27.990547 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.990568 kubelet[4238]: E0430 12:43:27.990556 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.990837 kubelet[4238]: E0430 12:43:27.990828 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.990859 kubelet[4238]: W0430 12:43:27.990837 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.990859 kubelet[4238]: E0430 12:43:27.990848 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.991101 kubelet[4238]: E0430 12:43:27.991089 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.991101 kubelet[4238]: W0430 12:43:27.991101 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.991146 kubelet[4238]: E0430 12:43:27.991114 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.991418 kubelet[4238]: E0430 12:43:27.991409 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.991443 kubelet[4238]: W0430 12:43:27.991418 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.991464 kubelet[4238]: E0430 12:43:27.991442 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.991652 kubelet[4238]: E0430 12:43:27.991644 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.991673 kubelet[4238]: W0430 12:43:27.991652 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.991694 kubelet[4238]: E0430 12:43:27.991671 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.991892 kubelet[4238]: E0430 12:43:27.991884 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.991913 kubelet[4238]: W0430 12:43:27.991891 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.991913 kubelet[4238]: E0430 12:43:27.991902 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.992096 kubelet[4238]: E0430 12:43:27.992085 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.992117 kubelet[4238]: W0430 12:43:27.992097 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.992117 kubelet[4238]: E0430 12:43:27.992112 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.992261 kubelet[4238]: E0430 12:43:27.992249 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.992261 kubelet[4238]: W0430 12:43:27.992256 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.992301 kubelet[4238]: E0430 12:43:27.992266 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.992477 kubelet[4238]: E0430 12:43:27.992467 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.992497 kubelet[4238]: W0430 12:43:27.992476 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.992497 kubelet[4238]: E0430 12:43:27.992488 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.992695 kubelet[4238]: E0430 12:43:27.992686 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.992717 kubelet[4238]: W0430 12:43:27.992694 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.992717 kubelet[4238]: E0430 12:43:27.992705 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.992966 kubelet[4238]: E0430 12:43:27.992956 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.992986 kubelet[4238]: W0430 12:43:27.992966 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.992986 kubelet[4238]: E0430 12:43:27.992979 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.993206 kubelet[4238]: E0430 12:43:27.993199 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.993227 kubelet[4238]: W0430 12:43:27.993206 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.993227 kubelet[4238]: E0430 12:43:27.993214 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.993380 kubelet[4238]: E0430 12:43:27.993373 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.993400 kubelet[4238]: W0430 12:43:27.993380 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.993400 kubelet[4238]: E0430 12:43:27.993389 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:27.993696 kubelet[4238]: E0430 12:43:27.993687 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:27.993716 kubelet[4238]: W0430 12:43:27.993696 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:27.993716 kubelet[4238]: E0430 12:43:27.993705 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.901014 kubelet[4238]: I0430 12:43:28.900993 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:28.996265 kubelet[4238]: E0430 12:43:28.996237 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.996265 kubelet[4238]: W0430 12:43:28.996259 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.996398 kubelet[4238]: E0430 12:43:28.996276 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.996544 kubelet[4238]: E0430 12:43:28.996534 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.996544 kubelet[4238]: W0430 12:43:28.996542 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.996582 kubelet[4238]: E0430 12:43:28.996550 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.996817 kubelet[4238]: E0430 12:43:28.996809 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.996837 kubelet[4238]: W0430 12:43:28.996817 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.996837 kubelet[4238]: E0430 12:43:28.996825 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.996985 kubelet[4238]: E0430 12:43:28.996977 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.997008 kubelet[4238]: W0430 12:43:28.996985 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.997008 kubelet[4238]: E0430 12:43:28.996992 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.997270 kubelet[4238]: E0430 12:43:28.997262 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.997290 kubelet[4238]: W0430 12:43:28.997270 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.997290 kubelet[4238]: E0430 12:43:28.997278 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.997434 kubelet[4238]: E0430 12:43:28.997426 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.997454 kubelet[4238]: W0430 12:43:28.997434 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.997454 kubelet[4238]: E0430 12:43:28.997441 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.997706 kubelet[4238]: E0430 12:43:28.997699 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.997726 kubelet[4238]: W0430 12:43:28.997706 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.997726 kubelet[4238]: E0430 12:43:28.997713 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.997862 kubelet[4238]: E0430 12:43:28.997854 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.997883 kubelet[4238]: W0430 12:43:28.997862 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.997883 kubelet[4238]: E0430 12:43:28.997869 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.998033 kubelet[4238]: E0430 12:43:28.998025 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.998053 kubelet[4238]: W0430 12:43:28.998033 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.998053 kubelet[4238]: E0430 12:43:28.998041 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.998218 kubelet[4238]: E0430 12:43:28.998210 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.998241 kubelet[4238]: W0430 12:43:28.998218 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.998241 kubelet[4238]: E0430 12:43:28.998225 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.998433 kubelet[4238]: E0430 12:43:28.998425 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.998452 kubelet[4238]: W0430 12:43:28.998432 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.998452 kubelet[4238]: E0430 12:43:28.998439 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.998649 kubelet[4238]: E0430 12:43:28.998641 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.998671 kubelet[4238]: W0430 12:43:28.998648 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.998671 kubelet[4238]: E0430 12:43:28.998656 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.998817 kubelet[4238]: E0430 12:43:28.998809 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.998841 kubelet[4238]: W0430 12:43:28.998817 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.998841 kubelet[4238]: E0430 12:43:28.998824 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.999090 kubelet[4238]: E0430 12:43:28.999082 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.999110 kubelet[4238]: W0430 12:43:28.999090 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.999110 kubelet[4238]: E0430 12:43:28.999097 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.999325 kubelet[4238]: E0430 12:43:28.999318 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.999345 kubelet[4238]: W0430 12:43:28.999325 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.999345 kubelet[4238]: E0430 12:43:28.999332 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.999679 kubelet[4238]: E0430 12:43:28.999670 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.999698 kubelet[4238]: W0430 12:43:28.999679 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.999698 kubelet[4238]: E0430 12:43:28.999687 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:28.999961 kubelet[4238]: E0430 12:43:28.999953 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:28.999981 kubelet[4238]: W0430 12:43:28.999961 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:28.999981 kubelet[4238]: E0430 12:43:28.999972 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.000181 kubelet[4238]: E0430 12:43:29.000166 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.000201 kubelet[4238]: W0430 12:43:29.000182 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.000220 kubelet[4238]: E0430 12:43:29.000202 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.000364 kubelet[4238]: E0430 12:43:29.000356 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.000384 kubelet[4238]: W0430 12:43:29.000364 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.000384 kubelet[4238]: E0430 12:43:29.000375 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.000607 kubelet[4238]: E0430 12:43:29.000599 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.000627 kubelet[4238]: W0430 12:43:29.000607 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.000627 kubelet[4238]: E0430 12:43:29.000617 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.000871 kubelet[4238]: E0430 12:43:29.000863 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.000891 kubelet[4238]: W0430 12:43:29.000871 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.000891 kubelet[4238]: E0430 12:43:29.000882 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.001250 kubelet[4238]: E0430 12:43:29.001237 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.001273 kubelet[4238]: W0430 12:43:29.001251 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.001273 kubelet[4238]: E0430 12:43:29.001267 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.001595 kubelet[4238]: E0430 12:43:29.001586 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.001615 kubelet[4238]: W0430 12:43:29.001595 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.001637 kubelet[4238]: E0430 12:43:29.001614 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.001832 kubelet[4238]: E0430 12:43:29.001824 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.001852 kubelet[4238]: W0430 12:43:29.001832 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.001872 kubelet[4238]: E0430 12:43:29.001858 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.001985 kubelet[4238]: E0430 12:43:29.001977 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.002007 kubelet[4238]: W0430 12:43:29.001985 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.002007 kubelet[4238]: E0430 12:43:29.001996 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.002224 kubelet[4238]: E0430 12:43:29.002213 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.002245 kubelet[4238]: W0430 12:43:29.002225 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.002245 kubelet[4238]: E0430 12:43:29.002240 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.002498 kubelet[4238]: E0430 12:43:29.002490 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.002518 kubelet[4238]: W0430 12:43:29.002498 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.002518 kubelet[4238]: E0430 12:43:29.002509 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.002801 kubelet[4238]: E0430 12:43:29.002792 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.002821 kubelet[4238]: W0430 12:43:29.002801 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.002821 kubelet[4238]: E0430 12:43:29.002812 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.003128 kubelet[4238]: E0430 12:43:29.003120 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.003147 kubelet[4238]: W0430 12:43:29.003128 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.003147 kubelet[4238]: E0430 12:43:29.003139 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.003369 kubelet[4238]: E0430 12:43:29.003361 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.003389 kubelet[4238]: W0430 12:43:29.003369 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.003389 kubelet[4238]: E0430 12:43:29.003380 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.003578 kubelet[4238]: E0430 12:43:29.003570 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.003597 kubelet[4238]: W0430 12:43:29.003578 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.003597 kubelet[4238]: E0430 12:43:29.003589 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.003958 kubelet[4238]: E0430 12:43:29.003947 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.003979 kubelet[4238]: W0430 12:43:29.003959 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.003979 kubelet[4238]: E0430 12:43:29.003974 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.004272 kubelet[4238]: E0430 12:43:29.004264 4238 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Apr 30 12:43:29.004292 kubelet[4238]: W0430 12:43:29.004272 4238 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Apr 30 12:43:29.004292 kubelet[4238]: E0430 12:43:29.004283 4238 plugins.go:691] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Apr 30 12:43:29.379547 containerd[2700]: time="2025-04-30T12:43:29.379510476Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:29.379859 containerd[2700]: time="2025-04-30T12:43:29.379551157Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3: active requests=0, bytes read=5122903" Apr 30 12:43:29.380298 containerd[2700]: time="2025-04-30T12:43:29.380277417Z" level=info msg="ImageCreate event name:\"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:29.382071 containerd[2700]: time="2025-04-30T12:43:29.382050465Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:29.382764 containerd[2700]: time="2025-04-30T12:43:29.382743124Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" with image id \"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:eeaa2bb4f9b1aa61adde43ce6dea95eee89291f96963548e108d9a2dfbc5edd1\", size \"6492045\" in 2.392526823s" Apr 30 12:43:29.382805 containerd[2700]: time="2025-04-30T12:43:29.382770245Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.3\" returns image reference \"sha256:dd8e710a588cc6f5834c4d84f7e12458efae593d3dfe527ca9e757c89239ecb8\"" Apr 30 12:43:29.384320 containerd[2700]: time="2025-04-30T12:43:29.384300726Z" level=info msg="CreateContainer within sandbox \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Apr 30 12:43:29.390187 containerd[2700]: time="2025-04-30T12:43:29.390159805Z" level=info msg="CreateContainer within sandbox \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e\"" Apr 30 12:43:29.390505 containerd[2700]: time="2025-04-30T12:43:29.390480093Z" level=info msg="StartContainer for \"7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e\"" Apr 30 12:43:29.424907 systemd[1]: Started cri-containerd-7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e.scope - libcontainer container 7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e. Apr 30 12:43:29.445128 containerd[2700]: time="2025-04-30T12:43:29.445103493Z" level=info msg="StartContainer for \"7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e\" returns successfully" Apr 30 12:43:29.458257 systemd[1]: cri-containerd-7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e.scope: Deactivated successfully. Apr 30 12:43:29.596876 containerd[2700]: time="2025-04-30T12:43:29.596826484Z" level=info msg="shim disconnected" id=7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e namespace=k8s.io Apr 30 12:43:29.596876 containerd[2700]: time="2025-04-30T12:43:29.596874046Z" level=warning msg="cleaning up after shim disconnected" id=7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e namespace=k8s.io Apr 30 12:43:29.596876 containerd[2700]: time="2025-04-30T12:43:29.596881966Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 12:43:29.869303 kubelet[4238]: E0430 12:43:29.869249 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-nckqz" podUID="45ca62a1-734b-4197-a2e5-9df6c9a9ad92" Apr 30 12:43:29.904521 containerd[2700]: time="2025-04-30T12:43:29.904488460Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\"" Apr 30 12:43:30.390593 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7403f7657ebc2adc12d891817d4a533dc7509f35b47b1364813bd8cbc4953b1e-rootfs.mount: Deactivated successfully. Apr 30 12:43:31.047073 containerd[2700]: time="2025-04-30T12:43:31.047036491Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:31.047398 containerd[2700]: time="2025-04-30T12:43:31.047054172Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.3: active requests=0, bytes read=91256270" Apr 30 12:43:31.047735 containerd[2700]: time="2025-04-30T12:43:31.047714907Z" level=info msg="ImageCreate event name:\"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:31.051548 containerd[2700]: time="2025-04-30T12:43:31.051519358Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:31.052430 containerd[2700]: time="2025-04-30T12:43:31.052406779Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.3\" with image id \"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:4505ec8f976470994b6a94295a4dabac0cb98375db050e959a22603e00ada90b\", size \"92625452\" in 1.147884158s" Apr 30 12:43:31.052460 containerd[2700]: time="2025-04-30T12:43:31.052435940Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.3\" returns image reference \"sha256:add6372545fb406bb017769f222d84c50549ce13e3b19f1fbaee3d8a4aaef627\"" Apr 30 12:43:31.054277 containerd[2700]: time="2025-04-30T12:43:31.054253343Z" level=info msg="CreateContainer within sandbox \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Apr 30 12:43:31.060058 containerd[2700]: time="2025-04-30T12:43:31.060033721Z" level=info msg="CreateContainer within sandbox \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575\"" Apr 30 12:43:31.060446 containerd[2700]: time="2025-04-30T12:43:31.060423370Z" level=info msg="StartContainer for \"a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575\"" Apr 30 12:43:31.093934 systemd[1]: Started cri-containerd-a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575.scope - libcontainer container a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575. Apr 30 12:43:31.114609 containerd[2700]: time="2025-04-30T12:43:31.114572179Z" level=info msg="StartContainer for \"a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575\" returns successfully" Apr 30 12:43:31.487709 systemd[1]: cri-containerd-a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575.scope: Deactivated successfully. Apr 30 12:43:31.488025 systemd[1]: cri-containerd-a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575.scope: Consumed 840ms CPU time, 177.4M memory peak, 150.3M written to disk. Apr 30 12:43:31.502548 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575-rootfs.mount: Deactivated successfully. Apr 30 12:43:31.520662 kubelet[4238]: I0430 12:43:31.520632 4238 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Apr 30 12:43:31.537963 systemd[1]: Created slice kubepods-burstable-pod7cb08d0e_eb1c_4f65_869e_78b2a54830d9.slice - libcontainer container kubepods-burstable-pod7cb08d0e_eb1c_4f65_869e_78b2a54830d9.slice. Apr 30 12:43:31.542677 systemd[1]: Created slice kubepods-besteffort-pode620e8d6_60ca_49b3_b04c_d2b723f4466f.slice - libcontainer container kubepods-besteffort-pode620e8d6_60ca_49b3_b04c_d2b723f4466f.slice. Apr 30 12:43:31.545954 systemd[1]: Created slice kubepods-burstable-pod66fa8991_5cd2_45c8_9c64_43cbd360d9ba.slice - libcontainer container kubepods-burstable-pod66fa8991_5cd2_45c8_9c64_43cbd360d9ba.slice. Apr 30 12:43:31.550128 systemd[1]: Created slice kubepods-besteffort-pod4c75db99_dcfe_4abb_89af_789dc5ea3081.slice - libcontainer container kubepods-besteffort-pod4c75db99_dcfe_4abb_89af_789dc5ea3081.slice. Apr 30 12:43:31.553567 systemd[1]: Created slice kubepods-besteffort-pod608b409e_49ba_444f_b675_8d577d74ae39.slice - libcontainer container kubepods-besteffort-pod608b409e_49ba_444f_b675_8d577d74ae39.slice. Apr 30 12:43:31.620966 kubelet[4238]: I0430 12:43:31.620923 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/608b409e-49ba-444f-b675-8d577d74ae39-calico-apiserver-certs\") pod \"calico-apiserver-7d8d665688-9lz6f\" (UID: \"608b409e-49ba-444f-b675-8d577d74ae39\") " pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:31.620966 kubelet[4238]: I0430 12:43:31.620973 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7cb08d0e-eb1c-4f65-869e-78b2a54830d9-config-volume\") pod \"coredns-6f6b679f8f-j28hw\" (UID: \"7cb08d0e-eb1c-4f65-869e-78b2a54830d9\") " pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:31.621130 kubelet[4238]: I0430 12:43:31.620992 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rsxf\" (UniqueName: \"kubernetes.io/projected/7cb08d0e-eb1c-4f65-869e-78b2a54830d9-kube-api-access-5rsxf\") pod \"coredns-6f6b679f8f-j28hw\" (UID: \"7cb08d0e-eb1c-4f65-869e-78b2a54830d9\") " pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:31.621130 kubelet[4238]: I0430 12:43:31.621016 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/66fa8991-5cd2-45c8-9c64-43cbd360d9ba-config-volume\") pod \"coredns-6f6b679f8f-tphmp\" (UID: \"66fa8991-5cd2-45c8-9c64-43cbd360d9ba\") " pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:31.621130 kubelet[4238]: I0430 12:43:31.621032 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e620e8d6-60ca-49b3-b04c-d2b723f4466f-tigera-ca-bundle\") pod \"calico-kube-controllers-677bfd4f5b-6wq57\" (UID: \"e620e8d6-60ca-49b3-b04c-d2b723f4466f\") " pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:31.621202 kubelet[4238]: I0430 12:43:31.621150 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gltln\" (UniqueName: \"kubernetes.io/projected/e620e8d6-60ca-49b3-b04c-d2b723f4466f-kube-api-access-gltln\") pod \"calico-kube-controllers-677bfd4f5b-6wq57\" (UID: \"e620e8d6-60ca-49b3-b04c-d2b723f4466f\") " pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:31.621225 kubelet[4238]: I0430 12:43:31.621207 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7bzw\" (UniqueName: \"kubernetes.io/projected/4c75db99-dcfe-4abb-89af-789dc5ea3081-kube-api-access-v7bzw\") pod \"calico-apiserver-7d8d665688-fjfzq\" (UID: \"4c75db99-dcfe-4abb-89af-789dc5ea3081\") " pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:31.621265 kubelet[4238]: I0430 12:43:31.621247 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9dnk\" (UniqueName: \"kubernetes.io/projected/608b409e-49ba-444f-b675-8d577d74ae39-kube-api-access-l9dnk\") pod \"calico-apiserver-7d8d665688-9lz6f\" (UID: \"608b409e-49ba-444f-b675-8d577d74ae39\") " pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:31.621291 kubelet[4238]: I0430 12:43:31.621278 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/4c75db99-dcfe-4abb-89af-789dc5ea3081-calico-apiserver-certs\") pod \"calico-apiserver-7d8d665688-fjfzq\" (UID: \"4c75db99-dcfe-4abb-89af-789dc5ea3081\") " pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:31.621325 kubelet[4238]: I0430 12:43:31.621309 4238 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsdhp\" (UniqueName: \"kubernetes.io/projected/66fa8991-5cd2-45c8-9c64-43cbd360d9ba-kube-api-access-dsdhp\") pod \"coredns-6f6b679f8f-tphmp\" (UID: \"66fa8991-5cd2-45c8-9c64-43cbd360d9ba\") " pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:31.659855 containerd[2700]: time="2025-04-30T12:43:31.659805203Z" level=info msg="shim disconnected" id=a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575 namespace=k8s.io Apr 30 12:43:31.659901 containerd[2700]: time="2025-04-30T12:43:31.659855804Z" level=warning msg="cleaning up after shim disconnected" id=a429120ce1e7c61e7fdb5ff8f599c021a9ede4f39d9f0b75a08d9f905a637575 namespace=k8s.io Apr 30 12:43:31.659901 containerd[2700]: time="2025-04-30T12:43:31.659868125Z" level=info msg="cleaning up dead shim" namespace=k8s.io Apr 30 12:43:31.841461 containerd[2700]: time="2025-04-30T12:43:31.841417888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:0,}" Apr 30 12:43:31.844961 containerd[2700]: time="2025-04-30T12:43:31.844929012Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:0,}" Apr 30 12:43:31.848497 containerd[2700]: time="2025-04-30T12:43:31.848469056Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:0,}" Apr 30 12:43:31.852167 containerd[2700]: time="2025-04-30T12:43:31.852144823Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:0,}" Apr 30 12:43:31.855824 containerd[2700]: time="2025-04-30T12:43:31.855800350Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:0,}" Apr 30 12:43:31.873731 systemd[1]: Created slice kubepods-besteffort-pod45ca62a1_734b_4197_a2e5_9df6c9a9ad92.slice - libcontainer container kubepods-besteffort-pod45ca62a1_734b_4197_a2e5_9df6c9a9ad92.slice. Apr 30 12:43:31.875492 containerd[2700]: time="2025-04-30T12:43:31.875462459Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:0,}" Apr 30 12:43:31.901960 containerd[2700]: time="2025-04-30T12:43:31.901916969Z" level=error msg="Failed to destroy network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902162 containerd[2700]: time="2025-04-30T12:43:31.902134734Z" level=error msg="Failed to destroy network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902357 containerd[2700]: time="2025-04-30T12:43:31.902314138Z" level=error msg="Failed to destroy network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902468 containerd[2700]: time="2025-04-30T12:43:31.902322298Z" level=error msg="encountered an error cleaning up failed sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902528 containerd[2700]: time="2025-04-30T12:43:31.902511703Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902593 containerd[2700]: time="2025-04-30T12:43:31.902456822Z" level=error msg="encountered an error cleaning up failed sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902618 containerd[2700]: time="2025-04-30T12:43:31.902593665Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902723 kubelet[4238]: E0430 12:43:31.902691 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902770 containerd[2700]: time="2025-04-30T12:43:31.902697107Z" level=error msg="encountered an error cleaning up failed sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902770 containerd[2700]: time="2025-04-30T12:43:31.902751309Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902819 kubelet[4238]: E0430 12:43:31.902704 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902819 kubelet[4238]: E0430 12:43:31.902762 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:31.902819 kubelet[4238]: E0430 12:43:31.902783 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:31.902884 kubelet[4238]: E0430 12:43:31.902824 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" podUID="e620e8d6-60ca-49b3-b04c-d2b723f4466f" Apr 30 12:43:31.902884 kubelet[4238]: E0430 12:43:31.902763 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:31.902884 kubelet[4238]: E0430 12:43:31.902863 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:31.902964 kubelet[4238]: E0430 12:43:31.902880 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.902964 kubelet[4238]: E0430 12:43:31.902922 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:31.902964 kubelet[4238]: E0430 12:43:31.902939 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:31.903021 kubelet[4238]: E0430 12:43:31.902888 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" podUID="608b409e-49ba-444f-b675-8d577d74ae39" Apr 30 12:43:31.903021 kubelet[4238]: E0430 12:43:31.902980 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-j28hw" podUID="7cb08d0e-eb1c-4f65-869e-78b2a54830d9" Apr 30 12:43:31.903154 containerd[2700]: time="2025-04-30T12:43:31.903126117Z" level=error msg="Failed to destroy network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903241 containerd[2700]: time="2025-04-30T12:43:31.903126997Z" level=error msg="Failed to destroy network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903464 containerd[2700]: time="2025-04-30T12:43:31.903443045Z" level=error msg="encountered an error cleaning up failed sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903506 containerd[2700]: time="2025-04-30T12:43:31.903487646Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903554 containerd[2700]: time="2025-04-30T12:43:31.903496646Z" level=error msg="encountered an error cleaning up failed sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903585 containerd[2700]: time="2025-04-30T12:43:31.903574768Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903678 kubelet[4238]: E0430 12:43:31.903651 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903720 kubelet[4238]: E0430 12:43:31.903665 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.903720 kubelet[4238]: E0430 12:43:31.903685 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:31.903720 kubelet[4238]: E0430 12:43:31.903700 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:31.903720 kubelet[4238]: E0430 12:43:31.903701 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:31.903816 kubelet[4238]: E0430 12:43:31.903723 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:31.903816 kubelet[4238]: E0430 12:43:31.903727 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" podUID="4c75db99-dcfe-4abb-89af-789dc5ea3081" Apr 30 12:43:31.903816 kubelet[4238]: E0430 12:43:31.903759 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-tphmp" podUID="66fa8991-5cd2-45c8-9c64-43cbd360d9ba" Apr 30 12:43:31.906847 kubelet[4238]: I0430 12:43:31.906832 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229" Apr 30 12:43:31.907327 containerd[2700]: time="2025-04-30T12:43:31.907308897Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:43:31.907477 containerd[2700]: time="2025-04-30T12:43:31.907464541Z" level=info msg="Ensure that sandbox a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229 in task-service has been cleanup successfully" Apr 30 12:43:31.907650 containerd[2700]: time="2025-04-30T12:43:31.907635425Z" level=info msg="TearDown network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" successfully" Apr 30 12:43:31.907686 containerd[2700]: time="2025-04-30T12:43:31.907650825Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" returns successfully" Apr 30 12:43:31.907720 kubelet[4238]: I0430 12:43:31.907707 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064" Apr 30 12:43:31.908087 containerd[2700]: time="2025-04-30T12:43:31.908070595Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:1,}" Apr 30 12:43:31.908121 containerd[2700]: time="2025-04-30T12:43:31.908094436Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:43:31.908241 containerd[2700]: time="2025-04-30T12:43:31.908228159Z" level=info msg="Ensure that sandbox 5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064 in task-service has been cleanup successfully" Apr 30 12:43:31.908398 containerd[2700]: time="2025-04-30T12:43:31.908384483Z" level=info msg="TearDown network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" successfully" Apr 30 12:43:31.908424 containerd[2700]: time="2025-04-30T12:43:31.908399603Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" returns successfully" Apr 30 12:43:31.908720 containerd[2700]: time="2025-04-30T12:43:31.908702450Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:1,}" Apr 30 12:43:31.909721 containerd[2700]: time="2025-04-30T12:43:31.909703354Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\"" Apr 30 12:43:31.909860 kubelet[4238]: I0430 12:43:31.909847 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba" Apr 30 12:43:31.910213 containerd[2700]: time="2025-04-30T12:43:31.910197166Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:43:31.910342 containerd[2700]: time="2025-04-30T12:43:31.910329369Z" level=info msg="Ensure that sandbox 7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba in task-service has been cleanup successfully" Apr 30 12:43:31.910494 containerd[2700]: time="2025-04-30T12:43:31.910481733Z" level=info msg="TearDown network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" successfully" Apr 30 12:43:31.910525 containerd[2700]: time="2025-04-30T12:43:31.910494333Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" returns successfully" Apr 30 12:43:31.910705 kubelet[4238]: I0430 12:43:31.910663 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a" Apr 30 12:43:31.911607 containerd[2700]: time="2025-04-30T12:43:31.911585799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:1,}" Apr 30 12:43:31.911801 containerd[2700]: time="2025-04-30T12:43:31.911781564Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:43:31.911927 containerd[2700]: time="2025-04-30T12:43:31.911915447Z" level=info msg="Ensure that sandbox c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a in task-service has been cleanup successfully" Apr 30 12:43:31.912128 containerd[2700]: time="2025-04-30T12:43:31.912114331Z" level=info msg="TearDown network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" successfully" Apr 30 12:43:31.912148 containerd[2700]: time="2025-04-30T12:43:31.912128212Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" returns successfully" Apr 30 12:43:31.912929 kubelet[4238]: I0430 12:43:31.912906 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50" Apr 30 12:43:31.913591 containerd[2700]: time="2025-04-30T12:43:31.913569726Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:43:31.913636 containerd[2700]: time="2025-04-30T12:43:31.913615647Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:1,}" Apr 30 12:43:31.913838 containerd[2700]: time="2025-04-30T12:43:31.913815492Z" level=info msg="Ensure that sandbox ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50 in task-service has been cleanup successfully" Apr 30 12:43:31.914006 containerd[2700]: time="2025-04-30T12:43:31.913990096Z" level=info msg="TearDown network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" successfully" Apr 30 12:43:31.914029 containerd[2700]: time="2025-04-30T12:43:31.914006097Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" returns successfully" Apr 30 12:43:31.914369 containerd[2700]: time="2025-04-30T12:43:31.914347905Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:1,}" Apr 30 12:43:31.931252 containerd[2700]: time="2025-04-30T12:43:31.931130864Z" level=error msg="Failed to destroy network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.931826 containerd[2700]: time="2025-04-30T12:43:31.931801600Z" level=error msg="encountered an error cleaning up failed sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.931898 containerd[2700]: time="2025-04-30T12:43:31.931860482Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.932094 kubelet[4238]: E0430 12:43:31.932057 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.932138 kubelet[4238]: E0430 12:43:31.932118 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:31.932164 kubelet[4238]: E0430 12:43:31.932138 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:31.932201 kubelet[4238]: E0430 12:43:31.932177 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-nckqz_calico-system(45ca62a1-734b-4197-a2e5-9df6c9a9ad92)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-nckqz_calico-system(45ca62a1-734b-4197-a2e5-9df6c9a9ad92)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nckqz" podUID="45ca62a1-734b-4197-a2e5-9df6c9a9ad92" Apr 30 12:43:31.954941 containerd[2700]: time="2025-04-30T12:43:31.954891230Z" level=error msg="Failed to destroy network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955315 containerd[2700]: time="2025-04-30T12:43:31.955292160Z" level=error msg="encountered an error cleaning up failed sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955356 containerd[2700]: time="2025-04-30T12:43:31.955315400Z" level=error msg="Failed to destroy network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955377 containerd[2700]: time="2025-04-30T12:43:31.955359841Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955568 kubelet[4238]: E0430 12:43:31.955539 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955605 kubelet[4238]: E0430 12:43:31.955593 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:31.955625 kubelet[4238]: E0430 12:43:31.955612 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:31.955649 containerd[2700]: time="2025-04-30T12:43:31.955618407Z" level=error msg="encountered an error cleaning up failed sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955675 kubelet[4238]: E0430 12:43:31.955651 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" podUID="608b409e-49ba-444f-b675-8d577d74ae39" Apr 30 12:43:31.955710 containerd[2700]: time="2025-04-30T12:43:31.955665489Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955865 kubelet[4238]: E0430 12:43:31.955826 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.955905 kubelet[4238]: E0430 12:43:31.955887 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:31.955933 kubelet[4238]: E0430 12:43:31.955905 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:31.955964 kubelet[4238]: E0430 12:43:31.955943 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-tphmp" podUID="66fa8991-5cd2-45c8-9c64-43cbd360d9ba" Apr 30 12:43:31.958706 containerd[2700]: time="2025-04-30T12:43:31.958673000Z" level=error msg="Failed to destroy network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.959016 containerd[2700]: time="2025-04-30T12:43:31.958995168Z" level=error msg="encountered an error cleaning up failed sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.959058 containerd[2700]: time="2025-04-30T12:43:31.959042049Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.959181 kubelet[4238]: E0430 12:43:31.959160 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.959206 kubelet[4238]: E0430 12:43:31.959195 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:31.959229 kubelet[4238]: E0430 12:43:31.959212 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:31.959271 kubelet[4238]: E0430 12:43:31.959243 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" podUID="4c75db99-dcfe-4abb-89af-789dc5ea3081" Apr 30 12:43:31.962205 containerd[2700]: time="2025-04-30T12:43:31.962177884Z" level=error msg="Failed to destroy network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.962486 containerd[2700]: time="2025-04-30T12:43:31.962466211Z" level=error msg="encountered an error cleaning up failed sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.962527 containerd[2700]: time="2025-04-30T12:43:31.962512172Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.962651 kubelet[4238]: E0430 12:43:31.962620 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.962679 kubelet[4238]: E0430 12:43:31.962668 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:31.962705 kubelet[4238]: E0430 12:43:31.962684 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:31.962734 kubelet[4238]: E0430 12:43:31.962714 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-j28hw" podUID="7cb08d0e-eb1c-4f65-869e-78b2a54830d9" Apr 30 12:43:31.963981 containerd[2700]: time="2025-04-30T12:43:31.963958006Z" level=error msg="Failed to destroy network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.964283 containerd[2700]: time="2025-04-30T12:43:31.964262453Z" level=error msg="encountered an error cleaning up failed sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.964320 containerd[2700]: time="2025-04-30T12:43:31.964306374Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.964433 kubelet[4238]: E0430 12:43:31.964412 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:31.964457 kubelet[4238]: E0430 12:43:31.964447 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:31.964480 kubelet[4238]: E0430 12:43:31.964463 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:31.964511 kubelet[4238]: E0430 12:43:31.964494 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" podUID="e620e8d6-60ca-49b3-b04c-d2b723f4466f" Apr 30 12:43:32.915373 kubelet[4238]: I0430 12:43:32.915347 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748" Apr 30 12:43:32.915874 containerd[2700]: time="2025-04-30T12:43:32.915850752Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" Apr 30 12:43:32.916038 containerd[2700]: time="2025-04-30T12:43:32.916010795Z" level=info msg="Ensure that sandbox 000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748 in task-service has been cleanup successfully" Apr 30 12:43:32.916195 containerd[2700]: time="2025-04-30T12:43:32.916181119Z" level=info msg="TearDown network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" successfully" Apr 30 12:43:32.916217 containerd[2700]: time="2025-04-30T12:43:32.916193999Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" returns successfully" Apr 30 12:43:32.916239 kubelet[4238]: I0430 12:43:32.916183 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7" Apr 30 12:43:32.916389 containerd[2700]: time="2025-04-30T12:43:32.916370963Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:43:32.916453 containerd[2700]: time="2025-04-30T12:43:32.916443125Z" level=info msg="TearDown network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" successfully" Apr 30 12:43:32.916473 containerd[2700]: time="2025-04-30T12:43:32.916453845Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" returns successfully" Apr 30 12:43:32.916563 containerd[2700]: time="2025-04-30T12:43:32.916552447Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" Apr 30 12:43:32.916681 containerd[2700]: time="2025-04-30T12:43:32.916669770Z" level=info msg="Ensure that sandbox 1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7 in task-service has been cleanup successfully" Apr 30 12:43:32.916849 containerd[2700]: time="2025-04-30T12:43:32.916835494Z" level=info msg="TearDown network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" successfully" Apr 30 12:43:32.916878 containerd[2700]: time="2025-04-30T12:43:32.916852734Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" returns successfully" Apr 30 12:43:32.916910 containerd[2700]: time="2025-04-30T12:43:32.916839014Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:2,}" Apr 30 12:43:32.917136 containerd[2700]: time="2025-04-30T12:43:32.917123620Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:43:32.917194 containerd[2700]: time="2025-04-30T12:43:32.917184542Z" level=info msg="TearDown network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" successfully" Apr 30 12:43:32.917214 containerd[2700]: time="2025-04-30T12:43:32.917194742Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" returns successfully" Apr 30 12:43:32.917314 kubelet[4238]: I0430 12:43:32.917298 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4" Apr 30 12:43:32.917542 containerd[2700]: time="2025-04-30T12:43:32.917521589Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:2,}" Apr 30 12:43:32.917718 containerd[2700]: time="2025-04-30T12:43:32.917697553Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" Apr 30 12:43:32.917753 systemd[1]: run-netns-cni\x2d8e1559e7\x2d0b83\x2dab3d\x2dc68d\x2d5c4b6406d6cc.mount: Deactivated successfully. Apr 30 12:43:32.917929 containerd[2700]: time="2025-04-30T12:43:32.917871717Z" level=info msg="Ensure that sandbox 1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4 in task-service has been cleanup successfully" Apr 30 12:43:32.918047 containerd[2700]: time="2025-04-30T12:43:32.918033280Z" level=info msg="TearDown network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" successfully" Apr 30 12:43:32.918070 containerd[2700]: time="2025-04-30T12:43:32.918046881Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" returns successfully" Apr 30 12:43:32.918258 containerd[2700]: time="2025-04-30T12:43:32.918239805Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:43:32.918322 containerd[2700]: time="2025-04-30T12:43:32.918311847Z" level=info msg="TearDown network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" successfully" Apr 30 12:43:32.918344 containerd[2700]: time="2025-04-30T12:43:32.918322247Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" returns successfully" Apr 30 12:43:32.918453 kubelet[4238]: I0430 12:43:32.918439 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45" Apr 30 12:43:32.918693 containerd[2700]: time="2025-04-30T12:43:32.918674855Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:2,}" Apr 30 12:43:32.918804 containerd[2700]: time="2025-04-30T12:43:32.918789697Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" Apr 30 12:43:32.918930 containerd[2700]: time="2025-04-30T12:43:32.918917820Z" level=info msg="Ensure that sandbox c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45 in task-service has been cleanup successfully" Apr 30 12:43:32.919080 containerd[2700]: time="2025-04-30T12:43:32.919067504Z" level=info msg="TearDown network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" successfully" Apr 30 12:43:32.919101 containerd[2700]: time="2025-04-30T12:43:32.919080224Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" returns successfully" Apr 30 12:43:32.919273 containerd[2700]: time="2025-04-30T12:43:32.919258948Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:43:32.919292 kubelet[4238]: I0430 12:43:32.919264 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620" Apr 30 12:43:32.919335 containerd[2700]: time="2025-04-30T12:43:32.919325909Z" level=info msg="TearDown network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" successfully" Apr 30 12:43:32.919359 containerd[2700]: time="2025-04-30T12:43:32.919335350Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" returns successfully" Apr 30 12:43:32.919634 containerd[2700]: time="2025-04-30T12:43:32.919616836Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" Apr 30 12:43:32.919697 systemd[1]: run-netns-cni\x2df2a5e0f1\x2d395e\x2d0e52\x2df9c2\x2d1a6b7c71d20f.mount: Deactivated successfully. Apr 30 12:43:32.919761 containerd[2700]: time="2025-04-30T12:43:32.919738639Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:2,}" Apr 30 12:43:32.919782 containerd[2700]: time="2025-04-30T12:43:32.919764999Z" level=info msg="Ensure that sandbox d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620 in task-service has been cleanup successfully" Apr 30 12:43:32.919781 systemd[1]: run-netns-cni\x2db6e0a217\x2d471e\x2dcb5d\x2d89df\x2dda8a75bb4d87.mount: Deactivated successfully. Apr 30 12:43:32.919937 containerd[2700]: time="2025-04-30T12:43:32.919924403Z" level=info msg="TearDown network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" successfully" Apr 30 12:43:32.919959 containerd[2700]: time="2025-04-30T12:43:32.919937723Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" returns successfully" Apr 30 12:43:32.920133 containerd[2700]: time="2025-04-30T12:43:32.920118847Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:43:32.920194 containerd[2700]: time="2025-04-30T12:43:32.920183769Z" level=info msg="TearDown network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" successfully" Apr 30 12:43:32.920213 containerd[2700]: time="2025-04-30T12:43:32.920193969Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" returns successfully" Apr 30 12:43:32.920319 kubelet[4238]: I0430 12:43:32.920306 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb" Apr 30 12:43:32.920480 containerd[2700]: time="2025-04-30T12:43:32.920464895Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:2,}" Apr 30 12:43:32.920677 containerd[2700]: time="2025-04-30T12:43:32.920661899Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" Apr 30 12:43:32.920806 containerd[2700]: time="2025-04-30T12:43:32.920794142Z" level=info msg="Ensure that sandbox e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb in task-service has been cleanup successfully" Apr 30 12:43:32.920980 containerd[2700]: time="2025-04-30T12:43:32.920965266Z" level=info msg="TearDown network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" successfully" Apr 30 12:43:32.920999 containerd[2700]: time="2025-04-30T12:43:32.920980586Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" returns successfully" Apr 30 12:43:32.921317 containerd[2700]: time="2025-04-30T12:43:32.921302953Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:1,}" Apr 30 12:43:32.921868 systemd[1]: run-netns-cni\x2d8490981d\x2d9624\x2d835d\x2d29fc\x2d9bd0edc204c1.mount: Deactivated successfully. Apr 30 12:43:32.921940 systemd[1]: run-netns-cni\x2d13ffa5bd\x2d0bfe\x2daa80\x2d9d01\x2dea75f4bf3180.mount: Deactivated successfully. Apr 30 12:43:32.923888 systemd[1]: run-netns-cni\x2dcc49fb80\x2d30be\x2d184c\x2d4dbd\x2deb520d0d92af.mount: Deactivated successfully. Apr 30 12:43:33.002550 containerd[2700]: time="2025-04-30T12:43:33.002500484Z" level=error msg="Failed to destroy network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.002757 containerd[2700]: time="2025-04-30T12:43:33.002718968Z" level=error msg="Failed to destroy network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.002818 containerd[2700]: time="2025-04-30T12:43:33.002729649Z" level=error msg="Failed to destroy network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.002910 containerd[2700]: time="2025-04-30T12:43:33.002881532Z" level=error msg="Failed to destroy network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.002981 containerd[2700]: time="2025-04-30T12:43:33.002959533Z" level=error msg="encountered an error cleaning up failed sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003034 containerd[2700]: time="2025-04-30T12:43:33.003019655Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003107 containerd[2700]: time="2025-04-30T12:43:33.003085336Z" level=error msg="encountered an error cleaning up failed sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003152 containerd[2700]: time="2025-04-30T12:43:33.003135617Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003193 containerd[2700]: time="2025-04-30T12:43:33.003171738Z" level=error msg="encountered an error cleaning up failed sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003214 containerd[2700]: time="2025-04-30T12:43:33.003193138Z" level=error msg="encountered an error cleaning up failed sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003234 containerd[2700]: time="2025-04-30T12:43:33.003218019Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003272 containerd[2700]: time="2025-04-30T12:43:33.003235019Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003309 kubelet[4238]: E0430 12:43:33.003217 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003309 kubelet[4238]: E0430 12:43:33.003270 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:33.003309 kubelet[4238]: E0430 12:43:33.003272 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003388 kubelet[4238]: E0430 12:43:33.003323 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:33.003388 kubelet[4238]: E0430 12:43:33.003341 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:33.003388 kubelet[4238]: E0430 12:43:33.003291 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:33.003388 kubelet[4238]: E0430 12:43:33.003355 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003468 kubelet[4238]: E0430 12:43:33.003393 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:33.003468 kubelet[4238]: E0430 12:43:33.003393 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003468 kubelet[4238]: E0430 12:43:33.003438 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:33.003468 kubelet[4238]: E0430 12:43:33.003455 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:33.003558 containerd[2700]: time="2025-04-30T12:43:33.003414343Z" level=error msg="Failed to destroy network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.003581 kubelet[4238]: E0430 12:43:33.003389 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" podUID="4c75db99-dcfe-4abb-89af-789dc5ea3081" Apr 30 12:43:33.003581 kubelet[4238]: E0430 12:43:33.003412 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:33.003635 kubelet[4238]: E0430 12:43:33.003486 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" podUID="e620e8d6-60ca-49b3-b04c-d2b723f4466f" Apr 30 12:43:33.003635 kubelet[4238]: E0430 12:43:33.003381 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-nckqz_calico-system(45ca62a1-734b-4197-a2e5-9df6c9a9ad92)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-nckqz_calico-system(45ca62a1-734b-4197-a2e5-9df6c9a9ad92)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nckqz" podUID="45ca62a1-734b-4197-a2e5-9df6c9a9ad92" Apr 30 12:43:33.003730 kubelet[4238]: E0430 12:43:33.003531 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" podUID="608b409e-49ba-444f-b675-8d577d74ae39" Apr 30 12:43:33.003832 containerd[2700]: time="2025-04-30T12:43:33.003802751Z" level=error msg="Failed to destroy network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004065 containerd[2700]: time="2025-04-30T12:43:33.004041956Z" level=error msg="encountered an error cleaning up failed sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004104 containerd[2700]: time="2025-04-30T12:43:33.004089077Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004187 containerd[2700]: time="2025-04-30T12:43:33.004164799Z" level=error msg="encountered an error cleaning up failed sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004228 kubelet[4238]: E0430 12:43:33.004206 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004254 kubelet[4238]: E0430 12:43:33.004239 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:33.004280 containerd[2700]: time="2025-04-30T12:43:33.004222200Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004304 kubelet[4238]: E0430 12:43:33.004255 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:33.004304 kubelet[4238]: E0430 12:43:33.004283 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-j28hw" podUID="7cb08d0e-eb1c-4f65-869e-78b2a54830d9" Apr 30 12:43:33.004362 kubelet[4238]: E0430 12:43:33.004317 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.004362 kubelet[4238]: E0430 12:43:33.004332 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:33.004362 kubelet[4238]: E0430 12:43:33.004344 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:33.004426 kubelet[4238]: E0430 12:43:33.004367 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-tphmp" podUID="66fa8991-5cd2-45c8-9c64-43cbd360d9ba" Apr 30 12:43:33.923135 kubelet[4238]: I0430 12:43:33.923109 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470" Apr 30 12:43:33.923565 containerd[2700]: time="2025-04-30T12:43:33.923537321Z" level=info msg="StopPodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\"" Apr 30 12:43:33.923711 containerd[2700]: time="2025-04-30T12:43:33.923695044Z" level=info msg="Ensure that sandbox 2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470 in task-service has been cleanup successfully" Apr 30 12:43:33.923886 containerd[2700]: time="2025-04-30T12:43:33.923871168Z" level=info msg="TearDown network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" successfully" Apr 30 12:43:33.923910 containerd[2700]: time="2025-04-30T12:43:33.923886168Z" level=info msg="StopPodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" returns successfully" Apr 30 12:43:33.924318 containerd[2700]: time="2025-04-30T12:43:33.924295897Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" Apr 30 12:43:33.924387 containerd[2700]: time="2025-04-30T12:43:33.924374778Z" level=info msg="TearDown network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" successfully" Apr 30 12:43:33.924406 containerd[2700]: time="2025-04-30T12:43:33.924386899Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" returns successfully" Apr 30 12:43:33.924506 kubelet[4238]: I0430 12:43:33.924487 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8" Apr 30 12:43:33.924607 containerd[2700]: time="2025-04-30T12:43:33.924591343Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:43:33.924677 containerd[2700]: time="2025-04-30T12:43:33.924666344Z" level=info msg="TearDown network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" successfully" Apr 30 12:43:33.924701 containerd[2700]: time="2025-04-30T12:43:33.924677625Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" returns successfully" Apr 30 12:43:33.924888 containerd[2700]: time="2025-04-30T12:43:33.924868589Z" level=info msg="StopPodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\"" Apr 30 12:43:33.925024 containerd[2700]: time="2025-04-30T12:43:33.925010592Z" level=info msg="Ensure that sandbox 411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8 in task-service has been cleanup successfully" Apr 30 12:43:33.925063 containerd[2700]: time="2025-04-30T12:43:33.925046072Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:3,}" Apr 30 12:43:33.925204 containerd[2700]: time="2025-04-30T12:43:33.925186715Z" level=info msg="TearDown network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" successfully" Apr 30 12:43:33.925227 containerd[2700]: time="2025-04-30T12:43:33.925204036Z" level=info msg="StopPodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" returns successfully" Apr 30 12:43:33.925430 containerd[2700]: time="2025-04-30T12:43:33.925412240Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" Apr 30 12:43:33.925499 containerd[2700]: time="2025-04-30T12:43:33.925488802Z" level=info msg="TearDown network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" successfully" Apr 30 12:43:33.925495 systemd[1]: run-netns-cni\x2d7653f24c\x2d59de\x2d613d\x2dddf7\x2d39ad142d0f74.mount: Deactivated successfully. Apr 30 12:43:33.925662 containerd[2700]: time="2025-04-30T12:43:33.925499322Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" returns successfully" Apr 30 12:43:33.925703 containerd[2700]: time="2025-04-30T12:43:33.925677926Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:43:33.925771 containerd[2700]: time="2025-04-30T12:43:33.925757087Z" level=info msg="TearDown network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" successfully" Apr 30 12:43:33.925801 containerd[2700]: time="2025-04-30T12:43:33.925770888Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" returns successfully" Apr 30 12:43:33.926053 kubelet[4238]: I0430 12:43:33.926037 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd" Apr 30 12:43:33.926102 containerd[2700]: time="2025-04-30T12:43:33.926079134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:3,}" Apr 30 12:43:33.926446 containerd[2700]: time="2025-04-30T12:43:33.926429501Z" level=info msg="StopPodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\"" Apr 30 12:43:33.926573 containerd[2700]: time="2025-04-30T12:43:33.926560304Z" level=info msg="Ensure that sandbox aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd in task-service has been cleanup successfully" Apr 30 12:43:33.926748 containerd[2700]: time="2025-04-30T12:43:33.926731148Z" level=info msg="TearDown network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" successfully" Apr 30 12:43:33.926768 containerd[2700]: time="2025-04-30T12:43:33.926749068Z" level=info msg="StopPodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" returns successfully" Apr 30 12:43:33.926956 containerd[2700]: time="2025-04-30T12:43:33.926912791Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" Apr 30 12:43:33.927015 kubelet[4238]: I0430 12:43:33.927004 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8" Apr 30 12:43:33.927038 containerd[2700]: time="2025-04-30T12:43:33.927013514Z" level=info msg="TearDown network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" successfully" Apr 30 12:43:33.927038 containerd[2700]: time="2025-04-30T12:43:33.927023514Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" returns successfully" Apr 30 12:43:33.927226 containerd[2700]: time="2025-04-30T12:43:33.927207518Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:43:33.927303 containerd[2700]: time="2025-04-30T12:43:33.927291479Z" level=info msg="TearDown network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" successfully" Apr 30 12:43:33.927323 containerd[2700]: time="2025-04-30T12:43:33.927303240Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" returns successfully" Apr 30 12:43:33.927386 containerd[2700]: time="2025-04-30T12:43:33.927367321Z" level=info msg="StopPodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\"" Apr 30 12:43:33.927461 systemd[1]: run-netns-cni\x2de851739a\x2de214\x2d45bc\x2d7d1f\x2debdc630cfb34.mount: Deactivated successfully. Apr 30 12:43:33.927536 containerd[2700]: time="2025-04-30T12:43:33.927522084Z" level=info msg="Ensure that sandbox 027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8 in task-service has been cleanup successfully" Apr 30 12:43:33.927639 containerd[2700]: time="2025-04-30T12:43:33.927622606Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:3,}" Apr 30 12:43:33.927694 containerd[2700]: time="2025-04-30T12:43:33.927680327Z" level=info msg="TearDown network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" successfully" Apr 30 12:43:33.927714 containerd[2700]: time="2025-04-30T12:43:33.927694768Z" level=info msg="StopPodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" returns successfully" Apr 30 12:43:33.927891 containerd[2700]: time="2025-04-30T12:43:33.927876052Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" Apr 30 12:43:33.927947 containerd[2700]: time="2025-04-30T12:43:33.927937653Z" level=info msg="TearDown network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" successfully" Apr 30 12:43:33.928376 containerd[2700]: time="2025-04-30T12:43:33.927947213Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" returns successfully" Apr 30 12:43:33.928376 containerd[2700]: time="2025-04-30T12:43:33.928238739Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:2,}" Apr 30 12:43:33.928413 kubelet[4238]: I0430 12:43:33.928112 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe" Apr 30 12:43:33.928506 containerd[2700]: time="2025-04-30T12:43:33.928487464Z" level=info msg="StopPodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\"" Apr 30 12:43:33.928634 containerd[2700]: time="2025-04-30T12:43:33.928621627Z" level=info msg="Ensure that sandbox fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe in task-service has been cleanup successfully" Apr 30 12:43:33.928812 containerd[2700]: time="2025-04-30T12:43:33.928797431Z" level=info msg="TearDown network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" successfully" Apr 30 12:43:33.928832 containerd[2700]: time="2025-04-30T12:43:33.928812391Z" level=info msg="StopPodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" returns successfully" Apr 30 12:43:33.928989 containerd[2700]: time="2025-04-30T12:43:33.928975195Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" Apr 30 12:43:33.929050 containerd[2700]: time="2025-04-30T12:43:33.929039716Z" level=info msg="TearDown network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" successfully" Apr 30 12:43:33.929072 containerd[2700]: time="2025-04-30T12:43:33.929050436Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" returns successfully" Apr 30 12:43:33.929213 containerd[2700]: time="2025-04-30T12:43:33.929199279Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:43:33.929232 kubelet[4238]: I0430 12:43:33.929213 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f" Apr 30 12:43:33.929280 containerd[2700]: time="2025-04-30T12:43:33.929268761Z" level=info msg="TearDown network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" successfully" Apr 30 12:43:33.929302 containerd[2700]: time="2025-04-30T12:43:33.929280041Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" returns successfully" Apr 30 12:43:33.929540 containerd[2700]: time="2025-04-30T12:43:33.929527166Z" level=info msg="StopPodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\"" Apr 30 12:43:33.929578 containerd[2700]: time="2025-04-30T12:43:33.929560647Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:3,}" Apr 30 12:43:33.929672 containerd[2700]: time="2025-04-30T12:43:33.929659489Z" level=info msg="Ensure that sandbox f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f in task-service has been cleanup successfully" Apr 30 12:43:33.929825 containerd[2700]: time="2025-04-30T12:43:33.929812252Z" level=info msg="TearDown network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" successfully" Apr 30 12:43:33.929868 containerd[2700]: time="2025-04-30T12:43:33.929827132Z" level=info msg="StopPodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" returns successfully" Apr 30 12:43:33.930031 containerd[2700]: time="2025-04-30T12:43:33.930014056Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" Apr 30 12:43:33.930112 containerd[2700]: time="2025-04-30T12:43:33.930101938Z" level=info msg="TearDown network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" successfully" Apr 30 12:43:33.930132 containerd[2700]: time="2025-04-30T12:43:33.930112618Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" returns successfully" Apr 30 12:43:33.930306 systemd[1]: run-netns-cni\x2d0ec475ae\x2d9c4e\x2d50a4\x2d0cc5\x2dc9c4e344cf2c.mount: Deactivated successfully. Apr 30 12:43:33.930341 containerd[2700]: time="2025-04-30T12:43:33.930321583Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:43:33.930376 systemd[1]: run-netns-cni\x2da89d9306\x2d9694\x2de449\x2d8056\x2deb80e0ee7dd1.mount: Deactivated successfully. Apr 30 12:43:33.930407 containerd[2700]: time="2025-04-30T12:43:33.930398984Z" level=info msg="TearDown network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" successfully" Apr 30 12:43:33.930427 containerd[2700]: time="2025-04-30T12:43:33.930408945Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" returns successfully" Apr 30 12:43:33.930742 containerd[2700]: time="2025-04-30T12:43:33.930717751Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:3,}" Apr 30 12:43:33.933680 systemd[1]: run-netns-cni\x2d1ed3ac93\x2df3c5\x2d18ba\x2d122a\x2d5e96ab4a2f4c.mount: Deactivated successfully. Apr 30 12:43:33.933755 systemd[1]: run-netns-cni\x2d3a8224b7\x2d9c8f\x2d5653\x2db49b\x2de089a7d01366.mount: Deactivated successfully. Apr 30 12:43:33.971959 containerd[2700]: time="2025-04-30T12:43:33.971909053Z" level=error msg="Failed to destroy network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.972331 containerd[2700]: time="2025-04-30T12:43:33.972309502Z" level=error msg="encountered an error cleaning up failed sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.972398 containerd[2700]: time="2025-04-30T12:43:33.972364623Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.972603 kubelet[4238]: E0430 12:43:33.972574 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.972642 kubelet[4238]: E0430 12:43:33.972627 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:33.972664 kubelet[4238]: E0430 12:43:33.972649 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" Apr 30 12:43:33.972718 kubelet[4238]: E0430 12:43:33.972697 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-677bfd4f5b-6wq57_calico-system(e620e8d6-60ca-49b3-b04c-d2b723f4466f)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" podUID="e620e8d6-60ca-49b3-b04c-d2b723f4466f" Apr 30 12:43:33.976701 containerd[2700]: time="2025-04-30T12:43:33.976663273Z" level=error msg="Failed to destroy network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977130 containerd[2700]: time="2025-04-30T12:43:33.977106362Z" level=error msg="Failed to destroy network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977216 containerd[2700]: time="2025-04-30T12:43:33.977123202Z" level=error msg="encountered an error cleaning up failed sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977272 containerd[2700]: time="2025-04-30T12:43:33.977255325Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977449 containerd[2700]: time="2025-04-30T12:43:33.977427729Z" level=error msg="encountered an error cleaning up failed sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977471 kubelet[4238]: E0430 12:43:33.977425 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977501 kubelet[4238]: E0430 12:43:33.977474 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:33.977523 containerd[2700]: time="2025-04-30T12:43:33.977485690Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977548 kubelet[4238]: E0430 12:43:33.977494 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-tphmp" Apr 30 12:43:33.977548 kubelet[4238]: E0430 12:43:33.977532 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-tphmp_kube-system(66fa8991-5cd2-45c8-9c64-43cbd360d9ba)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-tphmp" podUID="66fa8991-5cd2-45c8-9c64-43cbd360d9ba" Apr 30 12:43:33.977610 kubelet[4238]: E0430 12:43:33.977589 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.977643 kubelet[4238]: E0430 12:43:33.977627 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:33.977669 kubelet[4238]: E0430 12:43:33.977646 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-6f6b679f8f-j28hw" Apr 30 12:43:33.977700 kubelet[4238]: E0430 12:43:33.977680 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-6f6b679f8f-j28hw_kube-system(7cb08d0e-eb1c-4f65-869e-78b2a54830d9)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-6f6b679f8f-j28hw" podUID="7cb08d0e-eb1c-4f65-869e-78b2a54830d9" Apr 30 12:43:33.978302 containerd[2700]: time="2025-04-30T12:43:33.978276146Z" level=error msg="Failed to destroy network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978492 containerd[2700]: time="2025-04-30T12:43:33.978463750Z" level=error msg="Failed to destroy network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978652 containerd[2700]: time="2025-04-30T12:43:33.978631914Z" level=error msg="encountered an error cleaning up failed sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978695 containerd[2700]: time="2025-04-30T12:43:33.978680515Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978820 containerd[2700]: time="2025-04-30T12:43:33.978799557Z" level=error msg="encountered an error cleaning up failed sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978841 kubelet[4238]: E0430 12:43:33.978801 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978863 kubelet[4238]: E0430 12:43:33.978845 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:33.978888 containerd[2700]: time="2025-04-30T12:43:33.978847518Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978910 kubelet[4238]: E0430 12:43:33.978861 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" Apr 30 12:43:33.978910 kubelet[4238]: E0430 12:43:33.978894 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-9lz6f_calico-apiserver(608b409e-49ba-444f-b675-8d577d74ae39)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" podUID="608b409e-49ba-444f-b675-8d577d74ae39" Apr 30 12:43:33.978963 kubelet[4238]: E0430 12:43:33.978931 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.978963 kubelet[4238]: E0430 12:43:33.978948 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:33.979005 kubelet[4238]: E0430 12:43:33.978961 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-nckqz" Apr 30 12:43:33.979005 kubelet[4238]: E0430 12:43:33.978988 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-nckqz_calico-system(45ca62a1-734b-4197-a2e5-9df6c9a9ad92)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-nckqz_calico-system(45ca62a1-734b-4197-a2e5-9df6c9a9ad92)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-nckqz" podUID="45ca62a1-734b-4197-a2e5-9df6c9a9ad92" Apr 30 12:43:33.979650 containerd[2700]: time="2025-04-30T12:43:33.979621015Z" level=error msg="Failed to destroy network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.979950 containerd[2700]: time="2025-04-30T12:43:33.979929941Z" level=error msg="encountered an error cleaning up failed sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.979988 containerd[2700]: time="2025-04-30T12:43:33.979973622Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.980106 kubelet[4238]: E0430 12:43:33.980087 4238 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Apr 30 12:43:33.980130 kubelet[4238]: E0430 12:43:33.980116 4238 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:33.980154 kubelet[4238]: E0430 12:43:33.980134 4238 kuberuntime_manager.go:1168] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" Apr 30 12:43:33.980175 kubelet[4238]: E0430 12:43:33.980160 4238 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-7d8d665688-fjfzq_calico-apiserver(4c75db99-dcfe-4abb-89af-789dc5ea3081)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" podUID="4c75db99-dcfe-4abb-89af-789dc5ea3081" Apr 30 12:43:34.246876 containerd[2700]: time="2025-04-30T12:43:34.246765844Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.3: active requests=0, bytes read=138981893" Apr 30 12:43:34.246876 containerd[2700]: time="2025-04-30T12:43:34.246780684Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:34.247609 containerd[2700]: time="2025-04-30T12:43:34.247579660Z" level=info msg="ImageCreate event name:\"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:34.249298 containerd[2700]: time="2025-04-30T12:43:34.249272693Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:34.249865 containerd[2700]: time="2025-04-30T12:43:34.249843224Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.3\" with image id \"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node@sha256:750e267b4f8217e0ca9e4107228370190d1a2499b72112ad04370ab9b4553916\", size \"138981755\" in 2.340111269s" Apr 30 12:43:34.249901 containerd[2700]: time="2025-04-30T12:43:34.249869945Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.3\" returns image reference \"sha256:cdcce3ec4624a24c28cdc07b0ee29ddf6703628edee7452a3f8a8b4816bfd057\"" Apr 30 12:43:34.256062 containerd[2700]: time="2025-04-30T12:43:34.256028026Z" level=info msg="CreateContainer within sandbox \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Apr 30 12:43:34.276171 containerd[2700]: time="2025-04-30T12:43:34.276142940Z" level=info msg="CreateContainer within sandbox \"1f464dbae667f3fd4978ba4198752f8229a52c39e6a4a6bcb79af9af40c1b5f5\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"731f8d32ad0f6bc50dd0e42bbb0259e551fcd8f27f04dd40c67d6246b5a00b92\"" Apr 30 12:43:34.276482 containerd[2700]: time="2025-04-30T12:43:34.276459427Z" level=info msg="StartContainer for \"731f8d32ad0f6bc50dd0e42bbb0259e551fcd8f27f04dd40c67d6246b5a00b92\"" Apr 30 12:43:34.305902 systemd[1]: Started cri-containerd-731f8d32ad0f6bc50dd0e42bbb0259e551fcd8f27f04dd40c67d6246b5a00b92.scope - libcontainer container 731f8d32ad0f6bc50dd0e42bbb0259e551fcd8f27f04dd40c67d6246b5a00b92. Apr 30 12:43:34.328661 containerd[2700]: time="2025-04-30T12:43:34.328628690Z" level=info msg="StartContainer for \"731f8d32ad0f6bc50dd0e42bbb0259e551fcd8f27f04dd40c67d6246b5a00b92\" returns successfully" Apr 30 12:43:34.393961 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4-shm.mount: Deactivated successfully. Apr 30 12:43:34.394046 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9-shm.mount: Deactivated successfully. Apr 30 12:43:34.394097 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1618552560.mount: Deactivated successfully. Apr 30 12:43:34.439091 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Apr 30 12:43:34.439138 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Apr 30 12:43:34.931622 kubelet[4238]: I0430 12:43:34.931570 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9" Apr 30 12:43:34.932049 containerd[2700]: time="2025-04-30T12:43:34.932000449Z" level=info msg="StopPodSandbox for \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\"" Apr 30 12:43:34.932309 containerd[2700]: time="2025-04-30T12:43:34.932169453Z" level=info msg="Ensure that sandbox 88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9 in task-service has been cleanup successfully" Apr 30 12:43:34.932364 containerd[2700]: time="2025-04-30T12:43:34.932348856Z" level=info msg="TearDown network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\" successfully" Apr 30 12:43:34.932627 containerd[2700]: time="2025-04-30T12:43:34.932363417Z" level=info msg="StopPodSandbox for \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\" returns successfully" Apr 30 12:43:34.932627 containerd[2700]: time="2025-04-30T12:43:34.932580421Z" level=info msg="StopPodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\"" Apr 30 12:43:34.932683 containerd[2700]: time="2025-04-30T12:43:34.932655982Z" level=info msg="TearDown network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" successfully" Apr 30 12:43:34.932683 containerd[2700]: time="2025-04-30T12:43:34.932665262Z" level=info msg="StopPodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" returns successfully" Apr 30 12:43:34.932848 containerd[2700]: time="2025-04-30T12:43:34.932834746Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" Apr 30 12:43:34.932909 containerd[2700]: time="2025-04-30T12:43:34.932897907Z" level=info msg="TearDown network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" successfully" Apr 30 12:43:34.932952 containerd[2700]: time="2025-04-30T12:43:34.932908347Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" returns successfully" Apr 30 12:43:34.932981 kubelet[4238]: I0430 12:43:34.932916 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4" Apr 30 12:43:34.933083 containerd[2700]: time="2025-04-30T12:43:34.933070030Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:43:34.933138 containerd[2700]: time="2025-04-30T12:43:34.933127032Z" level=info msg="TearDown network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" successfully" Apr 30 12:43:34.933165 containerd[2700]: time="2025-04-30T12:43:34.933137072Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" returns successfully" Apr 30 12:43:34.933470 containerd[2700]: time="2025-04-30T12:43:34.933453598Z" level=info msg="StopPodSandbox for \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\"" Apr 30 12:43:34.933506 containerd[2700]: time="2025-04-30T12:43:34.933482679Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:4,}" Apr 30 12:43:34.933611 containerd[2700]: time="2025-04-30T12:43:34.933592161Z" level=info msg="Ensure that sandbox d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4 in task-service has been cleanup successfully" Apr 30 12:43:34.933783 containerd[2700]: time="2025-04-30T12:43:34.933767204Z" level=info msg="TearDown network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\" successfully" Apr 30 12:43:34.933813 containerd[2700]: time="2025-04-30T12:43:34.933784004Z" level=info msg="StopPodSandbox for \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\" returns successfully" Apr 30 12:43:34.933946 systemd[1]: run-netns-cni\x2d3a3b8ea8\x2dae20\x2d76a6\x2df0dc\x2dd11115838622.mount: Deactivated successfully. Apr 30 12:43:34.934122 containerd[2700]: time="2025-04-30T12:43:34.934000809Z" level=info msg="StopPodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\"" Apr 30 12:43:34.934122 containerd[2700]: time="2025-04-30T12:43:34.934086530Z" level=info msg="TearDown network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" successfully" Apr 30 12:43:34.934122 containerd[2700]: time="2025-04-30T12:43:34.934096171Z" level=info msg="StopPodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" returns successfully" Apr 30 12:43:34.934199 kubelet[4238]: I0430 12:43:34.934168 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f" Apr 30 12:43:34.934343 containerd[2700]: time="2025-04-30T12:43:34.934328055Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" Apr 30 12:43:34.934413 containerd[2700]: time="2025-04-30T12:43:34.934402297Z" level=info msg="TearDown network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" successfully" Apr 30 12:43:34.934441 containerd[2700]: time="2025-04-30T12:43:34.934412737Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" returns successfully" Apr 30 12:43:34.934553 containerd[2700]: time="2025-04-30T12:43:34.934532419Z" level=info msg="StopPodSandbox for \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\"" Apr 30 12:43:34.934626 containerd[2700]: time="2025-04-30T12:43:34.934610221Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:43:34.934690 containerd[2700]: time="2025-04-30T12:43:34.934678302Z" level=info msg="TearDown network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" successfully" Apr 30 12:43:34.934714 containerd[2700]: time="2025-04-30T12:43:34.934692062Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" returns successfully" Apr 30 12:43:34.934734 containerd[2700]: time="2025-04-30T12:43:34.934685422Z" level=info msg="Ensure that sandbox c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f in task-service has been cleanup successfully" Apr 30 12:43:34.934894 containerd[2700]: time="2025-04-30T12:43:34.934880106Z" level=info msg="TearDown network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\" successfully" Apr 30 12:43:34.934940 containerd[2700]: time="2025-04-30T12:43:34.934894226Z" level=info msg="StopPodSandbox for \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\" returns successfully" Apr 30 12:43:34.935050 containerd[2700]: time="2025-04-30T12:43:34.935033909Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:4,}" Apr 30 12:43:34.935099 containerd[2700]: time="2025-04-30T12:43:34.935082790Z" level=info msg="StopPodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\"" Apr 30 12:43:34.935164 containerd[2700]: time="2025-04-30T12:43:34.935148151Z" level=info msg="TearDown network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" successfully" Apr 30 12:43:34.935192 containerd[2700]: time="2025-04-30T12:43:34.935162231Z" level=info msg="StopPodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" returns successfully" Apr 30 12:43:34.935327 containerd[2700]: time="2025-04-30T12:43:34.935309394Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" Apr 30 12:43:34.935395 containerd[2700]: time="2025-04-30T12:43:34.935384356Z" level=info msg="TearDown network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" successfully" Apr 30 12:43:34.935472 containerd[2700]: time="2025-04-30T12:43:34.935394436Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" returns successfully" Apr 30 12:43:34.935568 containerd[2700]: time="2025-04-30T12:43:34.935548839Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:43:34.935599 kubelet[4238]: I0430 12:43:34.935570 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba" Apr 30 12:43:34.935641 containerd[2700]: time="2025-04-30T12:43:34.935629521Z" level=info msg="TearDown network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" successfully" Apr 30 12:43:34.935670 containerd[2700]: time="2025-04-30T12:43:34.935640081Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" returns successfully" Apr 30 12:43:34.935866 systemd[1]: run-netns-cni\x2de17c41ec\x2deed2\x2d11bc\x2d4683\x2d5ecb3cf9f019.mount: Deactivated successfully. Apr 30 12:43:34.935945 systemd[1]: run-netns-cni\x2dea2b4602\x2d45bb\x2d872a\x2d5c74\x2d94446b7509fa.mount: Deactivated successfully. Apr 30 12:43:34.935977 containerd[2700]: time="2025-04-30T12:43:34.935955527Z" level=info msg="StopPodSandbox for \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\"" Apr 30 12:43:34.936009 containerd[2700]: time="2025-04-30T12:43:34.935992368Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:4,}" Apr 30 12:43:34.936110 containerd[2700]: time="2025-04-30T12:43:34.936096370Z" level=info msg="Ensure that sandbox ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba in task-service has been cleanup successfully" Apr 30 12:43:34.936321 containerd[2700]: time="2025-04-30T12:43:34.936306854Z" level=info msg="TearDown network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\" successfully" Apr 30 12:43:34.936348 containerd[2700]: time="2025-04-30T12:43:34.936320654Z" level=info msg="StopPodSandbox for \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\" returns successfully" Apr 30 12:43:34.936554 containerd[2700]: time="2025-04-30T12:43:34.936534698Z" level=info msg="StopPodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\"" Apr 30 12:43:34.936627 containerd[2700]: time="2025-04-30T12:43:34.936615940Z" level=info msg="TearDown network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" successfully" Apr 30 12:43:34.936648 containerd[2700]: time="2025-04-30T12:43:34.936626540Z" level=info msg="StopPodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" returns successfully" Apr 30 12:43:34.936822 containerd[2700]: time="2025-04-30T12:43:34.936809184Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" Apr 30 12:43:34.936867 kubelet[4238]: I0430 12:43:34.936854 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4" Apr 30 12:43:34.936902 containerd[2700]: time="2025-04-30T12:43:34.936877665Z" level=info msg="TearDown network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" successfully" Apr 30 12:43:34.936902 containerd[2700]: time="2025-04-30T12:43:34.936886665Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" returns successfully" Apr 30 12:43:34.937078 containerd[2700]: time="2025-04-30T12:43:34.937063869Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:43:34.937140 containerd[2700]: time="2025-04-30T12:43:34.937129030Z" level=info msg="TearDown network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" successfully" Apr 30 12:43:34.937161 containerd[2700]: time="2025-04-30T12:43:34.937140150Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" returns successfully" Apr 30 12:43:34.937270 containerd[2700]: time="2025-04-30T12:43:34.937252672Z" level=info msg="StopPodSandbox for \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\"" Apr 30 12:43:34.937420 containerd[2700]: time="2025-04-30T12:43:34.937406836Z" level=info msg="Ensure that sandbox 63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4 in task-service has been cleanup successfully" Apr 30 12:43:34.937446 containerd[2700]: time="2025-04-30T12:43:34.937413796Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:4,}" Apr 30 12:43:34.937583 containerd[2700]: time="2025-04-30T12:43:34.937569959Z" level=info msg="TearDown network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\" successfully" Apr 30 12:43:34.937604 containerd[2700]: time="2025-04-30T12:43:34.937585159Z" level=info msg="StopPodSandbox for \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\" returns successfully" Apr 30 12:43:34.937795 containerd[2700]: time="2025-04-30T12:43:34.937779963Z" level=info msg="StopPodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\"" Apr 30 12:43:34.937872 containerd[2700]: time="2025-04-30T12:43:34.937861604Z" level=info msg="TearDown network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" successfully" Apr 30 12:43:34.937893 containerd[2700]: time="2025-04-30T12:43:34.937871765Z" level=info msg="StopPodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" returns successfully" Apr 30 12:43:34.938066 containerd[2700]: time="2025-04-30T12:43:34.938049168Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" Apr 30 12:43:34.938152 containerd[2700]: time="2025-04-30T12:43:34.938140290Z" level=info msg="TearDown network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" successfully" Apr 30 12:43:34.938174 containerd[2700]: time="2025-04-30T12:43:34.938153570Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" returns successfully" Apr 30 12:43:34.938476 containerd[2700]: time="2025-04-30T12:43:34.938458496Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:3,}" Apr 30 12:43:34.938765 systemd[1]: run-netns-cni\x2dabba4330\x2d87ac\x2d81a4\x2df728\x2d5fd0f03ab77f.mount: Deactivated successfully. Apr 30 12:43:34.940497 kubelet[4238]: I0430 12:43:34.940484 4238 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4" Apr 30 12:43:34.940881 containerd[2700]: time="2025-04-30T12:43:34.940861543Z" level=info msg="StopPodSandbox for \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\"" Apr 30 12:43:34.941016 containerd[2700]: time="2025-04-30T12:43:34.941003106Z" level=info msg="Ensure that sandbox 27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4 in task-service has been cleanup successfully" Apr 30 12:43:34.941171 containerd[2700]: time="2025-04-30T12:43:34.941158629Z" level=info msg="TearDown network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\" successfully" Apr 30 12:43:34.941193 containerd[2700]: time="2025-04-30T12:43:34.941171469Z" level=info msg="StopPodSandbox for \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\" returns successfully" Apr 30 12:43:34.942121 systemd[1]: run-netns-cni\x2d9f45f16c\x2dbad2\x2dde36\x2d7d2f\x2d363a3694fca4.mount: Deactivated successfully. Apr 30 12:43:34.942352 containerd[2700]: time="2025-04-30T12:43:34.942325612Z" level=info msg="StopPodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\"" Apr 30 12:43:34.942464 containerd[2700]: time="2025-04-30T12:43:34.942451214Z" level=info msg="TearDown network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" successfully" Apr 30 12:43:34.942486 containerd[2700]: time="2025-04-30T12:43:34.942465375Z" level=info msg="StopPodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" returns successfully" Apr 30 12:43:34.943148 containerd[2700]: time="2025-04-30T12:43:34.943108027Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" Apr 30 12:43:34.943802 containerd[2700]: time="2025-04-30T12:43:34.943778761Z" level=info msg="TearDown network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" successfully" Apr 30 12:43:34.943826 containerd[2700]: time="2025-04-30T12:43:34.943804481Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" returns successfully" Apr 30 12:43:34.944091 containerd[2700]: time="2025-04-30T12:43:34.944076566Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:43:34.944155 containerd[2700]: time="2025-04-30T12:43:34.944144648Z" level=info msg="TearDown network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" successfully" Apr 30 12:43:34.944176 containerd[2700]: time="2025-04-30T12:43:34.944154728Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" returns successfully" Apr 30 12:43:34.944542 containerd[2700]: time="2025-04-30T12:43:34.944524015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:4,}" Apr 30 12:43:34.949230 kubelet[4238]: I0430 12:43:34.949186 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-b4txl" podStartSLOduration=1.612582158 podStartE2EDuration="9.949172306s" podCreationTimestamp="2025-04-30 12:43:25 +0000 UTC" firstStartedPulling="2025-04-30 12:43:25.914575742 +0000 UTC m=+17.113935486" lastFinishedPulling="2025-04-30 12:43:34.25116585 +0000 UTC m=+25.450525634" observedRunningTime="2025-04-30 12:43:34.948818779 +0000 UTC m=+26.148178563" watchObservedRunningTime="2025-04-30 12:43:34.949172306 +0000 UTC m=+26.148532090" Apr 30 12:43:35.049813 systemd-networkd[2606]: cali3a3d9b22f7c: Link UP Apr 30 12:43:35.050174 systemd-networkd[2606]: cali3a3d9b22f7c: Gained carrier Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:34.961 [INFO][6872] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:34.973 [INFO][6872] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0 calico-apiserver-7d8d665688- calico-apiserver 608b409e-49ba-444f-b675-8d577d74ae39 657 0 2025-04-30 12:43:24 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7d8d665688 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4230.1.1-a-0ae8677ef9 calico-apiserver-7d8d665688-9lz6f eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali3a3d9b22f7c [] []}} ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:34.974 [INFO][6872] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.008 [INFO][7002] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" HandleID="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7002] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" HandleID="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40006a0d80), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4230.1.1-a-0ae8677ef9", "pod":"calico-apiserver-7d8d665688-9lz6f", "timestamp":"2025-04-30 12:43:35.008176735 +0000 UTC"}, Hostname:"ci-4230.1.1-a-0ae8677ef9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7002] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7002] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7002] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230.1.1-a-0ae8677ef9' Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.018 [INFO][7002] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.021 [INFO][7002] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.024 [INFO][7002] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.025 [INFO][7002] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.027 [INFO][7002] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.027 [INFO][7002] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.028 [INFO][7002] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99 Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.030 [INFO][7002] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.036 [INFO][7002] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.65/26] block=192.168.106.64/26 handle="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.036 [INFO][7002] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.65/26] handle="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.036 [INFO][7002] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 12:43:35.057916 containerd[2700]: 2025-04-30 12:43:35.036 [INFO][7002] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.65/26] IPv6=[] ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" HandleID="k8s-pod-network.5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.058325 containerd[2700]: 2025-04-30 12:43:35.039 [INFO][6872] cni-plugin/k8s.go 386: Populated endpoint ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0", GenerateName:"calico-apiserver-7d8d665688-", Namespace:"calico-apiserver", SelfLink:"", UID:"608b409e-49ba-444f-b675-8d577d74ae39", ResourceVersion:"657", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d8d665688", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"", Pod:"calico-apiserver-7d8d665688-9lz6f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.106.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3a3d9b22f7c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.058325 containerd[2700]: 2025-04-30 12:43:35.043 [INFO][6872] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.65/32] ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.058325 containerd[2700]: 2025-04-30 12:43:35.043 [INFO][6872] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3a3d9b22f7c ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.058325 containerd[2700]: 2025-04-30 12:43:35.050 [INFO][6872] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.058325 containerd[2700]: 2025-04-30 12:43:35.050 [INFO][6872] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0", GenerateName:"calico-apiserver-7d8d665688-", Namespace:"calico-apiserver", SelfLink:"", UID:"608b409e-49ba-444f-b675-8d577d74ae39", ResourceVersion:"657", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d8d665688", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99", Pod:"calico-apiserver-7d8d665688-9lz6f", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.106.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali3a3d9b22f7c", MAC:"0a:89:37:5a:b3:bb", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.058325 containerd[2700]: 2025-04-30 12:43:35.056 [INFO][6872] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-9lz6f" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--9lz6f-eth0" Apr 30 12:43:35.071863 containerd[2700]: time="2025-04-30T12:43:35.071802345Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:35.071863 containerd[2700]: time="2025-04-30T12:43:35.071857346Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:35.071959 containerd[2700]: time="2025-04-30T12:43:35.071869467Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.071959 containerd[2700]: time="2025-04-30T12:43:35.071939068Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.092863 systemd[1]: Started cri-containerd-5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99.scope - libcontainer container 5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99. Apr 30 12:43:35.116728 containerd[2700]: time="2025-04-30T12:43:35.116689371Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-9lz6f,Uid:608b409e-49ba-444f-b675-8d577d74ae39,Namespace:calico-apiserver,Attempt:4,} returns sandbox id \"5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99\"" Apr 30 12:43:35.117728 containerd[2700]: time="2025-04-30T12:43:35.117706190Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" Apr 30 12:43:35.147653 systemd-networkd[2606]: cali3793b446d51: Link UP Apr 30 12:43:35.147808 systemd-networkd[2606]: cali3793b446d51: Gained carrier Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:34.957 [INFO][6841] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:34.974 [INFO][6841] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0 calico-kube-controllers-677bfd4f5b- calico-system e620e8d6-60ca-49b3-b04c-d2b723f4466f 656 0 2025-04-30 12:43:25 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:677bfd4f5b projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4230.1.1-a-0ae8677ef9 calico-kube-controllers-677bfd4f5b-6wq57 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali3793b446d51 [] []}} ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:34.974 [INFO][6841] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.008 [INFO][7007] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" HandleID="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7007] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" HandleID="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40003880e0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4230.1.1-a-0ae8677ef9", "pod":"calico-kube-controllers-677bfd4f5b-6wq57", "timestamp":"2025-04-30 12:43:35.008167415 +0000 UTC"}, Hostname:"ci-4230.1.1-a-0ae8677ef9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7007] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.036 [INFO][7007] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.036 [INFO][7007] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230.1.1-a-0ae8677ef9' Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.119 [INFO][7007] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.122 [INFO][7007] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.125 [INFO][7007] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.126 [INFO][7007] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.128 [INFO][7007] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.128 [INFO][7007] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.129 [INFO][7007] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8 Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.141 [INFO][7007] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.145 [INFO][7007] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.66/26] block=192.168.106.64/26 handle="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.145 [INFO][7007] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.66/26] handle="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.145 [INFO][7007] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 12:43:35.154887 containerd[2700]: 2025-04-30 12:43:35.145 [INFO][7007] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.66/26] IPv6=[] ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" HandleID="k8s-pod-network.e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.155299 containerd[2700]: 2025-04-30 12:43:35.146 [INFO][6841] cni-plugin/k8s.go 386: Populated endpoint ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0", GenerateName:"calico-kube-controllers-677bfd4f5b-", Namespace:"calico-system", SelfLink:"", UID:"e620e8d6-60ca-49b3-b04c-d2b723f4466f", ResourceVersion:"656", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 25, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"677bfd4f5b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"", Pod:"calico-kube-controllers-677bfd4f5b-6wq57", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.106.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali3793b446d51", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.155299 containerd[2700]: 2025-04-30 12:43:35.146 [INFO][6841] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.66/32] ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.155299 containerd[2700]: 2025-04-30 12:43:35.146 [INFO][6841] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali3793b446d51 ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.155299 containerd[2700]: 2025-04-30 12:43:35.147 [INFO][6841] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.155299 containerd[2700]: 2025-04-30 12:43:35.148 [INFO][6841] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0", GenerateName:"calico-kube-controllers-677bfd4f5b-", Namespace:"calico-system", SelfLink:"", UID:"e620e8d6-60ca-49b3-b04c-d2b723f4466f", ResourceVersion:"656", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 25, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"677bfd4f5b", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8", Pod:"calico-kube-controllers-677bfd4f5b-6wq57", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.106.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali3793b446d51", MAC:"5a:a9:ca:61:30:ce", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.155299 containerd[2700]: 2025-04-30 12:43:35.153 [INFO][6841] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8" Namespace="calico-system" Pod="calico-kube-controllers-677bfd4f5b-6wq57" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--kube--controllers--677bfd4f5b--6wq57-eth0" Apr 30 12:43:35.180604 containerd[2700]: time="2025-04-30T12:43:35.180540746Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:35.180644 containerd[2700]: time="2025-04-30T12:43:35.180600067Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:35.180644 containerd[2700]: time="2025-04-30T12:43:35.180611227Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.180706 containerd[2700]: time="2025-04-30T12:43:35.180690708Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.200938 systemd[1]: Started cri-containerd-e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8.scope - libcontainer container e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8. Apr 30 12:43:35.224485 containerd[2700]: time="2025-04-30T12:43:35.224458514Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-677bfd4f5b-6wq57,Uid:e620e8d6-60ca-49b3-b04c-d2b723f4466f,Namespace:calico-system,Attempt:4,} returns sandbox id \"e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8\"" Apr 30 12:43:35.239944 systemd-networkd[2606]: cali2d3a4f8ce4b: Link UP Apr 30 12:43:35.240082 systemd-networkd[2606]: cali2d3a4f8ce4b: Gained carrier Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:34.956 [INFO][6843] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:34.973 [INFO][6843] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0 coredns-6f6b679f8f- kube-system 7cb08d0e-eb1c-4f65-869e-78b2a54830d9 651 0 2025-04-30 12:43:14 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4230.1.1-a-0ae8677ef9 coredns-6f6b679f8f-j28hw eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali2d3a4f8ce4b [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:34.974 [INFO][6843] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.008 [INFO][7004] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" HandleID="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7004] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" HandleID="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000583330), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4230.1.1-a-0ae8677ef9", "pod":"coredns-6f6b679f8f-j28hw", "timestamp":"2025-04-30 12:43:35.008168415 +0000 UTC"}, Hostname:"ci-4230.1.1-a-0ae8677ef9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7004] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.145 [INFO][7004] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.145 [INFO][7004] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230.1.1-a-0ae8677ef9' Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.219 [INFO][7004] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.223 [INFO][7004] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.226 [INFO][7004] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.227 [INFO][7004] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.228 [INFO][7004] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.228 [INFO][7004] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.229 [INFO][7004] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1 Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.232 [INFO][7004] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.236 [INFO][7004] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.67/26] block=192.168.106.64/26 handle="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.236 [INFO][7004] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.67/26] handle="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.236 [INFO][7004] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 12:43:35.246913 containerd[2700]: 2025-04-30 12:43:35.236 [INFO][7004] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.67/26] IPv6=[] ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" HandleID="k8s-pod-network.7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.247311 containerd[2700]: 2025-04-30 12:43:35.238 [INFO][6843] cni-plugin/k8s.go 386: Populated endpoint ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"7cb08d0e-eb1c-4f65-869e-78b2a54830d9", ResourceVersion:"651", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"", Pod:"coredns-6f6b679f8f-j28hw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.106.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2d3a4f8ce4b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.247311 containerd[2700]: 2025-04-30 12:43:35.238 [INFO][6843] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.67/32] ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.247311 containerd[2700]: 2025-04-30 12:43:35.238 [INFO][6843] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali2d3a4f8ce4b ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.247311 containerd[2700]: 2025-04-30 12:43:35.240 [INFO][6843] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.247311 containerd[2700]: 2025-04-30 12:43:35.240 [INFO][6843] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"7cb08d0e-eb1c-4f65-869e-78b2a54830d9", ResourceVersion:"651", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1", Pod:"coredns-6f6b679f8f-j28hw", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.106.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali2d3a4f8ce4b", MAC:"1e:6e:3a:2f:a3:93", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.247311 containerd[2700]: 2025-04-30 12:43:35.245 [INFO][6843] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1" Namespace="kube-system" Pod="coredns-6f6b679f8f-j28hw" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--j28hw-eth0" Apr 30 12:43:35.261117 containerd[2700]: time="2025-04-30T12:43:35.261050187Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:35.261154 containerd[2700]: time="2025-04-30T12:43:35.261113268Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:35.261154 containerd[2700]: time="2025-04-30T12:43:35.261125388Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.261225 containerd[2700]: time="2025-04-30T12:43:35.261206710Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.288857 systemd[1]: Started cri-containerd-7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1.scope - libcontainer container 7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1. Apr 30 12:43:35.312779 containerd[2700]: time="2025-04-30T12:43:35.312732777Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-j28hw,Uid:7cb08d0e-eb1c-4f65-869e-78b2a54830d9,Namespace:kube-system,Attempt:4,} returns sandbox id \"7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1\"" Apr 30 12:43:35.314614 containerd[2700]: time="2025-04-30T12:43:35.314592252Z" level=info msg="CreateContainer within sandbox \"7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 30 12:43:35.320309 containerd[2700]: time="2025-04-30T12:43:35.320281396Z" level=info msg="CreateContainer within sandbox \"7b97f4b916a66473014bea933580f281232cfd67f0425077a5382d15929b5eb1\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"457899f27d850cd062008b89a501a54f97b5a7eeadedc72d6417de44e01f11ee\"" Apr 30 12:43:35.320622 containerd[2700]: time="2025-04-30T12:43:35.320600482Z" level=info msg="StartContainer for \"457899f27d850cd062008b89a501a54f97b5a7eeadedc72d6417de44e01f11ee\"" Apr 30 12:43:35.341072 systemd-networkd[2606]: cali29429174561: Link UP Apr 30 12:43:35.341234 systemd-networkd[2606]: cali29429174561: Gained carrier Apr 30 12:43:35.345868 systemd[1]: Started cri-containerd-457899f27d850cd062008b89a501a54f97b5a7eeadedc72d6417de44e01f11ee.scope - libcontainer container 457899f27d850cd062008b89a501a54f97b5a7eeadedc72d6417de44e01f11ee. Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:34.963 [INFO][6892] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:34.978 [INFO][6892] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0 coredns-6f6b679f8f- kube-system 66fa8991-5cd2-45c8-9c64-43cbd360d9ba 654 0 2025-04-30 12:43:14 +0000 UTC map[k8s-app:kube-dns pod-template-hash:6f6b679f8f projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4230.1.1-a-0ae8677ef9 coredns-6f6b679f8f-tphmp eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali29429174561 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:34.978 [INFO][6892] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.008 [INFO][7027] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" HandleID="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7027] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" HandleID="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400004c100), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4230.1.1-a-0ae8677ef9", "pod":"coredns-6f6b679f8f-tphmp", "timestamp":"2025-04-30 12:43:35.008169975 +0000 UTC"}, Hostname:"ci-4230.1.1-a-0ae8677ef9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.017 [INFO][7027] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.236 [INFO][7027] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.236 [INFO][7027] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230.1.1-a-0ae8677ef9' Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.319 [INFO][7027] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.323 [INFO][7027] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.327 [INFO][7027] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.329 [INFO][7027] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.330 [INFO][7027] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.330 [INFO][7027] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.331 [INFO][7027] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3 Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.333 [INFO][7027] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.338 [INFO][7027] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.68/26] block=192.168.106.64/26 handle="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.338 [INFO][7027] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.68/26] handle="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.338 [INFO][7027] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 12:43:35.347995 containerd[2700]: 2025-04-30 12:43:35.338 [INFO][7027] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.68/26] IPv6=[] ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" HandleID="k8s-pod-network.0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.348401 containerd[2700]: 2025-04-30 12:43:35.339 [INFO][6892] cni-plugin/k8s.go 386: Populated endpoint ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"66fa8991-5cd2-45c8-9c64-43cbd360d9ba", ResourceVersion:"654", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"", Pod:"coredns-6f6b679f8f-tphmp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.106.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali29429174561", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.348401 containerd[2700]: 2025-04-30 12:43:35.340 [INFO][6892] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.68/32] ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.348401 containerd[2700]: 2025-04-30 12:43:35.340 [INFO][6892] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali29429174561 ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.348401 containerd[2700]: 2025-04-30 12:43:35.341 [INFO][6892] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.348401 containerd[2700]: 2025-04-30 12:43:35.341 [INFO][6892] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0", GenerateName:"coredns-6f6b679f8f-", Namespace:"kube-system", SelfLink:"", UID:"66fa8991-5cd2-45c8-9c64-43cbd360d9ba", ResourceVersion:"654", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 14, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"6f6b679f8f", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3", Pod:"coredns-6f6b679f8f-tphmp", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.106.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali29429174561", MAC:"aa:71:70:fe:8f:d7", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.348401 containerd[2700]: 2025-04-30 12:43:35.347 [INFO][6892] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3" Namespace="kube-system" Pod="coredns-6f6b679f8f-tphmp" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-coredns--6f6b679f8f--tphmp-eth0" Apr 30 12:43:35.362506 containerd[2700]: time="2025-04-30T12:43:35.362449132Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:35.362827 containerd[2700]: time="2025-04-30T12:43:35.362807259Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:35.362850 containerd[2700]: time="2025-04-30T12:43:35.362822379Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.362911 containerd[2700]: time="2025-04-30T12:43:35.362895100Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.363767 containerd[2700]: time="2025-04-30T12:43:35.363735556Z" level=info msg="StartContainer for \"457899f27d850cd062008b89a501a54f97b5a7eeadedc72d6417de44e01f11ee\" returns successfully" Apr 30 12:43:35.392904 systemd[1]: Started cri-containerd-0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3.scope - libcontainer container 0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3. Apr 30 12:43:35.401757 systemd[1]: run-netns-cni\x2d7b1d36fa\x2dbd72\x2d0560\x2dd684\x2d8ccd4cea347c.mount: Deactivated successfully. Apr 30 12:43:35.420970 containerd[2700]: time="2025-04-30T12:43:35.420940208Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-6f6b679f8f-tphmp,Uid:66fa8991-5cd2-45c8-9c64-43cbd360d9ba,Namespace:kube-system,Attempt:4,} returns sandbox id \"0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3\"" Apr 30 12:43:35.422825 containerd[2700]: time="2025-04-30T12:43:35.422801322Z" level=info msg="CreateContainer within sandbox \"0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 30 12:43:35.430075 containerd[2700]: time="2025-04-30T12:43:35.430044535Z" level=info msg="CreateContainer within sandbox \"0075b43e9c39a8d0a33a954b77a37f89e4343cfbe1ab0a9e582e36bf649c30d3\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b543ffbfa73f00db756376771810a57a6ecd35c20475849a52725301c956985b\"" Apr 30 12:43:35.430424 containerd[2700]: time="2025-04-30T12:43:35.430403542Z" level=info msg="StartContainer for \"b543ffbfa73f00db756376771810a57a6ecd35c20475849a52725301c956985b\"" Apr 30 12:43:35.441013 systemd-networkd[2606]: calif809c448b78: Link UP Apr 30 12:43:35.441160 systemd-networkd[2606]: calif809c448b78: Gained carrier Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:34.963 [INFO][6895] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:34.973 [INFO][6895] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0 csi-node-driver- calico-system 45ca62a1-734b-4197-a2e5-9df6c9a9ad92 588 0 2025-04-30 12:43:25 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:5bcd8f69 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4230.1.1-a-0ae8677ef9 csi-node-driver-nckqz eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] calif809c448b78 [] []}} ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:34.974 [INFO][6895] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.008 [INFO][7009] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" HandleID="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.018 [INFO][7009] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" HandleID="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400038a470), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4230.1.1-a-0ae8677ef9", "pod":"csi-node-driver-nckqz", "timestamp":"2025-04-30 12:43:35.008167415 +0000 UTC"}, Hostname:"ci-4230.1.1-a-0ae8677ef9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.018 [INFO][7009] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.338 [INFO][7009] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.338 [INFO][7009] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230.1.1-a-0ae8677ef9' Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.420 [INFO][7009] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.424 [INFO][7009] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.428 [INFO][7009] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.429 [INFO][7009] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.431 [INFO][7009] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.431 [INFO][7009] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.432 [INFO][7009] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402 Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.434 [INFO][7009] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.438 [INFO][7009] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.69/26] block=192.168.106.64/26 handle="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.438 [INFO][7009] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.69/26] handle="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.438 [INFO][7009] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 12:43:35.448270 containerd[2700]: 2025-04-30 12:43:35.438 [INFO][7009] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.69/26] IPv6=[] ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" HandleID="k8s-pod-network.06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.448819 containerd[2700]: 2025-04-30 12:43:35.439 [INFO][6895] cni-plugin/k8s.go 386: Populated endpoint ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"45ca62a1-734b-4197-a2e5-9df6c9a9ad92", ResourceVersion:"588", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 25, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5bcd8f69", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"", Pod:"csi-node-driver-nckqz", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.106.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif809c448b78", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.448819 containerd[2700]: 2025-04-30 12:43:35.439 [INFO][6895] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.69/32] ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.448819 containerd[2700]: 2025-04-30 12:43:35.439 [INFO][6895] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calif809c448b78 ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.448819 containerd[2700]: 2025-04-30 12:43:35.441 [INFO][6895] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.448819 containerd[2700]: 2025-04-30 12:43:35.441 [INFO][6895] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"45ca62a1-734b-4197-a2e5-9df6c9a9ad92", ResourceVersion:"588", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 25, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"5bcd8f69", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402", Pod:"csi-node-driver-nckqz", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.106.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"calif809c448b78", MAC:"1e:31:fd:3a:d6:4a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.448819 containerd[2700]: 2025-04-30 12:43:35.446 [INFO][6895] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402" Namespace="calico-system" Pod="csi-node-driver-nckqz" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-csi--node--driver--nckqz-eth0" Apr 30 12:43:35.464918 systemd[1]: Started cri-containerd-b543ffbfa73f00db756376771810a57a6ecd35c20475849a52725301c956985b.scope - libcontainer container b543ffbfa73f00db756376771810a57a6ecd35c20475849a52725301c956985b. Apr 30 12:43:35.467291 containerd[2700]: time="2025-04-30T12:43:35.467227979Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:35.467291 containerd[2700]: time="2025-04-30T12:43:35.467280420Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:35.467291 containerd[2700]: time="2025-04-30T12:43:35.467290741Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.467397 containerd[2700]: time="2025-04-30T12:43:35.467361622Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.476480 systemd[1]: Started cri-containerd-06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402.scope - libcontainer container 06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402. Apr 30 12:43:35.484032 containerd[2700]: time="2025-04-30T12:43:35.484000168Z" level=info msg="StartContainer for \"b543ffbfa73f00db756376771810a57a6ecd35c20475849a52725301c956985b\" returns successfully" Apr 30 12:43:35.493647 containerd[2700]: time="2025-04-30T12:43:35.493620465Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-nckqz,Uid:45ca62a1-734b-4197-a2e5-9df6c9a9ad92,Namespace:calico-system,Attempt:3,} returns sandbox id \"06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402\"" Apr 30 12:43:35.555862 systemd-networkd[2606]: cali0447bfb3f6c: Link UP Apr 30 12:43:35.556038 systemd-networkd[2606]: cali0447bfb3f6c: Gained carrier Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:34.965 [INFO][6923] cni-plugin/utils.go 100: File /var/lib/calico/mtu does not exist Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:34.978 [INFO][6923] cni-plugin/plugin.go 340: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0 calico-apiserver-7d8d665688- calico-apiserver 4c75db99-dcfe-4abb-89af-789dc5ea3081 655 0 2025-04-30 12:43:24 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7d8d665688 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4230.1.1-a-0ae8677ef9 calico-apiserver-7d8d665688-fjfzq eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali0447bfb3f6c [] []}} ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:34.978 [INFO][6923] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.008 [INFO][7029] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" HandleID="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.020 [INFO][7029] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" HandleID="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40001b3bb0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4230.1.1-a-0ae8677ef9", "pod":"calico-apiserver-7d8d665688-fjfzq", "timestamp":"2025-04-30 12:43:35.008168735 +0000 UTC"}, Hostname:"ci-4230.1.1-a-0ae8677ef9", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.020 [INFO][7029] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.438 [INFO][7029] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.438 [INFO][7029] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4230.1.1-a-0ae8677ef9' Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.521 [INFO][7029] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.525 [INFO][7029] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.528 [INFO][7029] ipam/ipam.go 489: Trying affinity for 192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.530 [INFO][7029] ipam/ipam.go 155: Attempting to load block cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.532 [INFO][7029] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.106.64/26 host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.532 [INFO][7029] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.106.64/26 handle="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.533 [INFO][7029] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3 Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.536 [INFO][7029] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.106.64/26 handle="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.552 [INFO][7029] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.106.70/26] block=192.168.106.64/26 handle="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.553 [INFO][7029] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.106.70/26] handle="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" host="ci-4230.1.1-a-0ae8677ef9" Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.553 [INFO][7029] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Apr 30 12:43:35.562638 containerd[2700]: 2025-04-30 12:43:35.553 [INFO][7029] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.106.70/26] IPv6=[] ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" HandleID="k8s-pod-network.5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Workload="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.563107 containerd[2700]: 2025-04-30 12:43:35.554 [INFO][6923] cni-plugin/k8s.go 386: Populated endpoint ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0", GenerateName:"calico-apiserver-7d8d665688-", Namespace:"calico-apiserver", SelfLink:"", UID:"4c75db99-dcfe-4abb-89af-789dc5ea3081", ResourceVersion:"655", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d8d665688", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"", Pod:"calico-apiserver-7d8d665688-fjfzq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.106.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0447bfb3f6c", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.563107 containerd[2700]: 2025-04-30 12:43:35.554 [INFO][6923] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.106.70/32] ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.563107 containerd[2700]: 2025-04-30 12:43:35.554 [INFO][6923] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali0447bfb3f6c ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.563107 containerd[2700]: 2025-04-30 12:43:35.556 [INFO][6923] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.563107 containerd[2700]: 2025-04-30 12:43:35.556 [INFO][6923] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0", GenerateName:"calico-apiserver-7d8d665688-", Namespace:"calico-apiserver", SelfLink:"", UID:"4c75db99-dcfe-4abb-89af-789dc5ea3081", ResourceVersion:"655", Generation:0, CreationTimestamp:time.Date(2025, time.April, 30, 12, 43, 24, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7d8d665688", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4230.1.1-a-0ae8677ef9", ContainerID:"5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3", Pod:"calico-apiserver-7d8d665688-fjfzq", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.106.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali0447bfb3f6c", MAC:"b6:61:fa:eb:27:16", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Apr 30 12:43:35.563107 containerd[2700]: 2025-04-30 12:43:35.561 [INFO][6923] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3" Namespace="calico-apiserver" Pod="calico-apiserver-7d8d665688-fjfzq" WorkloadEndpoint="ci--4230.1.1--a--0ae8677ef9-k8s-calico--apiserver--7d8d665688--fjfzq-eth0" Apr 30 12:43:35.577907 containerd[2700]: time="2025-04-30T12:43:35.577272884Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 30 12:43:35.577907 containerd[2700]: time="2025-04-30T12:43:35.577327885Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 30 12:43:35.577907 containerd[2700]: time="2025-04-30T12:43:35.577338925Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.577907 containerd[2700]: time="2025-04-30T12:43:35.577408046Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 30 12:43:35.606868 systemd[1]: Started cri-containerd-5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3.scope - libcontainer container 5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3. Apr 30 12:43:35.630621 containerd[2700]: time="2025-04-30T12:43:35.630591544Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7d8d665688-fjfzq,Uid:4c75db99-dcfe-4abb-89af-789dc5ea3081,Namespace:calico-apiserver,Attempt:4,} returns sandbox id \"5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3\"" Apr 30 12:43:35.895337 containerd[2700]: time="2025-04-30T12:43:35.895296654Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:35.895481 containerd[2700]: time="2025-04-30T12:43:35.895320334Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=40247603" Apr 30 12:43:35.896011 containerd[2700]: time="2025-04-30T12:43:35.895992867Z" level=info msg="ImageCreate event name:\"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:35.897872 containerd[2700]: time="2025-04-30T12:43:35.897845061Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:35.898563 containerd[2700]: time="2025-04-30T12:43:35.898546474Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"41616801\" in 780.808563ms" Apr 30 12:43:35.898584 containerd[2700]: time="2025-04-30T12:43:35.898569474Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\"" Apr 30 12:43:35.899424 containerd[2700]: time="2025-04-30T12:43:35.899405929Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\"" Apr 30 12:43:35.900165 containerd[2700]: time="2025-04-30T12:43:35.900139863Z" level=info msg="CreateContainer within sandbox \"5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 30 12:43:35.904533 containerd[2700]: time="2025-04-30T12:43:35.904501183Z" level=info msg="CreateContainer within sandbox \"5c173f240c6f0d2f29e3a4a05a30f40dce43b7d22644c113f63e417a39065f99\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"5c15b2827609d5bb430b3c04a10607cebdafd066fd9232a4935eb8a32db2d897\"" Apr 30 12:43:35.904963 containerd[2700]: time="2025-04-30T12:43:35.904936631Z" level=info msg="StartContainer for \"5c15b2827609d5bb430b3c04a10607cebdafd066fd9232a4935eb8a32db2d897\"" Apr 30 12:43:35.935922 systemd[1]: Started cri-containerd-5c15b2827609d5bb430b3c04a10607cebdafd066fd9232a4935eb8a32db2d897.scope - libcontainer container 5c15b2827609d5bb430b3c04a10607cebdafd066fd9232a4935eb8a32db2d897. Apr 30 12:43:35.951247 kubelet[4238]: I0430 12:43:35.951219 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:35.961024 containerd[2700]: time="2025-04-30T12:43:35.960996102Z" level=info msg="StartContainer for \"5c15b2827609d5bb430b3c04a10607cebdafd066fd9232a4935eb8a32db2d897\" returns successfully" Apr 30 12:43:35.969694 kubelet[4238]: I0430 12:43:35.969654 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-tphmp" podStartSLOduration=21.969638461 podStartE2EDuration="21.969638461s" podCreationTimestamp="2025-04-30 12:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 12:43:35.961924119 +0000 UTC m=+27.161283903" watchObservedRunningTime="2025-04-30 12:43:35.969638461 +0000 UTC m=+27.168998245" Apr 30 12:43:35.976668 kubelet[4238]: I0430 12:43:35.976629 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-6f6b679f8f-j28hw" podStartSLOduration=21.97661615 podStartE2EDuration="21.97661615s" podCreationTimestamp="2025-04-30 12:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-04-30 12:43:35.969800504 +0000 UTC m=+27.169160288" watchObservedRunningTime="2025-04-30 12:43:35.97661615 +0000 UTC m=+27.175975934" Apr 30 12:43:36.491871 systemd-networkd[2606]: cali29429174561: Gained IPv6LL Apr 30 12:43:36.555825 systemd-networkd[2606]: cali2d3a4f8ce4b: Gained IPv6LL Apr 30 12:43:36.683787 systemd-networkd[2606]: cali3a3d9b22f7c: Gained IPv6LL Apr 30 12:43:36.807223 containerd[2700]: time="2025-04-30T12:43:36.807180901Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:36.807361 containerd[2700]: time="2025-04-30T12:43:36.807229542Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.3: active requests=0, bytes read=32554116" Apr 30 12:43:36.807965 containerd[2700]: time="2025-04-30T12:43:36.807944354Z" level=info msg="ImageCreate event name:\"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:36.809817 containerd[2700]: time="2025-04-30T12:43:36.809780866Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:36.810555 containerd[2700]: time="2025-04-30T12:43:36.810528479Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" with image id \"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:feaab0197035d474845e0f8137a99a78cab274f0a3cac4d5485cf9b1bdf9ffa9\", size \"33923266\" in 911.092589ms" Apr 30 12:43:36.810581 containerd[2700]: time="2025-04-30T12:43:36.810559119Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.3\" returns image reference \"sha256:ec7c64189a2fd01b24b044fea1840d441e9884a0df32c2e9d6982cfbbea1f814\"" Apr 30 12:43:36.811326 containerd[2700]: time="2025-04-30T12:43:36.811307092Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\"" Apr 30 12:43:36.815987 containerd[2700]: time="2025-04-30T12:43:36.815948332Z" level=info msg="CreateContainer within sandbox \"e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Apr 30 12:43:36.833174 containerd[2700]: time="2025-04-30T12:43:36.833142109Z" level=info msg="CreateContainer within sandbox \"e08f0bd62a4a4cfcefe80fa4ca2b5fc97b38aa7af8fc21a9cfe615e5ebea24e8\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"5cc35c5e4168462a83b07df314ff3b6ceca755e445a60eb427502488ba7d8303\"" Apr 30 12:43:36.833571 containerd[2700]: time="2025-04-30T12:43:36.833548156Z" level=info msg="StartContainer for \"5cc35c5e4168462a83b07df314ff3b6ceca755e445a60eb427502488ba7d8303\"" Apr 30 12:43:36.863852 systemd[1]: Started cri-containerd-5cc35c5e4168462a83b07df314ff3b6ceca755e445a60eb427502488ba7d8303.scope - libcontainer container 5cc35c5e4168462a83b07df314ff3b6ceca755e445a60eb427502488ba7d8303. Apr 30 12:43:36.890870 containerd[2700]: time="2025-04-30T12:43:36.890839544Z" level=info msg="StartContainer for \"5cc35c5e4168462a83b07df314ff3b6ceca755e445a60eb427502488ba7d8303\" returns successfully" Apr 30 12:43:36.962152 kubelet[4238]: I0430 12:43:36.962105 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-677bfd4f5b-6wq57" podStartSLOduration=10.376235612 podStartE2EDuration="11.962089253s" podCreationTimestamp="2025-04-30 12:43:25 +0000 UTC" firstStartedPulling="2025-04-30 12:43:35.225312249 +0000 UTC m=+26.424672033" lastFinishedPulling="2025-04-30 12:43:36.81116589 +0000 UTC m=+28.010525674" observedRunningTime="2025-04-30 12:43:36.961679765 +0000 UTC m=+28.161039549" watchObservedRunningTime="2025-04-30 12:43:36.962089253 +0000 UTC m=+28.161449037" Apr 30 12:43:36.968915 kubelet[4238]: I0430 12:43:36.968882 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7d8d665688-9lz6f" podStartSLOduration=12.187148109 podStartE2EDuration="12.968869129s" podCreationTimestamp="2025-04-30 12:43:24 +0000 UTC" firstStartedPulling="2025-04-30 12:43:35.117527467 +0000 UTC m=+26.316887211" lastFinishedPulling="2025-04-30 12:43:35.899248407 +0000 UTC m=+27.098608231" observedRunningTime="2025-04-30 12:43:36.968377481 +0000 UTC m=+28.167737305" watchObservedRunningTime="2025-04-30 12:43:36.968869129 +0000 UTC m=+28.168228913" Apr 30 12:43:37.068123 systemd-networkd[2606]: cali3793b446d51: Gained IPv6LL Apr 30 12:43:37.196096 containerd[2700]: time="2025-04-30T12:43:37.196049677Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.3: active requests=0, bytes read=7474935" Apr 30 12:43:37.196305 containerd[2700]: time="2025-04-30T12:43:37.196140278Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.197994 containerd[2700]: time="2025-04-30T12:43:37.197970708Z" level=info msg="ImageCreate event name:\"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.198689 containerd[2700]: time="2025-04-30T12:43:37.198669799Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.3\" with image id \"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\", size \"8844117\" in 387.335427ms" Apr 30 12:43:37.198714 containerd[2700]: time="2025-04-30T12:43:37.198696600Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.3\" returns image reference \"sha256:15faf29e8b518d846c91c15785ff89e783d356ea0f2b22826f47a556ea32645b\"" Apr 30 12:43:37.199134 containerd[2700]: time="2025-04-30T12:43:37.199119166Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:72455a36febc7c56ec8881007f4805caed5764026a0694e4f86a2503209b2d31\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.199408 containerd[2700]: time="2025-04-30T12:43:37.199393251Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\"" Apr 30 12:43:37.200300 containerd[2700]: time="2025-04-30T12:43:37.200279385Z" level=info msg="CreateContainer within sandbox \"06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Apr 30 12:43:37.222921 containerd[2700]: time="2025-04-30T12:43:37.222888191Z" level=info msg="CreateContainer within sandbox \"06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"069e1029dee167b1833be646cd9902491219c411ff75c356996942c56b2a3e32\"" Apr 30 12:43:37.223310 containerd[2700]: time="2025-04-30T12:43:37.223283277Z" level=info msg="StartContainer for \"069e1029dee167b1833be646cd9902491219c411ff75c356996942c56b2a3e32\"" Apr 30 12:43:37.254851 systemd[1]: Started cri-containerd-069e1029dee167b1833be646cd9902491219c411ff75c356996942c56b2a3e32.scope - libcontainer container 069e1029dee167b1833be646cd9902491219c411ff75c356996942c56b2a3e32. Apr 30 12:43:37.275886 containerd[2700]: time="2025-04-30T12:43:37.275853607Z" level=info msg="StartContainer for \"069e1029dee167b1833be646cd9902491219c411ff75c356996942c56b2a3e32\" returns successfully" Apr 30 12:43:37.321982 containerd[2700]: time="2025-04-30T12:43:37.321917912Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.322076 containerd[2700]: time="2025-04-30T12:43:37.321923952Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.3: active requests=0, bytes read=77" Apr 30 12:43:37.323855 systemd-networkd[2606]: cali0447bfb3f6c: Gained IPv6LL Apr 30 12:43:37.324559 containerd[2700]: time="2025-04-30T12:43:37.324529194Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" with image id \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:bcb659f25f9aebaa389ed1dbb65edb39478ddf82c57d07d8da474e8cab38d77b\", size \"41616801\" in 125.110103ms" Apr 30 12:43:37.324590 containerd[2700]: time="2025-04-30T12:43:37.324562955Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.3\" returns image reference \"sha256:eca64fb9fcc40e83ed2310ac1fab340ba460a939c54e10dc0b7428f02b9b6253\"" Apr 30 12:43:37.325244 containerd[2700]: time="2025-04-30T12:43:37.325223885Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\"" Apr 30 12:43:37.326158 containerd[2700]: time="2025-04-30T12:43:37.326135500Z" level=info msg="CreateContainer within sandbox \"5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Apr 30 12:43:37.330823 containerd[2700]: time="2025-04-30T12:43:37.330796615Z" level=info msg="CreateContainer within sandbox \"5246d9357a0e801fac1439836f7f966d8b1e6d7bdab055cf12ccdee2580433d3\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"5c2071549014c3d5d78cb249f842834fae60b3b2f6ef30b8f91759efbb429e1d\"" Apr 30 12:43:37.331158 containerd[2700]: time="2025-04-30T12:43:37.331132781Z" level=info msg="StartContainer for \"5c2071549014c3d5d78cb249f842834fae60b3b2f6ef30b8f91759efbb429e1d\"" Apr 30 12:43:37.358847 systemd[1]: Started cri-containerd-5c2071549014c3d5d78cb249f842834fae60b3b2f6ef30b8f91759efbb429e1d.scope - libcontainer container 5c2071549014c3d5d78cb249f842834fae60b3b2f6ef30b8f91759efbb429e1d. Apr 30 12:43:37.384984 containerd[2700]: time="2025-04-30T12:43:37.384950251Z" level=info msg="StartContainer for \"5c2071549014c3d5d78cb249f842834fae60b3b2f6ef30b8f91759efbb429e1d\" returns successfully" Apr 30 12:43:37.385290 kubelet[4238]: I0430 12:43:37.385268 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:37.451856 systemd-networkd[2606]: calif809c448b78: Gained IPv6LL Apr 30 12:43:37.785964 containerd[2700]: time="2025-04-30T12:43:37.785929894Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.786079 containerd[2700]: time="2025-04-30T12:43:37.785967094Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3: active requests=0, bytes read=13124299" Apr 30 12:43:37.786708 containerd[2700]: time="2025-04-30T12:43:37.786686186Z" level=info msg="ImageCreate event name:\"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.788429 containerd[2700]: time="2025-04-30T12:43:37.788403614Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Apr 30 12:43:37.789118 containerd[2700]: time="2025-04-30T12:43:37.789100505Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" with image id \"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:3f15090a9bb45773d1fd019455ec3d3f3746f3287c35d8013e497b38d8237324\", size \"14493433\" in 463.849179ms" Apr 30 12:43:37.789142 containerd[2700]: time="2025-04-30T12:43:37.789122145Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.3\" returns image reference \"sha256:a91b1f00752edc175f270a01b33683fa80818734aa2274388785eaf3364315dc\"" Apr 30 12:43:37.790808 containerd[2700]: time="2025-04-30T12:43:37.790783932Z" level=info msg="CreateContainer within sandbox \"06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Apr 30 12:43:37.796945 containerd[2700]: time="2025-04-30T12:43:37.796916951Z" level=info msg="CreateContainer within sandbox \"06e493daf1121d999862b4fcbf411ec2bc1a0b9d874ed8cce714c5aa19050402\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"cc97aba6d4dbe57f3282e4e9460bea040282c341862b32f61f10b1c04baaae1b\"" Apr 30 12:43:37.797252 containerd[2700]: time="2025-04-30T12:43:37.797233597Z" level=info msg="StartContainer for \"cc97aba6d4dbe57f3282e4e9460bea040282c341862b32f61f10b1c04baaae1b\"" Apr 30 12:43:37.822913 systemd[1]: Started cri-containerd-cc97aba6d4dbe57f3282e4e9460bea040282c341862b32f61f10b1c04baaae1b.scope - libcontainer container cc97aba6d4dbe57f3282e4e9460bea040282c341862b32f61f10b1c04baaae1b. Apr 30 12:43:37.843910 containerd[2700]: time="2025-04-30T12:43:37.843878551Z" level=info msg="StartContainer for \"cc97aba6d4dbe57f3282e4e9460bea040282c341862b32f61f10b1c04baaae1b\" returns successfully" Apr 30 12:43:37.916235 kubelet[4238]: I0430 12:43:37.916213 4238 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Apr 30 12:43:37.916298 kubelet[4238]: I0430 12:43:37.916249 4238 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Apr 30 12:43:37.961370 kubelet[4238]: I0430 12:43:37.961333 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:37.961738 kubelet[4238]: I0430 12:43:37.961375 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:37.966849 kubelet[4238]: I0430 12:43:37.966809 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7d8d665688-fjfzq" podStartSLOduration=12.273084174 podStartE2EDuration="13.966798218s" podCreationTimestamp="2025-04-30 12:43:24 +0000 UTC" firstStartedPulling="2025-04-30 12:43:35.631393159 +0000 UTC m=+26.830752943" lastFinishedPulling="2025-04-30 12:43:37.325107203 +0000 UTC m=+28.524466987" observedRunningTime="2025-04-30 12:43:37.966502973 +0000 UTC m=+29.165862757" watchObservedRunningTime="2025-04-30 12:43:37.966798218 +0000 UTC m=+29.166158002" Apr 30 12:43:37.974025 kubelet[4238]: I0430 12:43:37.973982 4238 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-nckqz" podStartSLOduration=10.678813181 podStartE2EDuration="12.973969614s" podCreationTimestamp="2025-04-30 12:43:25 +0000 UTC" firstStartedPulling="2025-04-30 12:43:35.494546282 +0000 UTC m=+26.693906066" lastFinishedPulling="2025-04-30 12:43:37.789702715 +0000 UTC m=+28.989062499" observedRunningTime="2025-04-30 12:43:37.973763611 +0000 UTC m=+29.173123395" watchObservedRunningTime="2025-04-30 12:43:37.973969614 +0000 UTC m=+29.173329398" Apr 30 12:43:38.963157 kubelet[4238]: I0430 12:43:38.963124 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:45.162851 kubelet[4238]: I0430 12:43:45.162769 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:52.707619 kubelet[4238]: I0430 12:43:52.707575 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:43:53.019761 kernel: bpftool[8785]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Apr 30 12:43:53.177107 systemd-networkd[2606]: vxlan.calico: Link UP Apr 30 12:43:53.177112 systemd-networkd[2606]: vxlan.calico: Gained carrier Apr 30 12:43:55.051896 systemd-networkd[2606]: vxlan.calico: Gained IPv6LL Apr 30 12:43:55.142627 kubelet[4238]: I0430 12:43:55.142592 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:44:08.678186 kubelet[4238]: I0430 12:44:08.678088 4238 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Apr 30 12:44:08.862066 containerd[2700]: time="2025-04-30T12:44:08.862020676Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:44:08.862360 containerd[2700]: time="2025-04-30T12:44:08.862129796Z" level=info msg="TearDown network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" successfully" Apr 30 12:44:08.862360 containerd[2700]: time="2025-04-30T12:44:08.862141676Z" level=info msg="StopPodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" returns successfully" Apr 30 12:44:08.862451 containerd[2700]: time="2025-04-30T12:44:08.862427356Z" level=info msg="RemovePodSandbox for \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:44:08.862475 containerd[2700]: time="2025-04-30T12:44:08.862459757Z" level=info msg="Forcibly stopping sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\"" Apr 30 12:44:08.862536 containerd[2700]: time="2025-04-30T12:44:08.862526877Z" level=info msg="TearDown network for sandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" successfully" Apr 30 12:44:08.865365 containerd[2700]: time="2025-04-30T12:44:08.865292723Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.865365 containerd[2700]: time="2025-04-30T12:44:08.865370003Z" level=info msg="RemovePodSandbox \"5e9698b3fa59a91b602920cbd4623e6b8ee81af9b8a0e05a30ef00af67e35064\" returns successfully" Apr 30 12:44:08.865696 containerd[2700]: time="2025-04-30T12:44:08.865675324Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" Apr 30 12:44:08.865780 containerd[2700]: time="2025-04-30T12:44:08.865766324Z" level=info msg="TearDown network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" successfully" Apr 30 12:44:08.865780 containerd[2700]: time="2025-04-30T12:44:08.865778724Z" level=info msg="StopPodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" returns successfully" Apr 30 12:44:08.866068 containerd[2700]: time="2025-04-30T12:44:08.866040244Z" level=info msg="RemovePodSandbox for \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" Apr 30 12:44:08.866113 containerd[2700]: time="2025-04-30T12:44:08.866069684Z" level=info msg="Forcibly stopping sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\"" Apr 30 12:44:08.866155 containerd[2700]: time="2025-04-30T12:44:08.866139565Z" level=info msg="TearDown network for sandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" successfully" Apr 30 12:44:08.867527 containerd[2700]: time="2025-04-30T12:44:08.867501688Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.867583 containerd[2700]: time="2025-04-30T12:44:08.867548048Z" level=info msg="RemovePodSandbox \"d6b60db7492bc3a270441f51e5d52fbfbdf0d6a98752a31d1f83f80222983620\" returns successfully" Apr 30 12:44:08.867778 containerd[2700]: time="2025-04-30T12:44:08.867761128Z" level=info msg="StopPodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\"" Apr 30 12:44:08.867853 containerd[2700]: time="2025-04-30T12:44:08.867841368Z" level=info msg="TearDown network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" successfully" Apr 30 12:44:08.867853 containerd[2700]: time="2025-04-30T12:44:08.867851848Z" level=info msg="StopPodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" returns successfully" Apr 30 12:44:08.868112 containerd[2700]: time="2025-04-30T12:44:08.868093449Z" level=info msg="RemovePodSandbox for \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\"" Apr 30 12:44:08.868156 containerd[2700]: time="2025-04-30T12:44:08.868119929Z" level=info msg="Forcibly stopping sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\"" Apr 30 12:44:08.868231 containerd[2700]: time="2025-04-30T12:44:08.868189289Z" level=info msg="TearDown network for sandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" successfully" Apr 30 12:44:08.869507 containerd[2700]: time="2025-04-30T12:44:08.869483772Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.869576 containerd[2700]: time="2025-04-30T12:44:08.869526972Z" level=info msg="RemovePodSandbox \"aa15d426e04b27015ecefdfb934002cf7c0836a6daec5d9d5e7e5a7378cda3cd\" returns successfully" Apr 30 12:44:08.869793 containerd[2700]: time="2025-04-30T12:44:08.869773093Z" level=info msg="StopPodSandbox for \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\"" Apr 30 12:44:08.870482 containerd[2700]: time="2025-04-30T12:44:08.869852453Z" level=info msg="TearDown network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\" successfully" Apr 30 12:44:08.870482 containerd[2700]: time="2025-04-30T12:44:08.869862853Z" level=info msg="StopPodSandbox for \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\" returns successfully" Apr 30 12:44:08.870482 containerd[2700]: time="2025-04-30T12:44:08.870103973Z" level=info msg="RemovePodSandbox for \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\"" Apr 30 12:44:08.870482 containerd[2700]: time="2025-04-30T12:44:08.870124573Z" level=info msg="Forcibly stopping sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\"" Apr 30 12:44:08.870482 containerd[2700]: time="2025-04-30T12:44:08.870200373Z" level=info msg="TearDown network for sandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\" successfully" Apr 30 12:44:08.876838 containerd[2700]: time="2025-04-30T12:44:08.876805388Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.876944 containerd[2700]: time="2025-04-30T12:44:08.876864988Z" level=info msg="RemovePodSandbox \"ef5fdc69e20449d1885c940ca11b5f6d519cefa00ce9c4a12059f5321f4bd1ba\" returns successfully" Apr 30 12:44:08.882755 containerd[2700]: time="2025-04-30T12:44:08.879854755Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:44:08.882755 containerd[2700]: time="2025-04-30T12:44:08.879937955Z" level=info msg="TearDown network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" successfully" Apr 30 12:44:08.882755 containerd[2700]: time="2025-04-30T12:44:08.879947035Z" level=info msg="StopPodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" returns successfully" Apr 30 12:44:08.883949 containerd[2700]: time="2025-04-30T12:44:08.883915203Z" level=info msg="RemovePodSandbox for \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:44:08.884033 containerd[2700]: time="2025-04-30T12:44:08.883956204Z" level=info msg="Forcibly stopping sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\"" Apr 30 12:44:08.884064 containerd[2700]: time="2025-04-30T12:44:08.884036484Z" level=info msg="TearDown network for sandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" successfully" Apr 30 12:44:08.890292 containerd[2700]: time="2025-04-30T12:44:08.890254897Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.890359 containerd[2700]: time="2025-04-30T12:44:08.890317537Z" level=info msg="RemovePodSandbox \"ac897820649c203f10cff6146702eb11e7a761b8ede0b8899dcd95a05d363b50\" returns successfully" Apr 30 12:44:08.890703 containerd[2700]: time="2025-04-30T12:44:08.890677818Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" Apr 30 12:44:08.890790 containerd[2700]: time="2025-04-30T12:44:08.890776738Z" level=info msg="TearDown network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" successfully" Apr 30 12:44:08.890790 containerd[2700]: time="2025-04-30T12:44:08.890787058Z" level=info msg="StopPodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" returns successfully" Apr 30 12:44:08.891074 containerd[2700]: time="2025-04-30T12:44:08.891051979Z" level=info msg="RemovePodSandbox for \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" Apr 30 12:44:08.891097 containerd[2700]: time="2025-04-30T12:44:08.891078059Z" level=info msg="Forcibly stopping sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\"" Apr 30 12:44:08.891154 containerd[2700]: time="2025-04-30T12:44:08.891142699Z" level=info msg="TearDown network for sandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" successfully" Apr 30 12:44:08.892586 containerd[2700]: time="2025-04-30T12:44:08.892562822Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.892618 containerd[2700]: time="2025-04-30T12:44:08.892607542Z" level=info msg="RemovePodSandbox \"1b01bf115478d252d40bc88e523234d6d55cfe0393a2be45af381da6a271a2f4\" returns successfully" Apr 30 12:44:08.892831 containerd[2700]: time="2025-04-30T12:44:08.892813743Z" level=info msg="StopPodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\"" Apr 30 12:44:08.892905 containerd[2700]: time="2025-04-30T12:44:08.892894143Z" level=info msg="TearDown network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" successfully" Apr 30 12:44:08.892930 containerd[2700]: time="2025-04-30T12:44:08.892905183Z" level=info msg="StopPodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" returns successfully" Apr 30 12:44:08.893136 containerd[2700]: time="2025-04-30T12:44:08.893119984Z" level=info msg="RemovePodSandbox for \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\"" Apr 30 12:44:08.893160 containerd[2700]: time="2025-04-30T12:44:08.893142024Z" level=info msg="Forcibly stopping sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\"" Apr 30 12:44:08.893204 containerd[2700]: time="2025-04-30T12:44:08.893194464Z" level=info msg="TearDown network for sandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" successfully" Apr 30 12:44:08.894507 containerd[2700]: time="2025-04-30T12:44:08.894484907Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.894544 containerd[2700]: time="2025-04-30T12:44:08.894532827Z" level=info msg="RemovePodSandbox \"411021060c7673dca2205de69173fed427092b9424c179b0d3ba101de47ae2c8\" returns successfully" Apr 30 12:44:08.894756 containerd[2700]: time="2025-04-30T12:44:08.894735747Z" level=info msg="StopPodSandbox for \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\"" Apr 30 12:44:08.894831 containerd[2700]: time="2025-04-30T12:44:08.894819787Z" level=info msg="TearDown network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\" successfully" Apr 30 12:44:08.894854 containerd[2700]: time="2025-04-30T12:44:08.894831427Z" level=info msg="StopPodSandbox for \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\" returns successfully" Apr 30 12:44:08.895069 containerd[2700]: time="2025-04-30T12:44:08.895050948Z" level=info msg="RemovePodSandbox for \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\"" Apr 30 12:44:08.895089 containerd[2700]: time="2025-04-30T12:44:08.895075028Z" level=info msg="Forcibly stopping sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\"" Apr 30 12:44:08.895152 containerd[2700]: time="2025-04-30T12:44:08.895140428Z" level=info msg="TearDown network for sandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\" successfully" Apr 30 12:44:08.896464 containerd[2700]: time="2025-04-30T12:44:08.896437791Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.896494 containerd[2700]: time="2025-04-30T12:44:08.896483751Z" level=info msg="RemovePodSandbox \"d16bb5123fb74d7f9260fd1ff566e84b3952460448982d111acb73c7766e1be4\" returns successfully" Apr 30 12:44:08.896712 containerd[2700]: time="2025-04-30T12:44:08.896692671Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:44:08.896796 containerd[2700]: time="2025-04-30T12:44:08.896783312Z" level=info msg="TearDown network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" successfully" Apr 30 12:44:08.896796 containerd[2700]: time="2025-04-30T12:44:08.896794632Z" level=info msg="StopPodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" returns successfully" Apr 30 12:44:08.898074 containerd[2700]: time="2025-04-30T12:44:08.897045392Z" level=info msg="RemovePodSandbox for \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:44:08.898074 containerd[2700]: time="2025-04-30T12:44:08.897079432Z" level=info msg="Forcibly stopping sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\"" Apr 30 12:44:08.898074 containerd[2700]: time="2025-04-30T12:44:08.897146712Z" level=info msg="TearDown network for sandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" successfully" Apr 30 12:44:08.899029 containerd[2700]: time="2025-04-30T12:44:08.898969716Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.899029 containerd[2700]: time="2025-04-30T12:44:08.899015116Z" level=info msg="RemovePodSandbox \"c5596f615da0413816e045b696eb353de7ecdfc0c2ce993eb8bcccbd6789517a\" returns successfully" Apr 30 12:44:08.899262 containerd[2700]: time="2025-04-30T12:44:08.899243597Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" Apr 30 12:44:08.899339 containerd[2700]: time="2025-04-30T12:44:08.899327157Z" level=info msg="TearDown network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" successfully" Apr 30 12:44:08.899370 containerd[2700]: time="2025-04-30T12:44:08.899338477Z" level=info msg="StopPodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" returns successfully" Apr 30 12:44:08.899557 containerd[2700]: time="2025-04-30T12:44:08.899540758Z" level=info msg="RemovePodSandbox for \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" Apr 30 12:44:08.899592 containerd[2700]: time="2025-04-30T12:44:08.899561078Z" level=info msg="Forcibly stopping sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\"" Apr 30 12:44:08.899647 containerd[2700]: time="2025-04-30T12:44:08.899636758Z" level=info msg="TearDown network for sandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" successfully" Apr 30 12:44:08.900933 containerd[2700]: time="2025-04-30T12:44:08.900910721Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.900966 containerd[2700]: time="2025-04-30T12:44:08.900951801Z" level=info msg="RemovePodSandbox \"1c1d1209b877c573d9c3b91828f930c32c586086ac8c7d6236a3afceea7e9bc7\" returns successfully" Apr 30 12:44:08.901332 containerd[2700]: time="2025-04-30T12:44:08.901187321Z" level=info msg="StopPodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\"" Apr 30 12:44:08.901332 containerd[2700]: time="2025-04-30T12:44:08.901270841Z" level=info msg="TearDown network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" successfully" Apr 30 12:44:08.901332 containerd[2700]: time="2025-04-30T12:44:08.901280601Z" level=info msg="StopPodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" returns successfully" Apr 30 12:44:08.901502 containerd[2700]: time="2025-04-30T12:44:08.901487282Z" level=info msg="RemovePodSandbox for \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\"" Apr 30 12:44:08.901535 containerd[2700]: time="2025-04-30T12:44:08.901505482Z" level=info msg="Forcibly stopping sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\"" Apr 30 12:44:08.901579 containerd[2700]: time="2025-04-30T12:44:08.901566682Z" level=info msg="TearDown network for sandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" successfully" Apr 30 12:44:08.902843 containerd[2700]: time="2025-04-30T12:44:08.902823525Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.902884 containerd[2700]: time="2025-04-30T12:44:08.902863845Z" level=info msg="RemovePodSandbox \"2bff726927d46985873d3c87f35cce09f410f4627cc44fd40756b20c11566470\" returns successfully" Apr 30 12:44:08.903226 containerd[2700]: time="2025-04-30T12:44:08.903097445Z" level=info msg="StopPodSandbox for \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\"" Apr 30 12:44:08.903226 containerd[2700]: time="2025-04-30T12:44:08.903169766Z" level=info msg="TearDown network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\" successfully" Apr 30 12:44:08.903226 containerd[2700]: time="2025-04-30T12:44:08.903178606Z" level=info msg="StopPodSandbox for \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\" returns successfully" Apr 30 12:44:08.903378 containerd[2700]: time="2025-04-30T12:44:08.903360286Z" level=info msg="RemovePodSandbox for \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\"" Apr 30 12:44:08.903412 containerd[2700]: time="2025-04-30T12:44:08.903382406Z" level=info msg="Forcibly stopping sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\"" Apr 30 12:44:08.903454 containerd[2700]: time="2025-04-30T12:44:08.903443566Z" level=info msg="TearDown network for sandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\" successfully" Apr 30 12:44:08.904703 containerd[2700]: time="2025-04-30T12:44:08.904681529Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.904735 containerd[2700]: time="2025-04-30T12:44:08.904725369Z" level=info msg="RemovePodSandbox \"88570af6586378889350bc2b315b78c78a51267eb70b0f1e92a4addbdfc9f2b9\" returns successfully" Apr 30 12:44:08.905003 containerd[2700]: time="2025-04-30T12:44:08.904985650Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" Apr 30 12:44:08.905084 containerd[2700]: time="2025-04-30T12:44:08.905072650Z" level=info msg="TearDown network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" successfully" Apr 30 12:44:08.905084 containerd[2700]: time="2025-04-30T12:44:08.905082650Z" level=info msg="StopPodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" returns successfully" Apr 30 12:44:08.905406 containerd[2700]: time="2025-04-30T12:44:08.905378050Z" level=info msg="RemovePodSandbox for \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" Apr 30 12:44:08.905431 containerd[2700]: time="2025-04-30T12:44:08.905411170Z" level=info msg="Forcibly stopping sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\"" Apr 30 12:44:08.905496 containerd[2700]: time="2025-04-30T12:44:08.905485091Z" level=info msg="TearDown network for sandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" successfully" Apr 30 12:44:08.906774 containerd[2700]: time="2025-04-30T12:44:08.906750813Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.906808 containerd[2700]: time="2025-04-30T12:44:08.906791973Z" level=info msg="RemovePodSandbox \"e243af3c5851ecb0806cd7cd7b4aa03f8bc299ed0c4e9a68a55817cb52a814cb\" returns successfully" Apr 30 12:44:08.907152 containerd[2700]: time="2025-04-30T12:44:08.907023694Z" level=info msg="StopPodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\"" Apr 30 12:44:08.907152 containerd[2700]: time="2025-04-30T12:44:08.907098894Z" level=info msg="TearDown network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" successfully" Apr 30 12:44:08.907152 containerd[2700]: time="2025-04-30T12:44:08.907107054Z" level=info msg="StopPodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" returns successfully" Apr 30 12:44:08.907293 containerd[2700]: time="2025-04-30T12:44:08.907267695Z" level=info msg="RemovePodSandbox for \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\"" Apr 30 12:44:08.907330 containerd[2700]: time="2025-04-30T12:44:08.907291135Z" level=info msg="Forcibly stopping sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\"" Apr 30 12:44:08.907372 containerd[2700]: time="2025-04-30T12:44:08.907359335Z" level=info msg="TearDown network for sandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" successfully" Apr 30 12:44:08.908724 containerd[2700]: time="2025-04-30T12:44:08.908681858Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.908724 containerd[2700]: time="2025-04-30T12:44:08.908726738Z" level=info msg="RemovePodSandbox \"027fd39abf58571bd2792896a3f677f5a56393fd54fe8d050f54de766e5365b8\" returns successfully" Apr 30 12:44:08.908988 containerd[2700]: time="2025-04-30T12:44:08.908971938Z" level=info msg="StopPodSandbox for \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\"" Apr 30 12:44:08.909057 containerd[2700]: time="2025-04-30T12:44:08.909045498Z" level=info msg="TearDown network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\" successfully" Apr 30 12:44:08.909084 containerd[2700]: time="2025-04-30T12:44:08.909056018Z" level=info msg="StopPodSandbox for \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\" returns successfully" Apr 30 12:44:08.909234 containerd[2700]: time="2025-04-30T12:44:08.909216859Z" level=info msg="RemovePodSandbox for \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\"" Apr 30 12:44:08.909262 containerd[2700]: time="2025-04-30T12:44:08.909239259Z" level=info msg="Forcibly stopping sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\"" Apr 30 12:44:08.909322 containerd[2700]: time="2025-04-30T12:44:08.909310419Z" level=info msg="TearDown network for sandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\" successfully" Apr 30 12:44:08.910575 containerd[2700]: time="2025-04-30T12:44:08.910554102Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.910604 containerd[2700]: time="2025-04-30T12:44:08.910594622Z" level=info msg="RemovePodSandbox \"63587b90a90b04b56f809e4d2015ddc6b42c2688a4eb1572521a9898e548d3e4\" returns successfully" Apr 30 12:44:08.910829 containerd[2700]: time="2025-04-30T12:44:08.910811582Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:44:08.910900 containerd[2700]: time="2025-04-30T12:44:08.910888462Z" level=info msg="TearDown network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" successfully" Apr 30 12:44:08.910934 containerd[2700]: time="2025-04-30T12:44:08.910899222Z" level=info msg="StopPodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" returns successfully" Apr 30 12:44:08.911112 containerd[2700]: time="2025-04-30T12:44:08.911096743Z" level=info msg="RemovePodSandbox for \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:44:08.911147 containerd[2700]: time="2025-04-30T12:44:08.911116343Z" level=info msg="Forcibly stopping sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\"" Apr 30 12:44:08.911185 containerd[2700]: time="2025-04-30T12:44:08.911174023Z" level=info msg="TearDown network for sandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" successfully" Apr 30 12:44:08.912425 containerd[2700]: time="2025-04-30T12:44:08.912402906Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.912483 containerd[2700]: time="2025-04-30T12:44:08.912443386Z" level=info msg="RemovePodSandbox \"7032bddf715fa0251bd059b5f7eca980c9c35ea8674cabed98b59cbd682958ba\" returns successfully" Apr 30 12:44:08.912668 containerd[2700]: time="2025-04-30T12:44:08.912650506Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" Apr 30 12:44:08.912731 containerd[2700]: time="2025-04-30T12:44:08.912718506Z" level=info msg="TearDown network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" successfully" Apr 30 12:44:08.912731 containerd[2700]: time="2025-04-30T12:44:08.912728386Z" level=info msg="StopPodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" returns successfully" Apr 30 12:44:08.912953 containerd[2700]: time="2025-04-30T12:44:08.912934347Z" level=info msg="RemovePodSandbox for \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" Apr 30 12:44:08.912975 containerd[2700]: time="2025-04-30T12:44:08.912956187Z" level=info msg="Forcibly stopping sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\"" Apr 30 12:44:08.913026 containerd[2700]: time="2025-04-30T12:44:08.913013187Z" level=info msg="TearDown network for sandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" successfully" Apr 30 12:44:08.914322 containerd[2700]: time="2025-04-30T12:44:08.914298910Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.914351 containerd[2700]: time="2025-04-30T12:44:08.914343150Z" level=info msg="RemovePodSandbox \"000233665f3acc60f6c24b7e63c7f3e217849b920863b203357d90d63de1b748\" returns successfully" Apr 30 12:44:08.914579 containerd[2700]: time="2025-04-30T12:44:08.914558910Z" level=info msg="StopPodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\"" Apr 30 12:44:08.914676 containerd[2700]: time="2025-04-30T12:44:08.914661191Z" level=info msg="TearDown network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" successfully" Apr 30 12:44:08.914701 containerd[2700]: time="2025-04-30T12:44:08.914673831Z" level=info msg="StopPodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" returns successfully" Apr 30 12:44:08.914898 containerd[2700]: time="2025-04-30T12:44:08.914881871Z" level=info msg="RemovePodSandbox for \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\"" Apr 30 12:44:08.914920 containerd[2700]: time="2025-04-30T12:44:08.914900991Z" level=info msg="Forcibly stopping sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\"" Apr 30 12:44:08.914972 containerd[2700]: time="2025-04-30T12:44:08.914959191Z" level=info msg="TearDown network for sandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" successfully" Apr 30 12:44:08.916234 containerd[2700]: time="2025-04-30T12:44:08.916211274Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.916262 containerd[2700]: time="2025-04-30T12:44:08.916252834Z" level=info msg="RemovePodSandbox \"f7c1d1afee3b69ea17ab4bb42ef928f0ea8197ad6af25ccc21bd83d14acbad3f\" returns successfully" Apr 30 12:44:08.916487 containerd[2700]: time="2025-04-30T12:44:08.916466915Z" level=info msg="StopPodSandbox for \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\"" Apr 30 12:44:08.916558 containerd[2700]: time="2025-04-30T12:44:08.916545315Z" level=info msg="TearDown network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\" successfully" Apr 30 12:44:08.916583 containerd[2700]: time="2025-04-30T12:44:08.916556315Z" level=info msg="StopPodSandbox for \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\" returns successfully" Apr 30 12:44:08.916771 containerd[2700]: time="2025-04-30T12:44:08.916754115Z" level=info msg="RemovePodSandbox for \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\"" Apr 30 12:44:08.916800 containerd[2700]: time="2025-04-30T12:44:08.916775035Z" level=info msg="Forcibly stopping sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\"" Apr 30 12:44:08.916852 containerd[2700]: time="2025-04-30T12:44:08.916839995Z" level=info msg="TearDown network for sandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\" successfully" Apr 30 12:44:08.918111 containerd[2700]: time="2025-04-30T12:44:08.918088318Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.918142 containerd[2700]: time="2025-04-30T12:44:08.918130238Z" level=info msg="RemovePodSandbox \"27c8dd52ab66846771101db63bc74cc61a5fcfe1891fedb5aa2f8199e3cf42e4\" returns successfully" Apr 30 12:44:08.918384 containerd[2700]: time="2025-04-30T12:44:08.918366759Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:44:08.918444 containerd[2700]: time="2025-04-30T12:44:08.918432759Z" level=info msg="TearDown network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" successfully" Apr 30 12:44:08.918444 containerd[2700]: time="2025-04-30T12:44:08.918442359Z" level=info msg="StopPodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" returns successfully" Apr 30 12:44:08.918684 containerd[2700]: time="2025-04-30T12:44:08.918665399Z" level=info msg="RemovePodSandbox for \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:44:08.918709 containerd[2700]: time="2025-04-30T12:44:08.918686559Z" level=info msg="Forcibly stopping sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\"" Apr 30 12:44:08.918768 containerd[2700]: time="2025-04-30T12:44:08.918755480Z" level=info msg="TearDown network for sandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" successfully" Apr 30 12:44:08.920018 containerd[2700]: time="2025-04-30T12:44:08.919990322Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.920042 containerd[2700]: time="2025-04-30T12:44:08.920032122Z" level=info msg="RemovePodSandbox \"a69d87af78aabfa7e098a7232b13506c8e2b3a0ff57bf07e37896e8e154bf229\" returns successfully" Apr 30 12:44:08.920242 containerd[2700]: time="2025-04-30T12:44:08.920226603Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" Apr 30 12:44:08.920303 containerd[2700]: time="2025-04-30T12:44:08.920291683Z" level=info msg="TearDown network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" successfully" Apr 30 12:44:08.920328 containerd[2700]: time="2025-04-30T12:44:08.920304563Z" level=info msg="StopPodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" returns successfully" Apr 30 12:44:08.920540 containerd[2700]: time="2025-04-30T12:44:08.920520884Z" level=info msg="RemovePodSandbox for \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" Apr 30 12:44:08.920565 containerd[2700]: time="2025-04-30T12:44:08.920543484Z" level=info msg="Forcibly stopping sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\"" Apr 30 12:44:08.920619 containerd[2700]: time="2025-04-30T12:44:08.920606044Z" level=info msg="TearDown network for sandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" successfully" Apr 30 12:44:08.921915 containerd[2700]: time="2025-04-30T12:44:08.921890127Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.921950 containerd[2700]: time="2025-04-30T12:44:08.921939527Z" level=info msg="RemovePodSandbox \"c3f64980ab0df63bcd7d6ff70c9de401254594fafa72b2a116edb5f8865b1a45\" returns successfully" Apr 30 12:44:08.922191 containerd[2700]: time="2025-04-30T12:44:08.922174767Z" level=info msg="StopPodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\"" Apr 30 12:44:08.922262 containerd[2700]: time="2025-04-30T12:44:08.922248967Z" level=info msg="TearDown network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" successfully" Apr 30 12:44:08.922262 containerd[2700]: time="2025-04-30T12:44:08.922259567Z" level=info msg="StopPodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" returns successfully" Apr 30 12:44:08.922454 containerd[2700]: time="2025-04-30T12:44:08.922440128Z" level=info msg="RemovePodSandbox for \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\"" Apr 30 12:44:08.922475 containerd[2700]: time="2025-04-30T12:44:08.922459408Z" level=info msg="Forcibly stopping sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\"" Apr 30 12:44:08.922522 containerd[2700]: time="2025-04-30T12:44:08.922512168Z" level=info msg="TearDown network for sandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" successfully" Apr 30 12:44:08.923803 containerd[2700]: time="2025-04-30T12:44:08.923780811Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.923839 containerd[2700]: time="2025-04-30T12:44:08.923827251Z" level=info msg="RemovePodSandbox \"fb14f27d9b30e043575b39ee667d0c78d76576897083263ba1ea63c5548262fe\" returns successfully" Apr 30 12:44:08.924076 containerd[2700]: time="2025-04-30T12:44:08.924062931Z" level=info msg="StopPodSandbox for \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\"" Apr 30 12:44:08.924145 containerd[2700]: time="2025-04-30T12:44:08.924134771Z" level=info msg="TearDown network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\" successfully" Apr 30 12:44:08.924165 containerd[2700]: time="2025-04-30T12:44:08.924144771Z" level=info msg="StopPodSandbox for \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\" returns successfully" Apr 30 12:44:08.924348 containerd[2700]: time="2025-04-30T12:44:08.924333532Z" level=info msg="RemovePodSandbox for \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\"" Apr 30 12:44:08.924369 containerd[2700]: time="2025-04-30T12:44:08.924355692Z" level=info msg="Forcibly stopping sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\"" Apr 30 12:44:08.924427 containerd[2700]: time="2025-04-30T12:44:08.924416852Z" level=info msg="TearDown network for sandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\" successfully" Apr 30 12:44:08.925684 containerd[2700]: time="2025-04-30T12:44:08.925661455Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Apr 30 12:44:08.925716 containerd[2700]: time="2025-04-30T12:44:08.925707055Z" level=info msg="RemovePodSandbox \"c213b8e9969999fea207cb3e9b786b88d40a93e62236de73679a963fdf9d488f\" returns successfully" Apr 30 12:44:53.386653 systemd[1]: sshd@7-147.75.61.11:22-218.92.0.158:25852.service: Deactivated successfully. Apr 30 12:45:16.218824 systemd[1]: Started sshd@9-147.75.61.11:22-218.92.0.158:55106.service - OpenSSH per-connection server daemon (218.92.0.158:55106). Apr 30 12:45:19.209220 sshd-session[9362]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.158 user=root Apr 30 12:45:20.703706 sshd[9360]: PAM: Permission denied for root from 218.92.0.158 Apr 30 12:45:21.154200 sshd-session[9363]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.158 user=root Apr 30 12:45:23.591548 sshd[9360]: PAM: Permission denied for root from 218.92.0.158 Apr 30 12:45:24.041151 sshd-session[9364]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.158 user=root Apr 30 12:45:25.555396 sshd[9360]: PAM: Permission denied for root from 218.92.0.158 Apr 30 12:45:25.777802 sshd[9360]: Received disconnect from 218.92.0.158 port 55106:11: [preauth] Apr 30 12:45:25.777802 sshd[9360]: Disconnected from authenticating user root 218.92.0.158 port 55106 [preauth] Apr 30 12:45:25.779618 systemd[1]: sshd@9-147.75.61.11:22-218.92.0.158:55106.service: Deactivated successfully. Apr 30 12:47:39.281820 systemd[1]: Started sshd@10-147.75.61.11:22-218.92.0.158:26073.service - OpenSSH per-connection server daemon (218.92.0.158:26073). Apr 30 12:47:41.937880 sshd-session[9768]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.158 user=root Apr 30 12:47:44.064448 sshd[9765]: PAM: Permission denied for root from 218.92.0.158 Apr 30 12:47:44.476737 sshd-session[9769]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.158 user=root Apr 30 12:47:46.015896 sshd[9765]: PAM: Permission denied for root from 218.92.0.158 Apr 30 12:47:46.427382 sshd-session[9772]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=218.92.0.158 user=root Apr 30 12:47:48.574483 sshd[9765]: PAM: Permission denied for root from 218.92.0.158 Apr 30 12:47:48.780185 sshd[9765]: Received disconnect from 218.92.0.158 port 26073:11: [preauth] Apr 30 12:47:48.780185 sshd[9765]: Disconnected from authenticating user root 218.92.0.158 port 26073 [preauth] Apr 30 12:47:48.783074 systemd[1]: sshd@10-147.75.61.11:22-218.92.0.158:26073.service: Deactivated successfully. Apr 30 12:48:14.665890 update_engine[2695]: I20250430 12:48:14.665825 2695 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Apr 30 12:48:14.665890 update_engine[2695]: I20250430 12:48:14.665877 2695 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Apr 30 12:48:14.666366 update_engine[2695]: I20250430 12:48:14.666136 2695 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Apr 30 12:48:14.666471 update_engine[2695]: I20250430 12:48:14.666455 2695 omaha_request_params.cc:62] Current group set to beta Apr 30 12:48:14.666550 update_engine[2695]: I20250430 12:48:14.666538 2695 update_attempter.cc:499] Already updated boot flags. Skipping. Apr 30 12:48:14.666577 update_engine[2695]: I20250430 12:48:14.666548 2695 update_attempter.cc:643] Scheduling an action processor start. Apr 30 12:48:14.666577 update_engine[2695]: I20250430 12:48:14.666561 2695 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 30 12:48:14.666615 update_engine[2695]: I20250430 12:48:14.666587 2695 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Apr 30 12:48:14.666640 update_engine[2695]: I20250430 12:48:14.666630 2695 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 30 12:48:14.666660 update_engine[2695]: I20250430 12:48:14.666637 2695 omaha_request_action.cc:272] Request: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: Apr 30 12:48:14.666660 update_engine[2695]: I20250430 12:48:14.666643 2695 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 30 12:48:14.666858 locksmithd[2725]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Apr 30 12:48:14.667628 update_engine[2695]: I20250430 12:48:14.667611 2695 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 30 12:48:14.667926 update_engine[2695]: I20250430 12:48:14.667905 2695 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 30 12:48:14.668404 update_engine[2695]: E20250430 12:48:14.668390 2695 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 30 12:48:14.668448 update_engine[2695]: I20250430 12:48:14.668435 2695 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Apr 30 12:48:24.575449 update_engine[2695]: I20250430 12:48:24.575379 2695 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 30 12:48:24.575889 update_engine[2695]: I20250430 12:48:24.575665 2695 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 30 12:48:24.575925 update_engine[2695]: I20250430 12:48:24.575904 2695 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 30 12:48:24.576319 update_engine[2695]: E20250430 12:48:24.576303 2695 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 30 12:48:24.576347 update_engine[2695]: I20250430 12:48:24.576337 2695 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Apr 30 12:48:34.576129 update_engine[2695]: I20250430 12:48:34.576010 2695 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 30 12:48:34.576555 update_engine[2695]: I20250430 12:48:34.576220 2695 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 30 12:48:34.576555 update_engine[2695]: I20250430 12:48:34.576438 2695 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 30 12:48:34.576842 update_engine[2695]: E20250430 12:48:34.576823 2695 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 30 12:48:34.576872 update_engine[2695]: I20250430 12:48:34.576860 2695 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Apr 30 12:48:44.574970 update_engine[2695]: I20250430 12:48:44.574900 2695 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 30 12:48:44.575404 update_engine[2695]: I20250430 12:48:44.575141 2695 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 30 12:48:44.575404 update_engine[2695]: I20250430 12:48:44.575338 2695 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 30 12:48:44.575725 update_engine[2695]: E20250430 12:48:44.575708 2695 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 30 12:48:44.575770 update_engine[2695]: I20250430 12:48:44.575758 2695 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 30 12:48:44.575796 update_engine[2695]: I20250430 12:48:44.575768 2695 omaha_request_action.cc:617] Omaha request response: Apr 30 12:48:44.575849 update_engine[2695]: E20250430 12:48:44.575837 2695 omaha_request_action.cc:636] Omaha request network transfer failed. Apr 30 12:48:44.575871 update_engine[2695]: I20250430 12:48:44.575853 2695 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Apr 30 12:48:44.575871 update_engine[2695]: I20250430 12:48:44.575860 2695 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 30 12:48:44.575871 update_engine[2695]: I20250430 12:48:44.575863 2695 update_attempter.cc:306] Processing Done. Apr 30 12:48:44.575925 update_engine[2695]: E20250430 12:48:44.575876 2695 update_attempter.cc:619] Update failed. Apr 30 12:48:44.575925 update_engine[2695]: I20250430 12:48:44.575881 2695 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Apr 30 12:48:44.575925 update_engine[2695]: I20250430 12:48:44.575886 2695 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Apr 30 12:48:44.575925 update_engine[2695]: I20250430 12:48:44.575891 2695 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Apr 30 12:48:44.576000 update_engine[2695]: I20250430 12:48:44.575955 2695 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 30 12:48:44.576000 update_engine[2695]: I20250430 12:48:44.575974 2695 omaha_request_action.cc:271] Posting an Omaha request to disabled Apr 30 12:48:44.576000 update_engine[2695]: I20250430 12:48:44.575979 2695 omaha_request_action.cc:272] Request: Apr 30 12:48:44.576000 update_engine[2695]: Apr 30 12:48:44.576000 update_engine[2695]: Apr 30 12:48:44.576000 update_engine[2695]: Apr 30 12:48:44.576000 update_engine[2695]: Apr 30 12:48:44.576000 update_engine[2695]: Apr 30 12:48:44.576000 update_engine[2695]: Apr 30 12:48:44.576000 update_engine[2695]: I20250430 12:48:44.575984 2695 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 30 12:48:44.576163 update_engine[2695]: I20250430 12:48:44.576097 2695 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 30 12:48:44.576311 update_engine[2695]: I20250430 12:48:44.576244 2695 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 30 12:48:44.576427 locksmithd[2725]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Apr 30 12:48:44.576632 update_engine[2695]: E20250430 12:48:44.576524 2695 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576553 2695 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576558 2695 omaha_request_action.cc:617] Omaha request response: Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576565 2695 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576568 2695 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576572 2695 update_attempter.cc:306] Processing Done. Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576577 2695 update_attempter.cc:310] Error event sent. Apr 30 12:48:44.576632 update_engine[2695]: I20250430 12:48:44.576584 2695 update_check_scheduler.cc:74] Next update check in 43m16s Apr 30 12:48:44.576916 locksmithd[2725]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Apr 30 12:50:06.851705 systemd[1]: Started sshd@11-147.75.61.11:22-218.92.0.158:35422.service - OpenSSH per-connection server daemon (218.92.0.158:35422). Apr 30 12:52:05.613755 systemd[1]: Started sshd@12-147.75.61.11:22-139.178.68.195:33616.service - OpenSSH per-connection server daemon (139.178.68.195:33616). Apr 30 12:52:06.023788 sshd[10407]: Accepted publickey for core from 139.178.68.195 port 33616 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:06.024853 sshd-session[10407]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:06.028725 systemd-logind[2684]: New session 10 of user core. Apr 30 12:52:06.041843 systemd[1]: Started session-10.scope - Session 10 of User core. Apr 30 12:52:06.379255 sshd[10409]: Connection closed by 139.178.68.195 port 33616 Apr 30 12:52:06.379590 sshd-session[10407]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:06.382393 systemd[1]: sshd@12-147.75.61.11:22-139.178.68.195:33616.service: Deactivated successfully. Apr 30 12:52:06.384095 systemd[1]: session-10.scope: Deactivated successfully. Apr 30 12:52:06.384642 systemd-logind[2684]: Session 10 logged out. Waiting for processes to exit. Apr 30 12:52:06.385257 systemd-logind[2684]: Removed session 10. Apr 30 12:52:06.872225 systemd[1]: sshd@11-147.75.61.11:22-218.92.0.158:35422.service: Deactivated successfully. Apr 30 12:52:11.452821 systemd[1]: Started sshd@13-147.75.61.11:22-139.178.68.195:33628.service - OpenSSH per-connection server daemon (139.178.68.195:33628). Apr 30 12:52:11.865909 sshd[10485]: Accepted publickey for core from 139.178.68.195 port 33628 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:11.866922 sshd-session[10485]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:11.870109 systemd-logind[2684]: New session 11 of user core. Apr 30 12:52:11.886903 systemd[1]: Started session-11.scope - Session 11 of User core. Apr 30 12:52:12.214420 sshd[10487]: Connection closed by 139.178.68.195 port 33628 Apr 30 12:52:12.214912 sshd-session[10485]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:12.218409 systemd[1]: sshd@13-147.75.61.11:22-139.178.68.195:33628.service: Deactivated successfully. Apr 30 12:52:12.220235 systemd[1]: session-11.scope: Deactivated successfully. Apr 30 12:52:12.221271 systemd-logind[2684]: Session 11 logged out. Waiting for processes to exit. Apr 30 12:52:12.221874 systemd-logind[2684]: Removed session 11. Apr 30 12:52:12.285584 systemd[1]: Started sshd@14-147.75.61.11:22-139.178.68.195:33636.service - OpenSSH per-connection server daemon (139.178.68.195:33636). Apr 30 12:52:12.690408 sshd[10522]: Accepted publickey for core from 139.178.68.195 port 33636 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:12.691410 sshd-session[10522]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:12.694400 systemd-logind[2684]: New session 12 of user core. Apr 30 12:52:12.704907 systemd[1]: Started session-12.scope - Session 12 of User core. Apr 30 12:52:13.062430 sshd[10524]: Connection closed by 139.178.68.195 port 33636 Apr 30 12:52:13.062815 sshd-session[10522]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:13.065688 systemd[1]: sshd@14-147.75.61.11:22-139.178.68.195:33636.service: Deactivated successfully. Apr 30 12:52:13.067484 systemd[1]: session-12.scope: Deactivated successfully. Apr 30 12:52:13.068084 systemd-logind[2684]: Session 12 logged out. Waiting for processes to exit. Apr 30 12:52:13.068665 systemd-logind[2684]: Removed session 12. Apr 30 12:52:13.134654 systemd[1]: Started sshd@15-147.75.61.11:22-139.178.68.195:33648.service - OpenSSH per-connection server daemon (139.178.68.195:33648). Apr 30 12:52:13.541725 sshd[10560]: Accepted publickey for core from 139.178.68.195 port 33648 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:13.542863 sshd-session[10560]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:13.545843 systemd-logind[2684]: New session 13 of user core. Apr 30 12:52:13.556912 systemd[1]: Started session-13.scope - Session 13 of User core. Apr 30 12:52:13.891725 sshd[10570]: Connection closed by 139.178.68.195 port 33648 Apr 30 12:52:13.892058 sshd-session[10560]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:13.894947 systemd[1]: sshd@15-147.75.61.11:22-139.178.68.195:33648.service: Deactivated successfully. Apr 30 12:52:13.897463 systemd[1]: session-13.scope: Deactivated successfully. Apr 30 12:52:13.898128 systemd-logind[2684]: Session 13 logged out. Waiting for processes to exit. Apr 30 12:52:13.898708 systemd-logind[2684]: Removed session 13. Apr 30 12:52:18.969801 systemd[1]: Started sshd@16-147.75.61.11:22-139.178.68.195:50656.service - OpenSSH per-connection server daemon (139.178.68.195:50656). Apr 30 12:52:19.374938 sshd[10617]: Accepted publickey for core from 139.178.68.195 port 50656 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:19.376051 sshd-session[10617]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:19.379266 systemd-logind[2684]: New session 14 of user core. Apr 30 12:52:19.393912 systemd[1]: Started session-14.scope - Session 14 of User core. Apr 30 12:52:19.720328 sshd[10619]: Connection closed by 139.178.68.195 port 50656 Apr 30 12:52:19.720848 sshd-session[10617]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:19.724087 systemd[1]: sshd@16-147.75.61.11:22-139.178.68.195:50656.service: Deactivated successfully. Apr 30 12:52:19.726330 systemd[1]: session-14.scope: Deactivated successfully. Apr 30 12:52:19.726916 systemd-logind[2684]: Session 14 logged out. Waiting for processes to exit. Apr 30 12:52:19.727441 systemd-logind[2684]: Removed session 14. Apr 30 12:52:19.800631 systemd[1]: Started sshd@17-147.75.61.11:22-139.178.68.195:50664.service - OpenSSH per-connection server daemon (139.178.68.195:50664). Apr 30 12:52:20.208983 sshd[10652]: Accepted publickey for core from 139.178.68.195 port 50664 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:20.210020 sshd-session[10652]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:20.213051 systemd-logind[2684]: New session 15 of user core. Apr 30 12:52:20.222857 systemd[1]: Started session-15.scope - Session 15 of User core. Apr 30 12:52:20.746428 sshd[10654]: Connection closed by 139.178.68.195 port 50664 Apr 30 12:52:20.746860 sshd-session[10652]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:20.750318 systemd[1]: sshd@17-147.75.61.11:22-139.178.68.195:50664.service: Deactivated successfully. Apr 30 12:52:20.752114 systemd[1]: session-15.scope: Deactivated successfully. Apr 30 12:52:20.752720 systemd-logind[2684]: Session 15 logged out. Waiting for processes to exit. Apr 30 12:52:20.753317 systemd-logind[2684]: Removed session 15. Apr 30 12:52:20.818595 systemd[1]: Started sshd@18-147.75.61.11:22-139.178.68.195:50672.service - OpenSSH per-connection server daemon (139.178.68.195:50672). Apr 30 12:52:21.225881 sshd[10689]: Accepted publickey for core from 139.178.68.195 port 50672 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:21.226894 sshd-session[10689]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:21.229854 systemd-logind[2684]: New session 16 of user core. Apr 30 12:52:21.246837 systemd[1]: Started session-16.scope - Session 16 of User core. Apr 30 12:52:22.534874 sshd[10691]: Connection closed by 139.178.68.195 port 50672 Apr 30 12:52:22.535283 sshd-session[10689]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:22.538136 systemd[1]: sshd@18-147.75.61.11:22-139.178.68.195:50672.service: Deactivated successfully. Apr 30 12:52:22.539824 systemd[1]: session-16.scope: Deactivated successfully. Apr 30 12:52:22.540015 systemd[1]: session-16.scope: Consumed 3.450s CPU time, 117.1M memory peak. Apr 30 12:52:22.540370 systemd-logind[2684]: Session 16 logged out. Waiting for processes to exit. Apr 30 12:52:22.540923 systemd-logind[2684]: Removed session 16. Apr 30 12:52:22.606549 systemd[1]: Started sshd@19-147.75.61.11:22-139.178.68.195:50674.service - OpenSSH per-connection server daemon (139.178.68.195:50674). Apr 30 12:52:23.011753 sshd[10787]: Accepted publickey for core from 139.178.68.195 port 50674 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:23.012773 sshd-session[10787]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:23.015880 systemd-logind[2684]: New session 17 of user core. Apr 30 12:52:23.030841 systemd[1]: Started session-17.scope - Session 17 of User core. Apr 30 12:52:23.450246 sshd[10789]: Connection closed by 139.178.68.195 port 50674 Apr 30 12:52:23.450606 sshd-session[10787]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:23.453478 systemd[1]: sshd@19-147.75.61.11:22-139.178.68.195:50674.service: Deactivated successfully. Apr 30 12:52:23.455161 systemd[1]: session-17.scope: Deactivated successfully. Apr 30 12:52:23.455726 systemd-logind[2684]: Session 17 logged out. Waiting for processes to exit. Apr 30 12:52:23.456305 systemd-logind[2684]: Removed session 17. Apr 30 12:52:23.525519 systemd[1]: Started sshd@20-147.75.61.11:22-139.178.68.195:50690.service - OpenSSH per-connection server daemon (139.178.68.195:50690). Apr 30 12:52:23.935101 sshd[10833]: Accepted publickey for core from 139.178.68.195 port 50690 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:23.936095 sshd-session[10833]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:23.938956 systemd-logind[2684]: New session 18 of user core. Apr 30 12:52:23.948836 systemd[1]: Started session-18.scope - Session 18 of User core. Apr 30 12:52:24.282329 sshd[10835]: Connection closed by 139.178.68.195 port 50690 Apr 30 12:52:24.282728 sshd-session[10833]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:24.285543 systemd[1]: sshd@20-147.75.61.11:22-139.178.68.195:50690.service: Deactivated successfully. Apr 30 12:52:24.287215 systemd[1]: session-18.scope: Deactivated successfully. Apr 30 12:52:24.287781 systemd-logind[2684]: Session 18 logged out. Waiting for processes to exit. Apr 30 12:52:24.288403 systemd-logind[2684]: Removed session 18. Apr 30 12:52:24.743574 systemd[1]: Started sshd@21-147.75.61.11:22-218.92.0.158:21080.service - OpenSSH per-connection server daemon (218.92.0.158:21080). Apr 30 12:52:29.356703 systemd[1]: Started sshd@22-147.75.61.11:22-139.178.68.195:57452.service - OpenSSH per-connection server daemon (139.178.68.195:57452). Apr 30 12:52:29.766550 sshd[10896]: Accepted publickey for core from 139.178.68.195 port 57452 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:29.767527 sshd-session[10896]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:29.770533 systemd-logind[2684]: New session 19 of user core. Apr 30 12:52:29.781902 systemd[1]: Started session-19.scope - Session 19 of User core. Apr 30 12:52:30.115213 sshd[10898]: Connection closed by 139.178.68.195 port 57452 Apr 30 12:52:30.115601 sshd-session[10896]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:30.118437 systemd[1]: sshd@22-147.75.61.11:22-139.178.68.195:57452.service: Deactivated successfully. Apr 30 12:52:30.120162 systemd[1]: session-19.scope: Deactivated successfully. Apr 30 12:52:30.120727 systemd-logind[2684]: Session 19 logged out. Waiting for processes to exit. Apr 30 12:52:30.121290 systemd-logind[2684]: Removed session 19. Apr 30 12:52:35.191701 systemd[1]: Started sshd@23-147.75.61.11:22-139.178.68.195:33244.service - OpenSSH per-connection server daemon (139.178.68.195:33244). Apr 30 12:52:35.597638 sshd[10931]: Accepted publickey for core from 139.178.68.195 port 33244 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:35.598803 sshd-session[10931]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:35.601981 systemd-logind[2684]: New session 20 of user core. Apr 30 12:52:35.611846 systemd[1]: Started session-20.scope - Session 20 of User core. Apr 30 12:52:35.943596 sshd[10952]: Connection closed by 139.178.68.195 port 33244 Apr 30 12:52:35.943915 sshd-session[10931]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:35.946716 systemd[1]: sshd@23-147.75.61.11:22-139.178.68.195:33244.service: Deactivated successfully. Apr 30 12:52:35.949052 systemd[1]: session-20.scope: Deactivated successfully. Apr 30 12:52:35.949602 systemd-logind[2684]: Session 20 logged out. Waiting for processes to exit. Apr 30 12:52:35.950157 systemd-logind[2684]: Removed session 20. Apr 30 12:52:41.016749 systemd[1]: Started sshd@24-147.75.61.11:22-139.178.68.195:33254.service - OpenSSH per-connection server daemon (139.178.68.195:33254). Apr 30 12:52:41.426066 sshd[11011]: Accepted publickey for core from 139.178.68.195 port 33254 ssh2: RSA SHA256:17jUBniwDtjf/NDymVONNq8+sFmdschuf05DImChbOE Apr 30 12:52:41.427186 sshd-session[11011]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Apr 30 12:52:41.430135 systemd-logind[2684]: New session 21 of user core. Apr 30 12:52:41.439834 systemd[1]: Started session-21.scope - Session 21 of User core. Apr 30 12:52:41.773516 sshd[11013]: Connection closed by 139.178.68.195 port 33254 Apr 30 12:52:41.773872 sshd-session[11011]: pam_unix(sshd:session): session closed for user core Apr 30 12:52:41.776750 systemd[1]: sshd@24-147.75.61.11:22-139.178.68.195:33254.service: Deactivated successfully. Apr 30 12:52:41.778421 systemd[1]: session-21.scope: Deactivated successfully. Apr 30 12:52:41.778993 systemd-logind[2684]: Session 21 logged out. Waiting for processes to exit. Apr 30 12:52:41.779515 systemd-logind[2684]: Removed session 21.