May 13 23:41:50.369428 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] May 13 23:41:50.369451 kernel: Linux version 6.6.89-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.44 p1) 2.44.0) #1 SMP PREEMPT Tue May 13 22:16:18 -00 2025 May 13 23:41:50.369459 kernel: KASLR enabled May 13 23:41:50.369465 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') May 13 23:41:50.369472 kernel: printk: bootconsole [pl11] enabled May 13 23:41:50.369477 kernel: efi: EFI v2.7 by EDK II May 13 23:41:50.369484 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3f20e698 RNG=0x3fd5f998 MEMRESERVE=0x3e477598 May 13 23:41:50.369490 kernel: random: crng init done May 13 23:41:50.369496 kernel: secureboot: Secure boot disabled May 13 23:41:50.369502 kernel: ACPI: Early table checksum verification disabled May 13 23:41:50.369508 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) May 13 23:41:50.369513 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369519 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369527 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) May 13 23:41:50.369534 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369540 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369546 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369554 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369560 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369566 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369573 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) May 13 23:41:50.369578 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) May 13 23:41:50.369585 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 May 13 23:41:50.369591 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] May 13 23:41:50.369597 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] May 13 23:41:50.369603 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] May 13 23:41:50.369609 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] May 13 23:41:50.369615 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] May 13 23:41:50.369622 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] May 13 23:41:50.369628 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] May 13 23:41:50.369635 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] May 13 23:41:50.369641 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] May 13 23:41:50.369647 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] May 13 23:41:50.369653 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] May 13 23:41:50.369659 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] May 13 23:41:50.369665 kernel: NUMA: NODE_DATA [mem 0x1bf7ee800-0x1bf7f3fff] May 13 23:41:50.369672 kernel: Zone ranges: May 13 23:41:50.369692 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] May 13 23:41:50.369698 kernel: DMA32 empty May 13 23:41:50.369705 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] May 13 23:41:50.369716 kernel: Movable zone start for each node May 13 23:41:50.369722 kernel: Early memory node ranges May 13 23:41:50.369729 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] May 13 23:41:50.369735 kernel: node 0: [mem 0x0000000000824000-0x000000003e45ffff] May 13 23:41:50.369742 kernel: node 0: [mem 0x000000003e460000-0x000000003e46ffff] May 13 23:41:50.369750 kernel: node 0: [mem 0x000000003e470000-0x000000003e54ffff] May 13 23:41:50.369756 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] May 13 23:41:50.369763 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] May 13 23:41:50.369769 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] May 13 23:41:50.369775 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] May 13 23:41:50.369782 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] May 13 23:41:50.369788 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] May 13 23:41:50.369794 kernel: On node 0, zone DMA: 36 pages in unavailable ranges May 13 23:41:50.369801 kernel: psci: probing for conduit method from ACPI. May 13 23:41:50.369807 kernel: psci: PSCIv1.1 detected in firmware. May 13 23:41:50.369813 kernel: psci: Using standard PSCI v0.2 function IDs May 13 23:41:50.369820 kernel: psci: MIGRATE_INFO_TYPE not supported. May 13 23:41:50.369828 kernel: psci: SMC Calling Convention v1.4 May 13 23:41:50.369835 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 May 13 23:41:50.369841 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 May 13 23:41:50.369848 kernel: percpu: Embedded 31 pages/cpu s86632 r8192 d32152 u126976 May 13 23:41:50.369854 kernel: pcpu-alloc: s86632 r8192 d32152 u126976 alloc=31*4096 May 13 23:41:50.369861 kernel: pcpu-alloc: [0] 0 [0] 1 May 13 23:41:50.369867 kernel: Detected PIPT I-cache on CPU0 May 13 23:41:50.369874 kernel: CPU features: detected: GIC system register CPU interface May 13 23:41:50.369880 kernel: CPU features: detected: Hardware dirty bit management May 13 23:41:50.369886 kernel: CPU features: detected: Spectre-BHB May 13 23:41:50.369893 kernel: CPU features: kernel page table isolation forced ON by KASLR May 13 23:41:50.369901 kernel: CPU features: detected: Kernel page table isolation (KPTI) May 13 23:41:50.369907 kernel: CPU features: detected: ARM erratum 1418040 May 13 23:41:50.369914 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) May 13 23:41:50.369920 kernel: CPU features: detected: SSBS not fully self-synchronizing May 13 23:41:50.369927 kernel: alternatives: applying boot alternatives May 13 23:41:50.369934 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=3174b2682629aa8ad4069807ed6fd62c10f62266ee1e150a1104f2a2fb6489b5 May 13 23:41:50.369941 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 13 23:41:50.371983 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 13 23:41:50.372000 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) May 13 23:41:50.372007 kernel: Fallback order for Node 0: 0 May 13 23:41:50.372013 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 May 13 23:41:50.372026 kernel: Policy zone: Normal May 13 23:41:50.372033 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 13 23:41:50.372039 kernel: software IO TLB: area num 2. May 13 23:41:50.372046 kernel: software IO TLB: mapped [mem 0x0000000036520000-0x000000003a520000] (64MB) May 13 23:41:50.372053 kernel: Memory: 3983460K/4194160K available (10368K kernel code, 2186K rwdata, 8100K rodata, 38464K init, 897K bss, 210700K reserved, 0K cma-reserved) May 13 23:41:50.372060 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 13 23:41:50.372066 kernel: rcu: Preemptible hierarchical RCU implementation. May 13 23:41:50.372074 kernel: rcu: RCU event tracing is enabled. May 13 23:41:50.372080 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 13 23:41:50.372087 kernel: Trampoline variant of Tasks RCU enabled. May 13 23:41:50.372093 kernel: Tracing variant of Tasks RCU enabled. May 13 23:41:50.372102 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 13 23:41:50.372108 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 13 23:41:50.372115 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 May 13 23:41:50.372121 kernel: GICv3: 960 SPIs implemented May 13 23:41:50.372127 kernel: GICv3: 0 Extended SPIs implemented May 13 23:41:50.372134 kernel: Root IRQ handler: gic_handle_irq May 13 23:41:50.372140 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI May 13 23:41:50.372146 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 May 13 23:41:50.372153 kernel: ITS: No ITS available, not enabling LPIs May 13 23:41:50.372160 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. May 13 23:41:50.372166 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 13 23:41:50.372173 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). May 13 23:41:50.372181 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns May 13 23:41:50.372188 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns May 13 23:41:50.372194 kernel: Console: colour dummy device 80x25 May 13 23:41:50.372201 kernel: printk: console [tty1] enabled May 13 23:41:50.372208 kernel: ACPI: Core revision 20230628 May 13 23:41:50.372215 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) May 13 23:41:50.372222 kernel: pid_max: default: 32768 minimum: 301 May 13 23:41:50.372228 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity May 13 23:41:50.372235 kernel: landlock: Up and running. May 13 23:41:50.372243 kernel: SELinux: Initializing. May 13 23:41:50.372250 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 13 23:41:50.372257 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) May 13 23:41:50.372264 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 13 23:41:50.372270 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. May 13 23:41:50.372277 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 May 13 23:41:50.372284 kernel: Hyper-V: Host Build 10.0.22477.1619-1-0 May 13 23:41:50.372298 kernel: Hyper-V: enabling crash_kexec_post_notifiers May 13 23:41:50.372305 kernel: rcu: Hierarchical SRCU implementation. May 13 23:41:50.372312 kernel: rcu: Max phase no-delay instances is 400. May 13 23:41:50.372319 kernel: Remapping and enabling EFI services. May 13 23:41:50.372326 kernel: smp: Bringing up secondary CPUs ... May 13 23:41:50.372334 kernel: Detected PIPT I-cache on CPU1 May 13 23:41:50.372342 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 May 13 23:41:50.372349 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 May 13 23:41:50.372356 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] May 13 23:41:50.372363 kernel: smp: Brought up 1 node, 2 CPUs May 13 23:41:50.372371 kernel: SMP: Total of 2 processors activated. May 13 23:41:50.372378 kernel: CPU features: detected: 32-bit EL0 Support May 13 23:41:50.372385 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence May 13 23:41:50.372392 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence May 13 23:41:50.372399 kernel: CPU features: detected: CRC32 instructions May 13 23:41:50.372406 kernel: CPU features: detected: RCpc load-acquire (LDAPR) May 13 23:41:50.372413 kernel: CPU features: detected: LSE atomic instructions May 13 23:41:50.372420 kernel: CPU features: detected: Privileged Access Never May 13 23:41:50.372427 kernel: CPU: All CPU(s) started at EL1 May 13 23:41:50.372436 kernel: alternatives: applying system-wide alternatives May 13 23:41:50.372443 kernel: devtmpfs: initialized May 13 23:41:50.372450 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 13 23:41:50.372457 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 13 23:41:50.372464 kernel: pinctrl core: initialized pinctrl subsystem May 13 23:41:50.372471 kernel: SMBIOS 3.1.0 present. May 13 23:41:50.372478 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 May 13 23:41:50.372485 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 13 23:41:50.372492 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations May 13 23:41:50.372500 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations May 13 23:41:50.372507 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations May 13 23:41:50.372514 kernel: audit: initializing netlink subsys (disabled) May 13 23:41:50.372522 kernel: audit: type=2000 audit(0.046:1): state=initialized audit_enabled=0 res=1 May 13 23:41:50.372529 kernel: thermal_sys: Registered thermal governor 'step_wise' May 13 23:41:50.372535 kernel: cpuidle: using governor menu May 13 23:41:50.372542 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. May 13 23:41:50.372549 kernel: ASID allocator initialised with 32768 entries May 13 23:41:50.372556 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 13 23:41:50.372565 kernel: Serial: AMBA PL011 UART driver May 13 23:41:50.372572 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL May 13 23:41:50.372579 kernel: Modules: 0 pages in range for non-PLT usage May 13 23:41:50.372586 kernel: Modules: 509232 pages in range for PLT usage May 13 23:41:50.372593 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages May 13 23:41:50.372600 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page May 13 23:41:50.372607 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages May 13 23:41:50.372614 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page May 13 23:41:50.372621 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages May 13 23:41:50.372629 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page May 13 23:41:50.372636 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages May 13 23:41:50.372643 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page May 13 23:41:50.372650 kernel: ACPI: Added _OSI(Module Device) May 13 23:41:50.372657 kernel: ACPI: Added _OSI(Processor Device) May 13 23:41:50.372664 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 13 23:41:50.372671 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 13 23:41:50.372678 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded May 13 23:41:50.372685 kernel: ACPI: Interpreter enabled May 13 23:41:50.372693 kernel: ACPI: Using GIC for interrupt routing May 13 23:41:50.372700 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA May 13 23:41:50.372707 kernel: printk: console [ttyAMA0] enabled May 13 23:41:50.372714 kernel: printk: bootconsole [pl11] disabled May 13 23:41:50.372721 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA May 13 23:41:50.372729 kernel: iommu: Default domain type: Translated May 13 23:41:50.372735 kernel: iommu: DMA domain TLB invalidation policy: strict mode May 13 23:41:50.372742 kernel: efivars: Registered efivars operations May 13 23:41:50.372749 kernel: vgaarb: loaded May 13 23:41:50.372759 kernel: clocksource: Switched to clocksource arch_sys_counter May 13 23:41:50.372766 kernel: VFS: Disk quotas dquot_6.6.0 May 13 23:41:50.372773 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 13 23:41:50.372780 kernel: pnp: PnP ACPI init May 13 23:41:50.372787 kernel: pnp: PnP ACPI: found 0 devices May 13 23:41:50.372794 kernel: NET: Registered PF_INET protocol family May 13 23:41:50.372801 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) May 13 23:41:50.372809 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) May 13 23:41:50.372816 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 13 23:41:50.372825 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) May 13 23:41:50.372832 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) May 13 23:41:50.372839 kernel: TCP: Hash tables configured (established 32768 bind 32768) May 13 23:41:50.372846 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) May 13 23:41:50.372853 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) May 13 23:41:50.372860 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 13 23:41:50.372867 kernel: PCI: CLS 0 bytes, default 64 May 13 23:41:50.372873 kernel: kvm [1]: HYP mode not available May 13 23:41:50.372880 kernel: Initialise system trusted keyrings May 13 23:41:50.372889 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 May 13 23:41:50.372896 kernel: Key type asymmetric registered May 13 23:41:50.372902 kernel: Asymmetric key parser 'x509' registered May 13 23:41:50.372909 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) May 13 23:41:50.372916 kernel: io scheduler mq-deadline registered May 13 23:41:50.372923 kernel: io scheduler kyber registered May 13 23:41:50.372930 kernel: io scheduler bfq registered May 13 23:41:50.372937 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 13 23:41:50.372944 kernel: thunder_xcv, ver 1.0 May 13 23:41:50.372963 kernel: thunder_bgx, ver 1.0 May 13 23:41:50.372970 kernel: nicpf, ver 1.0 May 13 23:41:50.372977 kernel: nicvf, ver 1.0 May 13 23:41:50.373114 kernel: rtc-efi rtc-efi.0: registered as rtc0 May 13 23:41:50.373184 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-05-13T23:41:49 UTC (1747179709) May 13 23:41:50.373194 kernel: efifb: probing for efifb May 13 23:41:50.373201 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k May 13 23:41:50.373208 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 May 13 23:41:50.373218 kernel: efifb: scrolling: redraw May 13 23:41:50.373225 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 May 13 23:41:50.373232 kernel: Console: switching to colour frame buffer device 128x48 May 13 23:41:50.373239 kernel: fb0: EFI VGA frame buffer device May 13 23:41:50.373246 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... May 13 23:41:50.373253 kernel: hid: raw HID events driver (C) Jiri Kosina May 13 23:41:50.373260 kernel: No ACPI PMU IRQ for CPU0 May 13 23:41:50.373267 kernel: No ACPI PMU IRQ for CPU1 May 13 23:41:50.373274 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available May 13 23:41:50.373283 kernel: watchdog: Delayed init of the lockup detector failed: -19 May 13 23:41:50.373290 kernel: watchdog: Hard watchdog permanently disabled May 13 23:41:50.373296 kernel: NET: Registered PF_INET6 protocol family May 13 23:41:50.373303 kernel: Segment Routing with IPv6 May 13 23:41:50.373310 kernel: In-situ OAM (IOAM) with IPv6 May 13 23:41:50.373317 kernel: NET: Registered PF_PACKET protocol family May 13 23:41:50.373324 kernel: Key type dns_resolver registered May 13 23:41:50.373331 kernel: registered taskstats version 1 May 13 23:41:50.373338 kernel: Loading compiled-in X.509 certificates May 13 23:41:50.373346 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.89-flatcar: 568a15bbab977599d8f910f319ba50c03c8a57bd' May 13 23:41:50.373353 kernel: Key type .fscrypt registered May 13 23:41:50.373360 kernel: Key type fscrypt-provisioning registered May 13 23:41:50.373367 kernel: ima: No TPM chip found, activating TPM-bypass! May 13 23:41:50.373374 kernel: ima: Allocated hash algorithm: sha1 May 13 23:41:50.373381 kernel: ima: No architecture policies found May 13 23:41:50.373388 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) May 13 23:41:50.373396 kernel: clk: Disabling unused clocks May 13 23:41:50.373403 kernel: Freeing unused kernel memory: 38464K May 13 23:41:50.373411 kernel: Run /init as init process May 13 23:41:50.373418 kernel: with arguments: May 13 23:41:50.373426 kernel: /init May 13 23:41:50.373432 kernel: with environment: May 13 23:41:50.373439 kernel: HOME=/ May 13 23:41:50.373446 kernel: TERM=linux May 13 23:41:50.373453 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 13 23:41:50.373460 systemd[1]: Successfully made /usr/ read-only. May 13 23:41:50.373472 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 13 23:41:50.373480 systemd[1]: Detected virtualization microsoft. May 13 23:41:50.373487 systemd[1]: Detected architecture arm64. May 13 23:41:50.373495 systemd[1]: Running in initrd. May 13 23:41:50.373502 systemd[1]: No hostname configured, using default hostname. May 13 23:41:50.373510 systemd[1]: Hostname set to . May 13 23:41:50.373517 systemd[1]: Initializing machine ID from random generator. May 13 23:41:50.373525 systemd[1]: Queued start job for default target initrd.target. May 13 23:41:50.373534 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 13 23:41:50.373542 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 13 23:41:50.373550 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... May 13 23:41:50.373558 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 13 23:41:50.373566 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... May 13 23:41:50.373574 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... May 13 23:41:50.373583 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... May 13 23:41:50.373592 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... May 13 23:41:50.373600 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 13 23:41:50.373608 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 13 23:41:50.373615 systemd[1]: Reached target paths.target - Path Units. May 13 23:41:50.373623 systemd[1]: Reached target slices.target - Slice Units. May 13 23:41:50.373631 systemd[1]: Reached target swap.target - Swaps. May 13 23:41:50.373638 systemd[1]: Reached target timers.target - Timer Units. May 13 23:41:50.373646 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. May 13 23:41:50.373655 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 13 23:41:50.373663 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). May 13 23:41:50.373671 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. May 13 23:41:50.373678 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 13 23:41:50.373686 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 13 23:41:50.373694 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 13 23:41:50.373701 systemd[1]: Reached target sockets.target - Socket Units. May 13 23:41:50.373709 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... May 13 23:41:50.373717 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 13 23:41:50.373726 systemd[1]: Finished network-cleanup.service - Network Cleanup. May 13 23:41:50.373734 systemd[1]: Starting systemd-fsck-usr.service... May 13 23:41:50.373741 systemd[1]: Starting systemd-journald.service - Journal Service... May 13 23:41:50.373749 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 13 23:41:50.373774 systemd-journald[218]: Collecting audit messages is disabled. May 13 23:41:50.373795 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 13 23:41:50.373804 systemd-journald[218]: Journal started May 13 23:41:50.373822 systemd-journald[218]: Runtime Journal (/run/log/journal/687813609f7149508452628bf47b7b93) is 8M, max 78.5M, 70.5M free. May 13 23:41:50.386694 systemd-modules-load[220]: Inserted module 'overlay' May 13 23:41:50.403265 systemd[1]: Started systemd-journald.service - Journal Service. May 13 23:41:50.421375 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. May 13 23:41:50.441091 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 13 23:41:50.441115 kernel: Bridge firewalling registered May 13 23:41:50.433895 systemd-modules-load[220]: Inserted module 'br_netfilter' May 13 23:41:50.434572 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 13 23:41:50.450973 systemd[1]: Finished systemd-fsck-usr.service. May 13 23:41:50.461973 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 13 23:41:50.475711 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 13 23:41:50.498089 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 13 23:41:50.525070 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 13 23:41:50.552101 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... May 13 23:41:50.565174 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 13 23:41:50.590929 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 13 23:41:50.602117 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 13 23:41:50.618044 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. May 13 23:41:50.632933 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 13 23:41:50.652150 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... May 13 23:41:50.678088 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 13 23:41:50.694784 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 13 23:41:50.719781 dracut-cmdline[252]: dracut-dracut-053 May 13 23:41:50.719781 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=3174b2682629aa8ad4069807ed6fd62c10f62266ee1e150a1104f2a2fb6489b5 May 13 23:41:50.722574 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 13 23:41:50.793759 systemd-resolved[253]: Positive Trust Anchors: May 13 23:41:50.793773 systemd-resolved[253]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 13 23:41:50.793804 systemd-resolved[253]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 13 23:41:50.801385 systemd-resolved[253]: Defaulting to hostname 'linux'. May 13 23:41:50.802349 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 13 23:41:50.820561 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 13 23:41:50.899971 kernel: SCSI subsystem initialized May 13 23:41:50.908962 kernel: Loading iSCSI transport class v2.0-870. May 13 23:41:50.919984 kernel: iscsi: registered transport (tcp) May 13 23:41:50.938143 kernel: iscsi: registered transport (qla4xxx) May 13 23:41:50.938176 kernel: QLogic iSCSI HBA Driver May 13 23:41:50.971463 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. May 13 23:41:50.982089 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... May 13 23:41:51.035429 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 13 23:41:51.035485 kernel: device-mapper: uevent: version 1.0.3 May 13 23:41:51.042769 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com May 13 23:41:51.091974 kernel: raid6: neonx8 gen() 15763 MB/s May 13 23:41:51.112960 kernel: raid6: neonx4 gen() 15818 MB/s May 13 23:41:51.133962 kernel: raid6: neonx2 gen() 13211 MB/s May 13 23:41:51.155958 kernel: raid6: neonx1 gen() 10513 MB/s May 13 23:41:51.176957 kernel: raid6: int64x8 gen() 6792 MB/s May 13 23:41:51.197958 kernel: raid6: int64x4 gen() 7353 MB/s May 13 23:41:51.219962 kernel: raid6: int64x2 gen() 6112 MB/s May 13 23:41:51.245475 kernel: raid6: int64x1 gen() 5059 MB/s May 13 23:41:51.245486 kernel: raid6: using algorithm neonx4 gen() 15818 MB/s May 13 23:41:51.270858 kernel: raid6: .... xor() 12395 MB/s, rmw enabled May 13 23:41:51.270869 kernel: raid6: using neon recovery algorithm May 13 23:41:51.285646 kernel: xor: measuring software checksum speed May 13 23:41:51.285660 kernel: 8regs : 21618 MB/sec May 13 23:41:51.294364 kernel: 32regs : 20464 MB/sec May 13 23:41:51.294380 kernel: arm64_neon : 28051 MB/sec May 13 23:41:51.299436 kernel: xor: using function: arm64_neon (28051 MB/sec) May 13 23:41:51.350971 kernel: Btrfs loaded, zoned=no, fsverity=no May 13 23:41:51.360284 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. May 13 23:41:51.374773 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 13 23:41:51.413026 systemd-udevd[438]: Using default interface naming scheme 'v255'. May 13 23:41:51.419916 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 13 23:41:51.433076 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... May 13 23:41:51.467376 dracut-pre-trigger[451]: rd.md=0: removing MD RAID activation May 13 23:41:51.496996 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. May 13 23:41:51.506079 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 13 23:41:51.566739 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 13 23:41:51.584102 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... May 13 23:41:51.620455 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. May 13 23:41:51.630906 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. May 13 23:41:51.653645 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 13 23:41:51.686101 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 13 23:41:51.806089 kernel: hv_vmbus: Vmbus version:5.3 May 13 23:41:51.806123 kernel: pps_core: LinuxPPS API ver. 1 registered May 13 23:41:51.806133 kernel: hv_vmbus: registering driver hyperv_keyboard May 13 23:41:51.806141 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 May 13 23:41:51.806151 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti May 13 23:41:51.806160 kernel: hv_vmbus: registering driver hid_hyperv May 13 23:41:51.806169 kernel: hv_vmbus: registering driver hv_netvsc May 13 23:41:51.806177 kernel: PTP clock support registered May 13 23:41:51.806186 kernel: hv_vmbus: registering driver hv_storvsc May 13 23:41:51.806194 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 May 13 23:41:51.806205 kernel: scsi host1: storvsc_host_t May 13 23:41:51.806367 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on May 13 23:41:51.806451 kernel: scsi host0: storvsc_host_t May 13 23:41:51.806536 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 May 13 23:41:51.806558 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 May 13 23:41:51.709810 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... May 13 23:41:51.839358 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 13 23:41:51.863907 kernel: hv_netvsc 000d3af9-af51-000d-3af9-af51000d3af9 eth0: VF slot 1 added May 13 23:41:51.864124 kernel: hv_utils: Registering HyperV Utility Driver May 13 23:41:51.864136 kernel: hv_vmbus: registering driver hv_utils May 13 23:41:51.839528 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 13 23:41:52.308968 kernel: sr 0:0:0:2: [sr0] scsi-1 drive May 13 23:41:52.309161 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 May 13 23:41:52.309173 kernel: hv_vmbus: registering driver hv_pci May 13 23:41:52.309184 kernel: hv_utils: Heartbeat IC version 3.0 May 13 23:41:52.309193 kernel: hv_utils: Shutdown IC version 3.2 May 13 23:41:52.309202 kernel: hv_pci f7396fdb-4271-40f1-84ee-fae7beb987b6: PCI VMBus probing: Using version 0x10004 May 13 23:41:52.309304 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 May 13 23:41:52.309389 kernel: hv_utils: TimeSync IC version 4.0 May 13 23:41:52.309398 kernel: hv_pci f7396fdb-4271-40f1-84ee-fae7beb987b6: PCI host bridge to bus 4271:00 May 13 23:41:51.899140 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 13 23:41:52.333958 kernel: pci_bus 4271:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] May 13 23:41:52.334143 kernel: pci_bus 4271:00: No busn resource found for root bus, will use [bus 00-ff] May 13 23:41:51.924099 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 13 23:41:52.358078 kernel: pci 4271:00:02.0: [15b3:1018] type 00 class 0x020000 May 13 23:41:52.358129 kernel: pci 4271:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] May 13 23:41:51.924316 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 13 23:41:52.378791 kernel: pci 4271:00:02.0: enabling Extended Tags May 13 23:41:52.301526 systemd-resolved[253]: Clock change detected. Flushing caches. May 13 23:41:52.415696 kernel: pci 4271:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 4271:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) May 13 23:41:52.415902 kernel: pci_bus 4271:00: busn_res: [bus 00-ff] end is updated to 00 May 13 23:41:52.416022 kernel: pci 4271:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] May 13 23:41:52.399621 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 13 23:41:52.436081 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 13 23:41:52.468907 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) May 13 23:41:52.469088 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks May 13 23:41:52.469174 kernel: sd 0:0:0:0: [sda] Write Protect is off May 13 23:41:52.469255 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 May 13 23:41:52.469342 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA May 13 23:41:52.477738 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 13 23:41:52.492504 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. May 13 23:41:52.524904 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 13 23:41:52.524926 kernel: sd 0:0:0:0: [sda] Attached SCSI disk May 13 23:41:52.530474 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 13 23:41:52.530548 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 13 23:41:52.553900 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... May 13 23:41:52.575429 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 13 23:41:52.604772 kernel: mlx5_core 4271:00:02.0: enabling device (0000 -> 0002) May 13 23:41:52.604959 kernel: mlx5_core 4271:00:02.0: firmware version: 16.30.1284 May 13 23:41:52.590489 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 13 23:41:52.625233 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 13 23:41:52.641971 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... May 13 23:41:52.683193 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 13 23:41:52.813294 kernel: hv_netvsc 000d3af9-af51-000d-3af9-af51000d3af9 eth0: VF registering: eth1 May 13 23:41:52.813545 kernel: mlx5_core 4271:00:02.0 eth1: joined to eth0 May 13 23:41:52.821863 kernel: mlx5_core 4271:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) May 13 23:41:52.833850 kernel: mlx5_core 4271:00:02.0 enP17009s1: renamed from eth1 May 13 23:41:53.017442 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. May 13 23:41:53.047847 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by (udev-worker) (500) May 13 23:41:53.074876 kernel: BTRFS: device fsid ee830c17-a93d-4109-bd12-3fec8ef6763d devid 1 transid 41 /dev/sda3 scanned by (udev-worker) (485) May 13 23:41:53.077646 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. May 13 23:41:53.107999 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. May 13 23:41:53.116201 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. May 13 23:41:53.140493 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. May 13 23:41:53.158958 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... May 13 23:41:53.194969 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 13 23:41:53.205841 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 13 23:41:54.214834 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 13 23:41:54.215190 disk-uuid[608]: The operation has completed successfully. May 13 23:41:54.271374 systemd[1]: disk-uuid.service: Deactivated successfully. May 13 23:41:54.273540 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. May 13 23:41:54.324902 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... May 13 23:41:54.356197 sh[694]: Success May 13 23:41:54.384076 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" May 13 23:41:54.543014 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. May 13 23:41:54.559920 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... May 13 23:41:54.574151 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. May 13 23:41:54.606477 kernel: BTRFS info (device dm-0): first mount of filesystem ee830c17-a93d-4109-bd12-3fec8ef6763d May 13 23:41:54.606522 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm May 13 23:41:54.615044 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead May 13 23:41:54.621117 kernel: BTRFS info (device dm-0): disabling log replay at mount time May 13 23:41:54.625961 kernel: BTRFS info (device dm-0): using free space tree May 13 23:41:54.915054 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. May 13 23:41:54.921070 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. May 13 23:41:54.923943 systemd[1]: Starting ignition-setup.service - Ignition (setup)... May 13 23:41:54.947175 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... May 13 23:41:55.002193 kernel: BTRFS info (device sda6): first mount of filesystem e7b30525-8b14-4004-ad68-68a99b3959db May 13 23:41:55.002257 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm May 13 23:41:55.009771 kernel: BTRFS info (device sda6): using free space tree May 13 23:41:55.027854 kernel: BTRFS info (device sda6): auto enabling async discard May 13 23:41:55.040851 kernel: BTRFS info (device sda6): last unmount of filesystem e7b30525-8b14-4004-ad68-68a99b3959db May 13 23:41:55.047636 systemd[1]: Finished ignition-setup.service - Ignition (setup). May 13 23:41:55.057956 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... May 13 23:41:55.093221 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 13 23:41:55.113263 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 13 23:41:55.158191 systemd-networkd[875]: lo: Link UP May 13 23:41:55.158205 systemd-networkd[875]: lo: Gained carrier May 13 23:41:55.159793 systemd-networkd[875]: Enumeration completed May 13 23:41:55.159891 systemd[1]: Started systemd-networkd.service - Network Configuration. May 13 23:41:55.169795 systemd-networkd[875]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 13 23:41:55.169799 systemd-networkd[875]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 13 23:41:55.170892 systemd[1]: Reached target network.target - Network. May 13 23:41:55.266877 kernel: mlx5_core 4271:00:02.0 enP17009s1: Link up May 13 23:41:55.322851 kernel: hv_netvsc 000d3af9-af51-000d-3af9-af51000d3af9 eth0: Data path switched to VF: enP17009s1 May 13 23:41:55.323267 systemd-networkd[875]: enP17009s1: Link UP May 13 23:41:55.323335 systemd-networkd[875]: eth0: Link UP May 13 23:41:55.323428 systemd-networkd[875]: eth0: Gained carrier May 13 23:41:55.323437 systemd-networkd[875]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 13 23:41:55.337553 systemd-networkd[875]: enP17009s1: Gained carrier May 13 23:41:55.365893 systemd-networkd[875]: eth0: DHCPv4 address 10.200.20.30/24, gateway 10.200.20.1 acquired from 168.63.129.16 May 13 23:41:55.858304 ignition[856]: Ignition 2.20.0 May 13 23:41:55.861800 ignition[856]: Stage: fetch-offline May 13 23:41:55.861860 ignition[856]: no configs at "/usr/lib/ignition/base.d" May 13 23:41:55.863273 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). May 13 23:41:55.861869 ignition[856]: no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:41:55.877994 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... May 13 23:41:55.861979 ignition[856]: parsed url from cmdline: "" May 13 23:41:55.861982 ignition[856]: no config URL provided May 13 23:41:55.861987 ignition[856]: reading system config file "/usr/lib/ignition/user.ign" May 13 23:41:55.861994 ignition[856]: no config at "/usr/lib/ignition/user.ign" May 13 23:41:55.861999 ignition[856]: failed to fetch config: resource requires networking May 13 23:41:55.862188 ignition[856]: Ignition finished successfully May 13 23:41:55.913307 ignition[885]: Ignition 2.20.0 May 13 23:41:55.913313 ignition[885]: Stage: fetch May 13 23:41:55.913535 ignition[885]: no configs at "/usr/lib/ignition/base.d" May 13 23:41:55.913545 ignition[885]: no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:41:55.913677 ignition[885]: parsed url from cmdline: "" May 13 23:41:55.913680 ignition[885]: no config URL provided May 13 23:41:55.913685 ignition[885]: reading system config file "/usr/lib/ignition/user.ign" May 13 23:41:55.913692 ignition[885]: no config at "/usr/lib/ignition/user.ign" May 13 23:41:55.913722 ignition[885]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 May 13 23:41:56.039767 ignition[885]: GET result: OK May 13 23:41:56.039847 ignition[885]: config has been read from IMDS userdata May 13 23:41:56.039889 ignition[885]: parsing config with SHA512: 60e1f91be63d0fe9004af388815830ec52d89c15fd5fa4f62fe09fc2adf901ff004b6be860753d72e75c1139a474a46b5f833ab21e64c10245b24c260da9ab36 May 13 23:41:56.044655 unknown[885]: fetched base config from "system" May 13 23:41:56.045139 ignition[885]: fetch: fetch complete May 13 23:41:56.044662 unknown[885]: fetched base config from "system" May 13 23:41:56.045145 ignition[885]: fetch: fetch passed May 13 23:41:56.044668 unknown[885]: fetched user config from "azure" May 13 23:41:56.045189 ignition[885]: Ignition finished successfully May 13 23:41:56.051019 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). May 13 23:41:56.065017 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... May 13 23:41:56.106071 ignition[891]: Ignition 2.20.0 May 13 23:41:56.106081 ignition[891]: Stage: kargs May 13 23:41:56.111023 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). May 13 23:41:56.106229 ignition[891]: no configs at "/usr/lib/ignition/base.d" May 13 23:41:56.122959 systemd[1]: Starting ignition-disks.service - Ignition (disks)... May 13 23:41:56.106238 ignition[891]: no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:41:56.107099 ignition[891]: kargs: kargs passed May 13 23:41:56.107142 ignition[891]: Ignition finished successfully May 13 23:41:56.162100 ignition[898]: Ignition 2.20.0 May 13 23:41:56.162109 ignition[898]: Stage: disks May 13 23:41:56.162264 ignition[898]: no configs at "/usr/lib/ignition/base.d" May 13 23:41:56.167893 systemd[1]: Finished ignition-disks.service - Ignition (disks). May 13 23:41:56.162273 ignition[898]: no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:41:56.176164 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. May 13 23:41:56.166541 ignition[898]: disks: disks passed May 13 23:41:56.187903 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. May 13 23:41:56.166599 ignition[898]: Ignition finished successfully May 13 23:41:56.201388 systemd[1]: Reached target local-fs.target - Local File Systems. May 13 23:41:56.213847 systemd[1]: Reached target sysinit.target - System Initialization. May 13 23:41:56.226357 systemd[1]: Reached target basic.target - Basic System. May 13 23:41:56.238991 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... May 13 23:41:56.315970 systemd-fsck[907]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks May 13 23:41:56.325383 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. May 13 23:41:56.335917 systemd[1]: Mounting sysroot.mount - /sysroot... May 13 23:41:56.407849 kernel: EXT4-fs (sda9): mounted filesystem 9f8d74e6-c079-469f-823a-18a62077a2c7 r/w with ordered data mode. Quota mode: none. May 13 23:41:56.407917 systemd[1]: Mounted sysroot.mount - /sysroot. May 13 23:41:56.413447 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. May 13 23:41:56.452190 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 13 23:41:56.470638 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... May 13 23:41:56.484845 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... May 13 23:41:56.499724 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 13 23:41:56.540401 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (918) May 13 23:41:56.540425 kernel: BTRFS info (device sda6): first mount of filesystem e7b30525-8b14-4004-ad68-68a99b3959db May 13 23:41:56.540435 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm May 13 23:41:56.499770 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. May 13 23:41:56.552223 kernel: BTRFS info (device sda6): using free space tree May 13 23:41:56.540522 systemd-networkd[875]: enP17009s1: Gained IPv6LL May 13 23:41:56.553723 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. May 13 23:41:56.568003 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... May 13 23:41:56.590850 kernel: BTRFS info (device sda6): auto enabling async discard May 13 23:41:56.596163 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 13 23:41:56.664127 systemd-networkd[875]: eth0: Gained IPv6LL May 13 23:41:56.973507 coreos-metadata[920]: May 13 23:41:56.973 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 May 13 23:41:56.984309 coreos-metadata[920]: May 13 23:41:56.984 INFO Fetch successful May 13 23:41:56.990582 coreos-metadata[920]: May 13 23:41:56.986 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 May 13 23:41:57.014388 coreos-metadata[920]: May 13 23:41:57.014 INFO Fetch successful May 13 23:41:57.026031 coreos-metadata[920]: May 13 23:41:57.025 INFO wrote hostname ci-4284.0.0-n-f756a0c81e to /sysroot/etc/hostname May 13 23:41:57.035536 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 13 23:41:57.182845 initrd-setup-root[949]: cut: /sysroot/etc/passwd: No such file or directory May 13 23:41:57.210623 initrd-setup-root[956]: cut: /sysroot/etc/group: No such file or directory May 13 23:41:57.221453 initrd-setup-root[963]: cut: /sysroot/etc/shadow: No such file or directory May 13 23:41:57.231378 initrd-setup-root[970]: cut: /sysroot/etc/gshadow: No such file or directory May 13 23:41:57.952548 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. May 13 23:41:57.963938 systemd[1]: Starting ignition-mount.service - Ignition (mount)... May 13 23:41:57.986415 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... May 13 23:41:58.004834 systemd[1]: sysroot-oem.mount: Deactivated successfully. May 13 23:41:58.018067 kernel: BTRFS info (device sda6): last unmount of filesystem e7b30525-8b14-4004-ad68-68a99b3959db May 13 23:41:58.037856 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. May 13 23:41:58.045318 ignition[1038]: INFO : Ignition 2.20.0 May 13 23:41:58.045318 ignition[1038]: INFO : Stage: mount May 13 23:41:58.045318 ignition[1038]: INFO : no configs at "/usr/lib/ignition/base.d" May 13 23:41:58.045318 ignition[1038]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:41:58.045318 ignition[1038]: INFO : mount: mount passed May 13 23:41:58.045318 ignition[1038]: INFO : Ignition finished successfully May 13 23:41:58.051596 systemd[1]: Finished ignition-mount.service - Ignition (mount). May 13 23:41:58.064941 systemd[1]: Starting ignition-files.service - Ignition (files)... May 13 23:41:58.103182 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... May 13 23:41:58.132837 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/sda6 scanned by mount (1050) May 13 23:41:58.153204 kernel: BTRFS info (device sda6): first mount of filesystem e7b30525-8b14-4004-ad68-68a99b3959db May 13 23:41:58.153266 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm May 13 23:41:58.153277 kernel: BTRFS info (device sda6): using free space tree May 13 23:41:58.161845 kernel: BTRFS info (device sda6): auto enabling async discard May 13 23:41:58.162585 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. May 13 23:41:58.195554 ignition[1068]: INFO : Ignition 2.20.0 May 13 23:41:58.195554 ignition[1068]: INFO : Stage: files May 13 23:41:58.204275 ignition[1068]: INFO : no configs at "/usr/lib/ignition/base.d" May 13 23:41:58.204275 ignition[1068]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:41:58.204275 ignition[1068]: DEBUG : files: compiled without relabeling support, skipping May 13 23:41:58.224788 ignition[1068]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 13 23:41:58.224788 ignition[1068]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 13 23:41:58.266152 ignition[1068]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 13 23:41:58.274151 ignition[1068]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 13 23:41:58.274151 ignition[1068]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 13 23:41:58.266548 unknown[1068]: wrote ssh authorized keys file for user: core May 13 23:41:58.315877 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" May 13 23:41:58.315877 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 May 13 23:41:58.356455 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK May 13 23:41:58.505839 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" May 13 23:41:58.505839 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 13 23:41:58.532231 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 May 13 23:41:58.943690 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 13 23:41:59.019811 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 23:41:59.032287 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 May 13 23:41:59.415920 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK May 13 23:41:59.635281 ignition[1068]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" May 13 23:41:59.635281 ignition[1068]: INFO : files: op(c): [started] processing unit "prepare-helm.service" May 13 23:41:59.660899 ignition[1068]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 13 23:41:59.660899 ignition[1068]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 13 23:41:59.660899 ignition[1068]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" May 13 23:41:59.660899 ignition[1068]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" May 13 23:41:59.660899 ignition[1068]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" May 13 23:41:59.660899 ignition[1068]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" May 13 23:41:59.660899 ignition[1068]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" May 13 23:41:59.660899 ignition[1068]: INFO : files: files passed May 13 23:41:59.660899 ignition[1068]: INFO : Ignition finished successfully May 13 23:41:59.660314 systemd[1]: Finished ignition-files.service - Ignition (files). May 13 23:41:59.670962 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... May 13 23:41:59.712064 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... May 13 23:41:59.725102 systemd[1]: ignition-quench.service: Deactivated successfully. May 13 23:41:59.850289 initrd-setup-root-after-ignition[1096]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 13 23:41:59.850289 initrd-setup-root-after-ignition[1096]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory May 13 23:41:59.725227 systemd[1]: Finished ignition-quench.service - Ignition (record completion). May 13 23:41:59.877979 initrd-setup-root-after-ignition[1100]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 13 23:41:59.785187 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. May 13 23:41:59.799387 systemd[1]: Reached target ignition-complete.target - Ignition Complete. May 13 23:41:59.815960 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... May 13 23:41:59.921007 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 13 23:41:59.921137 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. May 13 23:41:59.935948 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. May 13 23:41:59.952094 systemd[1]: Reached target initrd.target - Initrd Default Target. May 13 23:41:59.964946 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. May 13 23:41:59.967954 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... May 13 23:42:00.009498 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 13 23:42:00.020991 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... May 13 23:42:00.058998 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. May 13 23:42:00.068037 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. May 13 23:42:00.082706 systemd[1]: Stopped target timers.target - Timer Units. May 13 23:42:00.095542 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 13 23:42:00.095667 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. May 13 23:42:00.118212 systemd[1]: Stopped target initrd.target - Initrd Default Target. May 13 23:42:00.125328 systemd[1]: Stopped target basic.target - Basic System. May 13 23:42:00.139431 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. May 13 23:42:00.152925 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. May 13 23:42:00.164466 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. May 13 23:42:00.176790 systemd[1]: Stopped target remote-fs.target - Remote File Systems. May 13 23:42:00.188668 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. May 13 23:42:00.203379 systemd[1]: Stopped target sysinit.target - System Initialization. May 13 23:42:00.215111 systemd[1]: Stopped target local-fs.target - Local File Systems. May 13 23:42:00.227714 systemd[1]: Stopped target swap.target - Swaps. May 13 23:42:00.237740 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 13 23:42:00.237876 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. May 13 23:42:00.254200 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. May 13 23:42:00.261579 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 13 23:42:00.274337 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. May 13 23:42:00.274422 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 13 23:42:00.288354 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 13 23:42:00.288477 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. May 13 23:42:00.308028 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 13 23:42:00.308213 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. May 13 23:42:00.321101 systemd[1]: ignition-files.service: Deactivated successfully. May 13 23:42:00.321266 systemd[1]: Stopped ignition-files.service - Ignition (files). May 13 23:42:00.431996 ignition[1121]: INFO : Ignition 2.20.0 May 13 23:42:00.431996 ignition[1121]: INFO : Stage: umount May 13 23:42:00.431996 ignition[1121]: INFO : no configs at "/usr/lib/ignition/base.d" May 13 23:42:00.431996 ignition[1121]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" May 13 23:42:00.431996 ignition[1121]: INFO : umount: umount passed May 13 23:42:00.431996 ignition[1121]: INFO : Ignition finished successfully May 13 23:42:00.334988 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. May 13 23:42:00.335149 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. May 13 23:42:00.350050 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... May 13 23:42:00.361343 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 13 23:42:00.361606 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. May 13 23:42:00.380075 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... May 13 23:42:00.399483 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 13 23:42:00.399775 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. May 13 23:42:00.411903 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 13 23:42:00.412068 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. May 13 23:42:00.434867 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 13 23:42:00.435948 systemd[1]: ignition-mount.service: Deactivated successfully. May 13 23:42:00.436057 systemd[1]: Stopped ignition-mount.service - Ignition (mount). May 13 23:42:00.446459 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 13 23:42:00.446553 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. May 13 23:42:00.468867 systemd[1]: sysroot-boot.service: Deactivated successfully. May 13 23:42:00.468968 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. May 13 23:42:00.484953 systemd[1]: ignition-disks.service: Deactivated successfully. May 13 23:42:00.485014 systemd[1]: Stopped ignition-disks.service - Ignition (disks). May 13 23:42:00.495684 systemd[1]: ignition-kargs.service: Deactivated successfully. May 13 23:42:00.495735 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). May 13 23:42:00.508143 systemd[1]: ignition-fetch.service: Deactivated successfully. May 13 23:42:00.508192 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). May 13 23:42:00.522245 systemd[1]: Stopped target network.target - Network. May 13 23:42:00.537137 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 13 23:42:00.537210 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). May 13 23:42:00.551370 systemd[1]: Stopped target paths.target - Path Units. May 13 23:42:00.563420 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 13 23:42:00.566839 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 13 23:42:00.911845 kernel: hv_netvsc 000d3af9-af51-000d-3af9-af51000d3af9 eth0: Data path switched from VF: enP17009s1 May 13 23:42:00.580119 systemd[1]: Stopped target slices.target - Slice Units. May 13 23:42:00.593450 systemd[1]: Stopped target sockets.target - Socket Units. May 13 23:42:00.611897 systemd[1]: iscsid.socket: Deactivated successfully. May 13 23:42:00.611961 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. May 13 23:42:00.625512 systemd[1]: iscsiuio.socket: Deactivated successfully. May 13 23:42:00.625544 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. May 13 23:42:00.638354 systemd[1]: ignition-setup.service: Deactivated successfully. May 13 23:42:00.638406 systemd[1]: Stopped ignition-setup.service - Ignition (setup). May 13 23:42:00.651671 systemd[1]: ignition-setup-pre.service: Deactivated successfully. May 13 23:42:00.651718 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. May 13 23:42:00.665120 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 13 23:42:00.665174 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. May 13 23:42:00.678953 systemd[1]: Stopping systemd-networkd.service - Network Configuration... May 13 23:42:00.692295 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... May 13 23:42:00.715429 systemd[1]: systemd-resolved.service: Deactivated successfully. May 13 23:42:00.715566 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. May 13 23:42:00.737017 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. May 13 23:42:00.737251 systemd[1]: systemd-networkd.service: Deactivated successfully. May 13 23:42:00.737474 systemd[1]: Stopped systemd-networkd.service - Network Configuration. May 13 23:42:00.758336 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. May 13 23:42:00.759345 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 13 23:42:00.759412 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. May 13 23:42:00.775934 systemd[1]: Stopping network-cleanup.service - Network Cleanup... May 13 23:42:00.786709 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 13 23:42:00.786774 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. May 13 23:42:00.800531 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 13 23:42:00.800586 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 13 23:42:00.807297 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 13 23:42:00.807339 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. May 13 23:42:00.816224 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. May 13 23:42:00.816266 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. May 13 23:42:00.824763 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... May 13 23:42:00.834704 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 13 23:42:00.834770 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. May 13 23:42:00.863775 systemd[1]: systemd-udevd.service: Deactivated successfully. May 13 23:42:00.863998 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. May 13 23:42:01.255604 systemd-journald[218]: Received SIGTERM from PID 1 (systemd). May 13 23:42:00.876719 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 13 23:42:00.876772 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. May 13 23:42:00.899047 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 13 23:42:00.899087 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. May 13 23:42:00.919433 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 13 23:42:00.919505 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. May 13 23:42:00.938502 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 13 23:42:00.938559 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. May 13 23:42:00.950492 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 13 23:42:00.950556 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. May 13 23:42:00.972983 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... May 13 23:42:00.992733 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. May 13 23:42:00.992931 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 13 23:42:01.013757 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 13 23:42:01.013841 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. May 13 23:42:01.029193 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. May 13 23:42:01.029265 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. May 13 23:42:01.029553 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 13 23:42:01.029653 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. May 13 23:42:01.044159 systemd[1]: network-cleanup.service: Deactivated successfully. May 13 23:42:01.044255 systemd[1]: Stopped network-cleanup.service - Network Cleanup. May 13 23:42:01.057801 systemd[1]: Reached target initrd-switch-root.target - Switch Root. May 13 23:42:01.073956 systemd[1]: Starting initrd-switch-root.service - Switch Root... May 13 23:42:01.135896 systemd[1]: Switching root. May 13 23:42:01.430296 systemd-journald[218]: Journal stopped May 13 23:42:05.972709 kernel: SELinux: policy capability network_peer_controls=1 May 13 23:42:05.972737 kernel: SELinux: policy capability open_perms=1 May 13 23:42:05.972747 kernel: SELinux: policy capability extended_socket_class=1 May 13 23:42:05.972756 kernel: SELinux: policy capability always_check_network=0 May 13 23:42:05.972766 kernel: SELinux: policy capability cgroup_seclabel=1 May 13 23:42:05.972774 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 13 23:42:05.972783 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 13 23:42:05.972792 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 13 23:42:05.972800 kernel: audit: type=1403 audit(1747179722.611:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 13 23:42:05.972811 systemd[1]: Successfully loaded SELinux policy in 135.207ms. May 13 23:42:05.972844 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 10.361ms. May 13 23:42:05.972855 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) May 13 23:42:05.972864 systemd[1]: Detected virtualization microsoft. May 13 23:42:05.972877 systemd[1]: Detected architecture arm64. May 13 23:42:05.972886 systemd[1]: Detected first boot. May 13 23:42:05.972898 systemd[1]: Hostname set to . May 13 23:42:05.972908 systemd[1]: Initializing machine ID from random generator. May 13 23:42:05.972918 zram_generator::config[1164]: No configuration found. May 13 23:42:05.972928 kernel: NET: Registered PF_VSOCK protocol family May 13 23:42:05.972937 systemd[1]: Populated /etc with preset unit settings. May 13 23:42:05.972948 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. May 13 23:42:05.972957 systemd[1]: initrd-switch-root.service: Deactivated successfully. May 13 23:42:05.972969 systemd[1]: Stopped initrd-switch-root.service - Switch Root. May 13 23:42:05.972978 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. May 13 23:42:05.972988 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. May 13 23:42:05.972998 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. May 13 23:42:05.973008 systemd[1]: Created slice system-getty.slice - Slice /system/getty. May 13 23:42:05.973017 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. May 13 23:42:05.973027 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. May 13 23:42:05.973038 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. May 13 23:42:05.973048 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. May 13 23:42:05.973058 systemd[1]: Created slice user.slice - User and Session Slice. May 13 23:42:05.973068 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. May 13 23:42:05.973078 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. May 13 23:42:05.973089 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. May 13 23:42:05.973099 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. May 13 23:42:05.973109 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. May 13 23:42:05.973120 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... May 13 23:42:05.973130 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... May 13 23:42:05.973140 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). May 13 23:42:05.973153 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. May 13 23:42:05.973163 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. May 13 23:42:05.973173 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. May 13 23:42:05.973183 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. May 13 23:42:05.973203 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. May 13 23:42:05.973214 systemd[1]: Reached target remote-fs.target - Remote File Systems. May 13 23:42:05.973224 systemd[1]: Reached target slices.target - Slice Units. May 13 23:42:05.973234 systemd[1]: Reached target swap.target - Swaps. May 13 23:42:05.973244 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. May 13 23:42:05.973254 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. May 13 23:42:05.973264 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. May 13 23:42:05.973277 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. May 13 23:42:05.973287 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. May 13 23:42:05.973297 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. May 13 23:42:05.973307 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. May 13 23:42:05.973319 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... May 13 23:42:05.973329 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... May 13 23:42:05.973339 systemd[1]: Mounting media.mount - External Media Directory... May 13 23:42:05.973350 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... May 13 23:42:05.973360 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... May 13 23:42:05.973370 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... May 13 23:42:05.973381 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 13 23:42:05.973391 systemd[1]: Reached target machines.target - Containers. May 13 23:42:05.973401 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... May 13 23:42:05.973412 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 13 23:42:05.973422 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... May 13 23:42:05.973433 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... May 13 23:42:05.973444 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 13 23:42:05.973454 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 13 23:42:05.973464 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 13 23:42:05.973474 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... May 13 23:42:05.973484 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 13 23:42:05.973494 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 13 23:42:05.973504 systemd[1]: systemd-fsck-root.service: Deactivated successfully. May 13 23:42:05.973516 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. May 13 23:42:05.973526 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. May 13 23:42:05.973537 systemd[1]: Stopped systemd-fsck-usr.service. May 13 23:42:05.973547 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 13 23:42:05.973557 kernel: loop: module loaded May 13 23:42:05.973566 systemd[1]: Starting systemd-journald.service - Journal Service... May 13 23:42:05.973576 kernel: fuse: init (API version 7.39) May 13 23:42:05.973585 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... May 13 23:42:05.973595 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... May 13 23:42:05.973607 kernel: ACPI: bus type drm_connector registered May 13 23:42:05.973616 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... May 13 23:42:05.973626 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... May 13 23:42:05.973636 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... May 13 23:42:05.973667 systemd-journald[1266]: Collecting audit messages is disabled. May 13 23:42:05.973690 systemd-journald[1266]: Journal started May 13 23:42:05.973710 systemd-journald[1266]: Runtime Journal (/run/log/journal/fee706df98c54d5989ae6f06b9eba71d) is 8M, max 78.5M, 70.5M free. May 13 23:42:04.950313 systemd[1]: Queued start job for default target multi-user.target. May 13 23:42:04.958526 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. May 13 23:42:04.958904 systemd[1]: systemd-journald.service: Deactivated successfully. May 13 23:42:04.959205 systemd[1]: systemd-journald.service: Consumed 3.860s CPU time. May 13 23:42:05.987008 systemd[1]: verity-setup.service: Deactivated successfully. May 13 23:42:05.987081 systemd[1]: Stopped verity-setup.service. May 13 23:42:06.012582 systemd[1]: Started systemd-journald.service - Journal Service. May 13 23:42:06.013286 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. May 13 23:42:06.021361 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. May 13 23:42:06.028338 systemd[1]: Mounted media.mount - External Media Directory. May 13 23:42:06.034458 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. May 13 23:42:06.041312 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. May 13 23:42:06.047630 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. May 13 23:42:06.053365 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. May 13 23:42:06.062356 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. May 13 23:42:06.070806 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 13 23:42:06.070991 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. May 13 23:42:06.078366 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 23:42:06.078513 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 13 23:42:06.087640 systemd[1]: modprobe@drm.service: Deactivated successfully. May 13 23:42:06.087791 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 13 23:42:06.094705 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 23:42:06.094880 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 13 23:42:06.103478 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 13 23:42:06.103630 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. May 13 23:42:06.111487 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 23:42:06.111632 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 13 23:42:06.121853 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. May 13 23:42:06.130034 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. May 13 23:42:06.139864 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. May 13 23:42:06.149505 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. May 13 23:42:06.157605 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. May 13 23:42:06.173856 systemd[1]: Reached target network-pre.target - Preparation for Network. May 13 23:42:06.182337 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... May 13 23:42:06.197056 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... May 13 23:42:06.204477 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 13 23:42:06.204517 systemd[1]: Reached target local-fs.target - Local File Systems. May 13 23:42:06.212387 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. May 13 23:42:06.221164 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... May 13 23:42:06.240968 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... May 13 23:42:06.247702 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 13 23:42:06.254651 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... May 13 23:42:06.263517 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... May 13 23:42:06.270937 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 23:42:06.271933 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... May 13 23:42:06.278112 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 13 23:42:06.279975 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 13 23:42:06.287950 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... May 13 23:42:06.299071 systemd[1]: Starting systemd-sysusers.service - Create System Users... May 13 23:42:06.307870 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... May 13 23:42:06.319621 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. May 13 23:42:06.327372 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. May 13 23:42:06.336874 systemd-journald[1266]: Time spent on flushing to /var/log/journal/fee706df98c54d5989ae6f06b9eba71d is 22.590ms for 919 entries. May 13 23:42:06.336874 systemd-journald[1266]: System Journal (/var/log/journal/fee706df98c54d5989ae6f06b9eba71d) is 8M, max 2.6G, 2.6G free. May 13 23:42:06.393456 systemd-journald[1266]: Received client request to flush runtime journal. May 13 23:42:06.393498 kernel: loop0: detected capacity change from 0 to 103832 May 13 23:42:06.345235 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. May 13 23:42:06.354563 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. May 13 23:42:06.366831 udevadm[1308]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. May 13 23:42:06.371473 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. May 13 23:42:06.383961 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... May 13 23:42:06.395976 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. May 13 23:42:06.431419 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 13 23:42:06.487225 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 13 23:42:06.488863 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. May 13 23:42:06.606320 systemd[1]: Finished systemd-sysusers.service - Create System Users. May 13 23:42:06.616001 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... May 13 23:42:06.684880 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 13 23:42:06.709494 systemd-tmpfiles[1321]: ACLs are not supported, ignoring. May 13 23:42:06.709509 systemd-tmpfiles[1321]: ACLs are not supported, ignoring. May 13 23:42:06.713260 kernel: loop1: detected capacity change from 0 to 194096 May 13 23:42:06.715764 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. May 13 23:42:06.768842 kernel: loop2: detected capacity change from 0 to 126448 May 13 23:42:07.036836 kernel: loop3: detected capacity change from 0 to 28888 May 13 23:42:07.280034 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. May 13 23:42:07.290159 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... May 13 23:42:07.330532 systemd-udevd[1329]: Using default interface naming scheme 'v255'. May 13 23:42:07.401845 kernel: loop4: detected capacity change from 0 to 103832 May 13 23:42:07.412841 kernel: loop5: detected capacity change from 0 to 194096 May 13 23:42:07.424595 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. May 13 23:42:07.433859 kernel: loop6: detected capacity change from 0 to 126448 May 13 23:42:07.449607 systemd[1]: Starting systemd-networkd.service - Network Configuration... May 13 23:42:07.471831 kernel: loop7: detected capacity change from 0 to 28888 May 13 23:42:07.482003 (sd-merge)[1331]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. May 13 23:42:07.482454 (sd-merge)[1331]: Merged extensions into '/usr'. May 13 23:42:07.488547 systemd[1]: Reload requested from client PID 1305 ('systemd-sysext') (unit systemd-sysext.service)... May 13 23:42:07.488565 systemd[1]: Reloading... May 13 23:42:07.583839 zram_generator::config[1384]: No configuration found. May 13 23:42:07.624837 kernel: mousedev: PS/2 mouse device common for all mice May 13 23:42:07.667854 kernel: hv_vmbus: registering driver hv_balloon May 13 23:42:07.673021 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 May 13 23:42:07.685782 kernel: hv_balloon: Memory hot add disabled on ARM64 May 13 23:42:07.686979 kernel: hv_vmbus: registering driver hyperv_fb May 13 23:42:07.687012 kernel: hyperv_fb: Synthvid Version major 3, minor 5 May 13 23:42:07.699675 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 May 13 23:42:07.711979 kernel: Console: switching to colour dummy device 80x25 May 13 23:42:07.720188 kernel: Console: switching to colour frame buffer device 128x48 May 13 23:42:07.783900 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1340) May 13 23:42:07.835992 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 23:42:07.948780 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. May 13 23:42:07.948905 systemd[1]: Reloading finished in 459 ms. May 13 23:42:07.963723 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. May 13 23:42:07.997736 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. May 13 23:42:08.015757 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. May 13 23:42:08.034101 systemd[1]: Starting ensure-sysext.service... May 13 23:42:08.042865 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... May 13 23:42:08.054597 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... May 13 23:42:08.064338 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... May 13 23:42:08.075956 systemd[1]: Starting systemd-userdbd.service - User Database Manager... May 13 23:42:08.095321 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... May 13 23:42:08.095425 systemd-tmpfiles[1517]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 13 23:42:08.096933 systemd-tmpfiles[1517]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. May 13 23:42:08.099630 systemd-tmpfiles[1517]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 13 23:42:08.099924 systemd-tmpfiles[1517]: ACLs are not supported, ignoring. May 13 23:42:08.099973 systemd-tmpfiles[1517]: ACLs are not supported, ignoring. May 13 23:42:08.115179 lvm[1515]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 13 23:42:08.119507 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. May 13 23:42:08.132849 systemd-tmpfiles[1517]: Detected autofs mount point /boot during canonicalization of boot. May 13 23:42:08.132858 systemd-tmpfiles[1517]: Skipping /boot May 13 23:42:08.134361 systemd[1]: Reload requested from client PID 1514 ('systemctl') (unit ensure-sysext.service)... May 13 23:42:08.134370 systemd[1]: Reloading... May 13 23:42:08.156097 systemd-tmpfiles[1517]: Detected autofs mount point /boot during canonicalization of boot. May 13 23:42:08.156111 systemd-tmpfiles[1517]: Skipping /boot May 13 23:42:08.229844 zram_generator::config[1557]: No configuration found. May 13 23:42:08.313200 systemd-networkd[1353]: lo: Link UP May 13 23:42:08.313209 systemd-networkd[1353]: lo: Gained carrier May 13 23:42:08.316208 systemd-networkd[1353]: Enumeration completed May 13 23:42:08.316630 systemd-networkd[1353]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 13 23:42:08.316706 systemd-networkd[1353]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 13 23:42:08.359304 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 23:42:08.367917 kernel: mlx5_core 4271:00:02.0 enP17009s1: Link up May 13 23:42:08.396891 kernel: hv_netvsc 000d3af9-af51-000d-3af9-af51000d3af9 eth0: Data path switched to VF: enP17009s1 May 13 23:42:08.397372 systemd-networkd[1353]: enP17009s1: Link UP May 13 23:42:08.397462 systemd-networkd[1353]: eth0: Link UP May 13 23:42:08.397466 systemd-networkd[1353]: eth0: Gained carrier May 13 23:42:08.397480 systemd-networkd[1353]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 13 23:42:08.401090 systemd-networkd[1353]: enP17009s1: Gained carrier May 13 23:42:08.406862 systemd-networkd[1353]: eth0: DHCPv4 address 10.200.20.30/24, gateway 10.200.20.1 acquired from 168.63.129.16 May 13 23:42:08.489720 systemd[1]: Reloading finished in 355 ms. May 13 23:42:08.500043 systemd[1]: Started systemd-userdbd.service - User Database Manager. May 13 23:42:08.506945 systemd[1]: Started systemd-networkd.service - Network Configuration. May 13 23:42:08.530625 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. May 13 23:42:08.540200 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. May 13 23:42:08.549488 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. May 13 23:42:08.563489 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. May 13 23:42:08.572569 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 13 23:42:08.590212 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... May 13 23:42:08.609487 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... May 13 23:42:08.621194 lvm[1625]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 13 23:42:08.621626 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... May 13 23:42:08.635982 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... May 13 23:42:08.649156 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... May 13 23:42:08.667030 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... May 13 23:42:08.680732 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... May 13 23:42:08.691145 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. May 13 23:42:08.709162 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 13 23:42:08.715346 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 13 23:42:08.728133 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 13 23:42:08.746043 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 13 23:42:08.758033 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 13 23:42:08.758186 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 13 23:42:08.762248 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. May 13 23:42:08.775828 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. May 13 23:42:08.786462 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 23:42:08.787846 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 13 23:42:08.798168 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 23:42:08.798359 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 13 23:42:08.809091 augenrules[1654]: No rules May 13 23:42:08.809399 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 23:42:08.809772 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 13 23:42:08.817189 systemd[1]: audit-rules.service: Deactivated successfully. May 13 23:42:08.817389 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 13 23:42:08.832146 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 13 23:42:08.833484 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 13 23:42:08.844290 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 13 23:42:08.854620 systemd-resolved[1634]: Positive Trust Anchors: May 13 23:42:08.854641 systemd-resolved[1634]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 13 23:42:08.854672 systemd-resolved[1634]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test May 13 23:42:08.857635 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 13 23:42:08.864391 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 13 23:42:08.864537 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 13 23:42:08.865899 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. May 13 23:42:08.876742 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 23:42:08.876856 systemd-resolved[1634]: Using system hostname 'ci-4284.0.0-n-f756a0c81e'. May 13 23:42:08.877236 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 13 23:42:08.888563 systemd[1]: Started systemd-resolved.service - Network Name Resolution. May 13 23:42:08.896426 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 23:42:08.896623 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 13 23:42:08.905347 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 23:42:08.905508 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 13 23:42:08.920355 systemd[1]: Reached target network.target - Network. May 13 23:42:08.925887 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. May 13 23:42:08.934651 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 13 23:42:08.940790 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. May 13 23:42:08.944951 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... May 13 23:42:08.955994 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... May 13 23:42:08.968798 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... May 13 23:42:08.978249 augenrules[1672]: /sbin/augenrules: No change May 13 23:42:08.979558 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... May 13 23:42:08.989131 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. May 13 23:42:08.989384 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). May 13 23:42:08.989654 systemd[1]: Reached target time-set.target - System Time Set. May 13 23:42:09.000102 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 13 23:42:09.000690 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. May 13 23:42:09.008901 augenrules[1693]: No rules May 13 23:42:09.009658 systemd[1]: modprobe@drm.service: Deactivated successfully. May 13 23:42:09.009856 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. May 13 23:42:09.018395 systemd[1]: audit-rules.service: Deactivated successfully. May 13 23:42:09.018630 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 13 23:42:09.024739 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 13 23:42:09.024927 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. May 13 23:42:09.035743 systemd[1]: modprobe@loop.service: Deactivated successfully. May 13 23:42:09.035994 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. May 13 23:42:09.047483 systemd[1]: Finished ensure-sysext.service. May 13 23:42:09.056564 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 13 23:42:09.056648 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. May 13 23:42:09.335953 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. May 13 23:42:09.345722 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 13 23:42:09.720006 systemd-networkd[1353]: eth0: Gained IPv6LL May 13 23:42:09.725493 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. May 13 23:42:09.735350 systemd[1]: Reached target network-online.target - Network is Online. May 13 23:42:10.167996 systemd-networkd[1353]: enP17009s1: Gained IPv6LL May 13 23:42:11.112847 ldconfig[1300]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 13 23:42:11.141858 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. May 13 23:42:11.150299 systemd[1]: Starting systemd-update-done.service - Update is Completed... May 13 23:42:11.171924 systemd[1]: Finished systemd-update-done.service - Update is Completed. May 13 23:42:11.179705 systemd[1]: Reached target sysinit.target - System Initialization. May 13 23:42:11.186494 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. May 13 23:42:11.194160 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. May 13 23:42:11.206204 systemd[1]: Started logrotate.timer - Daily rotation of log files. May 13 23:42:11.212419 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. May 13 23:42:11.221256 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. May 13 23:42:11.228772 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 13 23:42:11.228802 systemd[1]: Reached target paths.target - Path Units. May 13 23:42:11.234253 systemd[1]: Reached target timers.target - Timer Units. May 13 23:42:11.253475 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. May 13 23:42:11.261856 systemd[1]: Starting docker.socket - Docker Socket for the API... May 13 23:42:11.269344 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). May 13 23:42:11.277639 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). May 13 23:42:11.286679 systemd[1]: Reached target ssh-access.target - SSH Access Available. May 13 23:42:11.295888 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. May 13 23:42:11.302975 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. May 13 23:42:11.310981 systemd[1]: Listening on docker.socket - Docker Socket for the API. May 13 23:42:11.317961 systemd[1]: Reached target sockets.target - Socket Units. May 13 23:42:11.323851 systemd[1]: Reached target basic.target - Basic System. May 13 23:42:11.329618 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. May 13 23:42:11.329645 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. May 13 23:42:11.331949 systemd[1]: Starting chronyd.service - NTP client/server... May 13 23:42:11.346932 systemd[1]: Starting containerd.service - containerd container runtime... May 13 23:42:11.356955 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... May 13 23:42:11.369591 systemd[1]: Starting dbus.service - D-Bus System Message Bus... May 13 23:42:11.379465 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... May 13 23:42:11.386653 (chronyd)[1711]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS May 13 23:42:11.395994 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... May 13 23:42:11.396299 jq[1718]: false May 13 23:42:11.402674 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). May 13 23:42:11.402727 systemd[1]: hv_fcopy_daemon.service - Hyper-V FCOPY daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_fcopy). May 13 23:42:11.404342 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. May 13 23:42:11.412267 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). May 13 23:42:11.414534 chronyd[1723]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) May 13 23:42:11.419722 KVP[1720]: KVP starting; pid is:1720 May 13 23:42:11.420666 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:11.428394 KVP[1720]: KVP LIC Version: 3.1 May 13 23:42:11.430844 kernel: hv_utils: KVP IC version 4.0 May 13 23:42:11.436399 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... May 13 23:42:11.447536 chronyd[1723]: Timezone right/UTC failed leap second check, ignoring May 13 23:42:11.448177 chronyd[1723]: Loaded seccomp filter (level 2) May 13 23:42:11.450566 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... May 13 23:42:11.457326 extend-filesystems[1719]: Found loop4 May 13 23:42:11.457326 extend-filesystems[1719]: Found loop5 May 13 23:42:11.475960 extend-filesystems[1719]: Found loop6 May 13 23:42:11.475960 extend-filesystems[1719]: Found loop7 May 13 23:42:11.475960 extend-filesystems[1719]: Found sda May 13 23:42:11.475960 extend-filesystems[1719]: Found sda1 May 13 23:42:11.475960 extend-filesystems[1719]: Found sda2 May 13 23:42:11.475960 extend-filesystems[1719]: Found sda3 May 13 23:42:11.475960 extend-filesystems[1719]: Found usr May 13 23:42:11.475960 extend-filesystems[1719]: Found sda4 May 13 23:42:11.475960 extend-filesystems[1719]: Found sda6 May 13 23:42:11.475960 extend-filesystems[1719]: Found sda7 May 13 23:42:11.475960 extend-filesystems[1719]: Found sda9 May 13 23:42:11.475960 extend-filesystems[1719]: Checking size of /dev/sda9 May 13 23:42:11.461927 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... May 13 23:42:11.712536 extend-filesystems[1719]: Old size kept for /dev/sda9 May 13 23:42:11.712536 extend-filesystems[1719]: Found sr0 May 13 23:42:11.522911 dbus-daemon[1714]: [system] SELinux support is enabled May 13 23:42:11.745040 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (1764) May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.659 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.685 INFO Fetch successful May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.685 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.691 INFO Fetch successful May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.692 INFO Fetching http://168.63.129.16/machine/8f5ac59c-52d4-45e1-8dff-07f586552418/aeee2db6%2D5a64%2D4521%2Dbbc8%2Dfa3d9ca4f4c9.%5Fci%2D4284.0.0%2Dn%2Df756a0c81e?comp=config&type=sharedConfig&incarnation=1: Attempt #1 May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.694 INFO Fetch successful May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.694 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 May 13 23:42:11.745070 coreos-metadata[1713]: May 13 23:42:11.712 INFO Fetch successful May 13 23:42:11.480873 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... May 13 23:42:11.507802 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... May 13 23:42:11.537374 systemd[1]: Starting systemd-logind.service - User Login Management... May 13 23:42:11.560035 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). May 13 23:42:11.560668 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. May 13 23:42:11.562202 systemd[1]: Starting update-engine.service - Update Engine... May 13 23:42:11.755098 jq[1760]: true May 13 23:42:11.575007 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... May 13 23:42:11.755332 update_engine[1751]: I20250513 23:42:11.702569 1751 main.cc:92] Flatcar Update Engine starting May 13 23:42:11.755332 update_engine[1751]: I20250513 23:42:11.709156 1751 update_check_scheduler.cc:74] Next update check in 2m54s May 13 23:42:11.601632 systemd[1]: Started dbus.service - D-Bus System Message Bus. May 13 23:42:11.631206 systemd[1]: Started chronyd.service - NTP client/server. May 13 23:42:11.638310 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 13 23:42:11.638502 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. May 13 23:42:11.638737 systemd[1]: extend-filesystems.service: Deactivated successfully. May 13 23:42:11.764768 jq[1796]: true May 13 23:42:11.638930 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. May 13 23:42:11.663768 systemd[1]: motdgen.service: Deactivated successfully. May 13 23:42:11.663966 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. May 13 23:42:11.676585 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. May 13 23:42:11.698856 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 13 23:42:11.699070 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. May 13 23:42:11.740527 (ntainerd)[1799]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR May 13 23:42:11.787166 systemd[1]: Started update-engine.service - Update Engine. May 13 23:42:11.798852 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 13 23:42:11.798894 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. May 13 23:42:11.813626 systemd-logind[1744]: New seat seat0. May 13 23:42:11.825345 systemd-logind[1744]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 13 23:42:11.833639 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 13 23:42:11.833666 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. May 13 23:42:11.850636 systemd[1]: Started locksmithd.service - Cluster reboot manager. May 13 23:42:11.859225 systemd[1]: Started systemd-logind.service - User Login Management. May 13 23:42:11.868148 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. May 13 23:42:11.886423 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. May 13 23:42:11.888225 tar[1795]: linux-arm64/helm May 13 23:42:11.925293 bash[1848]: Updated "/home/core/.ssh/authorized_keys" May 13 23:42:11.926267 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. May 13 23:42:11.949348 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. May 13 23:42:12.158545 locksmithd[1849]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 13 23:42:12.347102 tar[1795]: linux-arm64/LICENSE May 13 23:42:12.347102 tar[1795]: linux-arm64/README.md May 13 23:42:12.352687 sshd_keygen[1752]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 13 23:42:12.364550 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. May 13 23:42:12.384658 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. May 13 23:42:12.397276 systemd[1]: Starting issuegen.service - Generate /run/issue... May 13 23:42:12.419195 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... May 13 23:42:12.429257 systemd[1]: issuegen.service: Deactivated successfully. May 13 23:42:12.429491 systemd[1]: Finished issuegen.service - Generate /run/issue. May 13 23:42:12.441133 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... May 13 23:42:12.462224 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. May 13 23:42:12.476132 systemd[1]: Started getty@tty1.service - Getty on tty1. May 13 23:42:12.484329 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. May 13 23:42:12.504946 containerd[1799]: time="2025-05-13T23:42:12Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 May 13 23:42:12.498243 systemd[1]: Reached target getty.target - Login Prompts. May 13 23:42:12.506746 containerd[1799]: time="2025-05-13T23:42:12.506707120Z" level=info msg="starting containerd" revision=88aa2f531d6c2922003cc7929e51daf1c14caa0a version=v2.0.1 May 13 23:42:12.512993 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. May 13 23:42:12.524736 containerd[1799]: time="2025-05-13T23:42:12.524693080Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="7.36µs" May 13 23:42:12.524901 containerd[1799]: time="2025-05-13T23:42:12.524882760Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 May 13 23:42:12.524961 containerd[1799]: time="2025-05-13T23:42:12.524948360Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 May 13 23:42:12.525185 containerd[1799]: time="2025-05-13T23:42:12.525164360Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 May 13 23:42:12.525256 containerd[1799]: time="2025-05-13T23:42:12.525243320Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 May 13 23:42:12.525324 containerd[1799]: time="2025-05-13T23:42:12.525311640Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 13 23:42:12.525452 containerd[1799]: time="2025-05-13T23:42:12.525434400Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 May 13 23:42:12.525512 containerd[1799]: time="2025-05-13T23:42:12.525498680Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 13 23:42:12.525871 containerd[1799]: time="2025-05-13T23:42:12.525806720Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 May 13 23:42:12.526062 containerd[1799]: time="2025-05-13T23:42:12.526041920Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 13 23:42:12.526207 containerd[1799]: time="2025-05-13T23:42:12.526189200Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 May 13 23:42:12.526523 containerd[1799]: time="2025-05-13T23:42:12.526384560Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 May 13 23:42:12.527143 containerd[1799]: time="2025-05-13T23:42:12.526946840Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 May 13 23:42:12.527582 containerd[1799]: time="2025-05-13T23:42:12.527557960Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 13 23:42:12.527742 containerd[1799]: time="2025-05-13T23:42:12.527723800Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 May 13 23:42:12.527812 containerd[1799]: time="2025-05-13T23:42:12.527798480Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 May 13 23:42:12.527933 containerd[1799]: time="2025-05-13T23:42:12.527918880Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 May 13 23:42:12.528304 containerd[1799]: time="2025-05-13T23:42:12.528276640Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 May 13 23:42:12.528433 containerd[1799]: time="2025-05-13T23:42:12.528417160Z" level=info msg="metadata content store policy set" policy=shared May 13 23:42:12.555053 containerd[1799]: time="2025-05-13T23:42:12.555007840Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 May 13 23:42:12.555220 containerd[1799]: time="2025-05-13T23:42:12.555205160Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 May 13 23:42:12.555416 containerd[1799]: time="2025-05-13T23:42:12.555397760Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 May 13 23:42:12.555658 containerd[1799]: time="2025-05-13T23:42:12.555517840Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 May 13 23:42:12.556775 containerd[1799]: time="2025-05-13T23:42:12.556731400Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 May 13 23:42:12.556775 containerd[1799]: time="2025-05-13T23:42:12.556768120Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 May 13 23:42:12.556775 containerd[1799]: time="2025-05-13T23:42:12.556783840Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.556798040Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.556812320Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.556837840Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.556848800Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.556861360Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.556998680Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557046200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557063320Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557073960Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557084840Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557095600Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557107360Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557119040Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557131200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 May 13 23:42:12.557354 containerd[1799]: time="2025-05-13T23:42:12.557143200Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 May 13 23:42:12.557840 containerd[1799]: time="2025-05-13T23:42:12.557154760Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 May 13 23:42:12.557840 containerd[1799]: time="2025-05-13T23:42:12.557232280Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" May 13 23:42:12.557840 containerd[1799]: time="2025-05-13T23:42:12.557248280Z" level=info msg="Start snapshots syncer" May 13 23:42:12.557840 containerd[1799]: time="2025-05-13T23:42:12.557274840Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557490680Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557538400Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557614880Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557714440Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557736480Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557747480Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557757720Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557770080Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557784640Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557795160Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557846200Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557862560Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557872480Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557916000Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557932240Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557941040Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557950200Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557958240Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557969120Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557979640Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.557997080Z" level=info msg="runtime interface created" May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.558002240Z" level=info msg="created NRI interface" May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.558010800Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 May 13 23:42:12.557955 containerd[1799]: time="2025-05-13T23:42:12.558022120Z" level=info msg="Connect containerd service" May 13 23:42:12.558872 containerd[1799]: time="2025-05-13T23:42:12.558049880Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" May 13 23:42:12.558872 containerd[1799]: time="2025-05-13T23:42:12.558668280Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 13 23:42:12.721170 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:12.792426 (kubelet)[1903]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.237621320Z" level=info msg="Start subscribing containerd event" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.237700880Z" level=info msg="Start recovering state" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.237925840Z" level=info msg="Start event monitor" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238037600Z" level=info msg="Start cni network conf syncer for default" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238047160Z" level=info msg="Start streaming server" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238057080Z" level=info msg="Registered namespace \"k8s.io\" with NRI" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238064320Z" level=info msg="runtime interface starting up..." May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238070520Z" level=info msg="starting plugins..." May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238085960Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238268520Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 13 23:42:13.239219 containerd[1799]: time="2025-05-13T23:42:13.238315400Z" level=info msg=serving... address=/run/containerd/containerd.sock May 13 23:42:13.239567 containerd[1799]: time="2025-05-13T23:42:13.239262920Z" level=info msg="containerd successfully booted in 0.742356s" May 13 23:42:13.239693 systemd[1]: Started containerd.service - containerd container runtime. May 13 23:42:13.249579 systemd[1]: Reached target multi-user.target - Multi-User System. May 13 23:42:13.257904 systemd[1]: Startup finished in 661ms (kernel) + 12.350s (initrd) + 10.780s (userspace) = 23.793s. May 13 23:42:13.281750 kubelet[1903]: E0513 23:42:13.281667 1903 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 23:42:13.284371 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 23:42:13.284506 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 23:42:13.284951 systemd[1]: kubelet.service: Consumed 707ms CPU time, 240.2M memory peak. May 13 23:42:13.487314 login[1888]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:13.488907 login[1889]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:13.497556 systemd[1]: Created slice user-500.slice - User Slice of UID 500. May 13 23:42:13.499055 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... May 13 23:42:13.505066 systemd-logind[1744]: New session 2 of user core. May 13 23:42:13.510926 systemd-logind[1744]: New session 1 of user core. May 13 23:42:13.519630 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. May 13 23:42:13.522518 systemd[1]: Starting user@500.service - User Manager for UID 500... May 13 23:42:13.546990 (systemd)[1926]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 13 23:42:13.549881 systemd-logind[1744]: New session c1 of user core. May 13 23:42:13.701594 systemd[1926]: Queued start job for default target default.target. May 13 23:42:13.711745 systemd[1926]: Created slice app.slice - User Application Slice. May 13 23:42:13.711924 systemd[1926]: Reached target paths.target - Paths. May 13 23:42:13.712035 systemd[1926]: Reached target timers.target - Timers. May 13 23:42:13.714974 systemd[1926]: Starting dbus.socket - D-Bus User Message Bus Socket... May 13 23:42:13.729929 systemd[1926]: Listening on dbus.socket - D-Bus User Message Bus Socket. May 13 23:42:13.730041 systemd[1926]: Reached target sockets.target - Sockets. May 13 23:42:13.730085 systemd[1926]: Reached target basic.target - Basic System. May 13 23:42:13.730115 systemd[1926]: Reached target default.target - Main User Target. May 13 23:42:13.730139 systemd[1926]: Startup finished in 173ms. May 13 23:42:13.730345 systemd[1]: Started user@500.service - User Manager for UID 500. May 13 23:42:13.740984 systemd[1]: Started session-1.scope - Session 1 of User core. May 13 23:42:13.741735 systemd[1]: Started session-2.scope - Session 2 of User core. May 13 23:42:13.915413 waagent[1892]: 2025-05-13T23:42:13.915280Z INFO Daemon Daemon Azure Linux Agent Version: 2.12.0.4 May 13 23:42:13.922261 waagent[1892]: 2025-05-13T23:42:13.922211Z INFO Daemon Daemon OS: flatcar 4284.0.0 May 13 23:42:13.927540 waagent[1892]: 2025-05-13T23:42:13.927499Z INFO Daemon Daemon Python: 3.11.11 May 13 23:42:13.933934 waagent[1892]: 2025-05-13T23:42:13.933883Z INFO Daemon Daemon Run daemon May 13 23:42:13.938919 waagent[1892]: 2025-05-13T23:42:13.938879Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4284.0.0' May 13 23:42:13.949842 waagent[1892]: 2025-05-13T23:42:13.948778Z INFO Daemon Daemon Using waagent for provisioning May 13 23:42:13.954392 waagent[1892]: 2025-05-13T23:42:13.954347Z INFO Daemon Daemon Activate resource disk May 13 23:42:13.959939 waagent[1892]: 2025-05-13T23:42:13.959899Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb May 13 23:42:13.972705 waagent[1892]: 2025-05-13T23:42:13.972660Z INFO Daemon Daemon Found device: None May 13 23:42:13.977860 waagent[1892]: 2025-05-13T23:42:13.977823Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology May 13 23:42:13.986916 waagent[1892]: 2025-05-13T23:42:13.986880Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 May 13 23:42:13.999024 waagent[1892]: 2025-05-13T23:42:13.998985Z INFO Daemon Daemon Clean protocol and wireserver endpoint May 13 23:42:14.006047 waagent[1892]: 2025-05-13T23:42:14.006012Z INFO Daemon Daemon Running default provisioning handler May 13 23:42:14.018041 waagent[1892]: 2025-05-13T23:42:14.017971Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. May 13 23:42:14.034244 waagent[1892]: 2025-05-13T23:42:14.034186Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' May 13 23:42:14.044940 waagent[1892]: 2025-05-13T23:42:14.044895Z INFO Daemon Daemon cloud-init is enabled: False May 13 23:42:14.051161 waagent[1892]: 2025-05-13T23:42:14.051118Z INFO Daemon Daemon Copying ovf-env.xml May 13 23:42:14.146004 waagent[1892]: 2025-05-13T23:42:14.145923Z INFO Daemon Daemon Successfully mounted dvd May 13 23:42:14.173015 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. May 13 23:42:14.174447 waagent[1892]: 2025-05-13T23:42:14.174377Z INFO Daemon Daemon Detect protocol endpoint May 13 23:42:14.180492 waagent[1892]: 2025-05-13T23:42:14.180444Z INFO Daemon Daemon Clean protocol and wireserver endpoint May 13 23:42:14.187476 waagent[1892]: 2025-05-13T23:42:14.187428Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler May 13 23:42:14.195886 waagent[1892]: 2025-05-13T23:42:14.195846Z INFO Daemon Daemon Test for route to 168.63.129.16 May 13 23:42:14.201894 waagent[1892]: 2025-05-13T23:42:14.201855Z INFO Daemon Daemon Route to 168.63.129.16 exists May 13 23:42:14.208123 waagent[1892]: 2025-05-13T23:42:14.208085Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 May 13 23:42:14.249392 waagent[1892]: 2025-05-13T23:42:14.249337Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 May 13 23:42:14.257264 waagent[1892]: 2025-05-13T23:42:14.257234Z INFO Daemon Daemon Wire protocol version:2012-11-30 May 13 23:42:14.263588 waagent[1892]: 2025-05-13T23:42:14.263549Z INFO Daemon Daemon Server preferred version:2015-04-05 May 13 23:42:14.524262 waagent[1892]: 2025-05-13T23:42:14.524128Z INFO Daemon Daemon Initializing goal state during protocol detection May 13 23:42:14.531111 waagent[1892]: 2025-05-13T23:42:14.531059Z INFO Daemon Daemon Forcing an update of the goal state. May 13 23:42:14.540525 waagent[1892]: 2025-05-13T23:42:14.540480Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] May 13 23:42:14.563068 waagent[1892]: 2025-05-13T23:42:14.563032Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.164 May 13 23:42:14.569413 waagent[1892]: 2025-05-13T23:42:14.569371Z INFO Daemon May 13 23:42:14.572529 waagent[1892]: 2025-05-13T23:42:14.572492Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: 40142007-df45-404d-af01-ffdb79a73e15 eTag: 13326329972855851017 source: Fabric] May 13 23:42:14.585135 waagent[1892]: 2025-05-13T23:42:14.585094Z INFO Daemon The vmSettings originated via Fabric; will ignore them. May 13 23:42:14.595020 waagent[1892]: 2025-05-13T23:42:14.594982Z INFO Daemon May 13 23:42:14.599564 waagent[1892]: 2025-05-13T23:42:14.599522Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] May 13 23:42:14.613660 waagent[1892]: 2025-05-13T23:42:14.613623Z INFO Daemon Daemon Downloading artifacts profile blob May 13 23:42:14.799061 waagent[1892]: 2025-05-13T23:42:14.798940Z INFO Daemon Downloaded certificate {'thumbprint': '871AEDEB83C93DA910E10E5E21B32D21A6B8BB07', 'hasPrivateKey': False} May 13 23:42:14.811031 waagent[1892]: 2025-05-13T23:42:14.810986Z INFO Daemon Downloaded certificate {'thumbprint': '0B578C39987D18560EC99028CCCD5492DCF1CA03', 'hasPrivateKey': True} May 13 23:42:14.822312 waagent[1892]: 2025-05-13T23:42:14.822270Z INFO Daemon Fetch goal state completed May 13 23:42:14.891962 waagent[1892]: 2025-05-13T23:42:14.891918Z INFO Daemon Daemon Starting provisioning May 13 23:42:14.897781 waagent[1892]: 2025-05-13T23:42:14.897731Z INFO Daemon Daemon Handle ovf-env.xml. May 13 23:42:14.903298 waagent[1892]: 2025-05-13T23:42:14.903257Z INFO Daemon Daemon Set hostname [ci-4284.0.0-n-f756a0c81e] May 13 23:42:14.945846 waagent[1892]: 2025-05-13T23:42:14.945269Z INFO Daemon Daemon Publish hostname [ci-4284.0.0-n-f756a0c81e] May 13 23:42:14.952993 waagent[1892]: 2025-05-13T23:42:14.952940Z INFO Daemon Daemon Examine /proc/net/route for primary interface May 13 23:42:14.960900 waagent[1892]: 2025-05-13T23:42:14.960859Z INFO Daemon Daemon Primary interface is [eth0] May 13 23:42:14.974330 systemd-networkd[1353]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. May 13 23:42:14.974344 systemd-networkd[1353]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 13 23:42:14.974373 systemd-networkd[1353]: eth0: DHCP lease lost May 13 23:42:14.975368 waagent[1892]: 2025-05-13T23:42:14.975293Z INFO Daemon Daemon Create user account if not exists May 13 23:42:14.982455 waagent[1892]: 2025-05-13T23:42:14.982403Z INFO Daemon Daemon User core already exists, skip useradd May 13 23:42:14.989307 waagent[1892]: 2025-05-13T23:42:14.989249Z INFO Daemon Daemon Configure sudoer May 13 23:42:14.994751 waagent[1892]: 2025-05-13T23:42:14.994701Z INFO Daemon Daemon Configure sshd May 13 23:42:15.000086 waagent[1892]: 2025-05-13T23:42:15.000042Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. May 13 23:42:15.015246 waagent[1892]: 2025-05-13T23:42:15.015195Z INFO Daemon Daemon Deploy ssh public key. May 13 23:42:15.030914 systemd-networkd[1353]: eth0: DHCPv4 address 10.200.20.30/24, gateway 10.200.20.1 acquired from 168.63.129.16 May 13 23:42:16.164192 waagent[1892]: 2025-05-13T23:42:16.164145Z INFO Daemon Daemon Provisioning complete May 13 23:42:16.183743 waagent[1892]: 2025-05-13T23:42:16.183703Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping May 13 23:42:16.190690 waagent[1892]: 2025-05-13T23:42:16.190642Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. May 13 23:42:16.202168 waagent[1892]: 2025-05-13T23:42:16.202122Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.12.0.4 is the most current agent May 13 23:42:16.338287 waagent[1983]: 2025-05-13T23:42:16.338040Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.12.0.4) May 13 23:42:16.338287 waagent[1983]: 2025-05-13T23:42:16.338178Z INFO ExtHandler ExtHandler OS: flatcar 4284.0.0 May 13 23:42:16.339057 waagent[1983]: 2025-05-13T23:42:16.338236Z INFO ExtHandler ExtHandler Python: 3.11.11 May 13 23:42:16.339057 waagent[1983]: 2025-05-13T23:42:16.338545Z INFO ExtHandler ExtHandler CPU Arch: aarch64 May 13 23:42:16.380448 waagent[1983]: 2025-05-13T23:42:16.380381Z INFO ExtHandler ExtHandler Distro: flatcar-4284.0.0; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.11; Arch: aarch64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; May 13 23:42:16.381837 waagent[1983]: 2025-05-13T23:42:16.380763Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file May 13 23:42:16.381837 waagent[1983]: 2025-05-13T23:42:16.380855Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 May 13 23:42:16.389359 waagent[1983]: 2025-05-13T23:42:16.389308Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] May 13 23:42:16.396181 waagent[1983]: 2025-05-13T23:42:16.396143Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.164 May 13 23:42:16.396851 waagent[1983]: 2025-05-13T23:42:16.396786Z INFO ExtHandler May 13 23:42:16.397000 waagent[1983]: 2025-05-13T23:42:16.396973Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 8ef778da-4103-4521-8e88-22c4c1897604 eTag: 13326329972855851017 source: Fabric] May 13 23:42:16.397373 waagent[1983]: 2025-05-13T23:42:16.397343Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. May 13 23:42:16.398005 waagent[1983]: 2025-05-13T23:42:16.397973Z INFO ExtHandler May 13 23:42:16.398144 waagent[1983]: 2025-05-13T23:42:16.398120Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] May 13 23:42:16.402877 waagent[1983]: 2025-05-13T23:42:16.402848Z INFO ExtHandler ExtHandler Downloading artifacts profile blob May 13 23:42:16.488713 waagent[1983]: 2025-05-13T23:42:16.488647Z INFO ExtHandler Downloaded certificate {'thumbprint': '871AEDEB83C93DA910E10E5E21B32D21A6B8BB07', 'hasPrivateKey': False} May 13 23:42:16.489299 waagent[1983]: 2025-05-13T23:42:16.489268Z INFO ExtHandler Downloaded certificate {'thumbprint': '0B578C39987D18560EC99028CCCD5492DCF1CA03', 'hasPrivateKey': True} May 13 23:42:16.489796 waagent[1983]: 2025-05-13T23:42:16.489764Z INFO ExtHandler Fetch goal state completed May 13 23:42:16.505651 waagent[1983]: 2025-05-13T23:42:16.505604Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.3.3 11 Feb 2025 (Library: OpenSSL 3.3.3 11 Feb 2025) May 13 23:42:16.510309 waagent[1983]: 2025-05-13T23:42:16.510262Z INFO ExtHandler ExtHandler WALinuxAgent-2.12.0.4 running as process 1983 May 13 23:42:16.511834 waagent[1983]: 2025-05-13T23:42:16.510529Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** May 13 23:42:16.511834 waagent[1983]: 2025-05-13T23:42:16.510862Z INFO ExtHandler ExtHandler ******** AutoUpdate.UpdateToLatestVersion is set to False, not processing the operation ******** May 13 23:42:16.512491 waagent[1983]: 2025-05-13T23:42:16.512457Z INFO ExtHandler ExtHandler [CGI] Cgroup monitoring is not supported on ['flatcar', '4284.0.0', '', 'Flatcar Container Linux by Kinvolk'] May 13 23:42:16.512984 waagent[1983]: 2025-05-13T23:42:16.512952Z INFO ExtHandler ExtHandler [CGI] Agent will reset the quotas in case distro: ['flatcar', '4284.0.0', '', 'Flatcar Container Linux by Kinvolk'] went from supported to unsupported May 13 23:42:16.513231 waagent[1983]: 2025-05-13T23:42:16.513201Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: False May 13 23:42:16.513919 waagent[1983]: 2025-05-13T23:42:16.513883Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules May 13 23:42:16.540560 waagent[1983]: 2025-05-13T23:42:16.540526Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service May 13 23:42:16.540892 waagent[1983]: 2025-05-13T23:42:16.540860Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup May 13 23:42:16.546963 waagent[1983]: 2025-05-13T23:42:16.546935Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now May 13 23:42:16.552947 systemd[1]: Reload requested from client PID 2000 ('systemctl') (unit waagent.service)... May 13 23:42:16.552962 systemd[1]: Reloading... May 13 23:42:16.645856 zram_generator::config[2036]: No configuration found. May 13 23:42:16.747517 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 23:42:16.848798 systemd[1]: Reloading finished in 295 ms. May 13 23:42:16.863204 waagent[1983]: 2025-05-13T23:42:16.862418Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service May 13 23:42:16.863204 waagent[1983]: 2025-05-13T23:42:16.862567Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully May 13 23:42:17.916531 waagent[1983]: 2025-05-13T23:42:17.916470Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. May 13 23:42:17.917280 waagent[1983]: 2025-05-13T23:42:17.917244Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: 1. configuration enabled [True], 2. cgroups v1 enabled [False] OR cgroups v2 is in use and v2 resource limiting configuration enabled [False], 3. python supported: [True] May 13 23:42:17.918095 waagent[1983]: 2025-05-13T23:42:17.918050Z INFO ExtHandler ExtHandler Starting env monitor service. May 13 23:42:17.918201 waagent[1983]: 2025-05-13T23:42:17.918164Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file May 13 23:42:17.918388 waagent[1983]: 2025-05-13T23:42:17.918289Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 May 13 23:42:17.918636 waagent[1983]: 2025-05-13T23:42:17.918593Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. May 13 23:42:17.919108 waagent[1983]: 2025-05-13T23:42:17.919062Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: May 13 23:42:17.919108 waagent[1983]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT May 13 23:42:17.919108 waagent[1983]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 May 13 23:42:17.919108 waagent[1983]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 May 13 23:42:17.919108 waagent[1983]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 May 13 23:42:17.919108 waagent[1983]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 May 13 23:42:17.919108 waagent[1983]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 May 13 23:42:17.919581 waagent[1983]: 2025-05-13T23:42:17.919515Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. May 13 23:42:17.919762 waagent[1983]: 2025-05-13T23:42:17.919629Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file May 13 23:42:17.919841 waagent[1983]: 2025-05-13T23:42:17.919784Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 May 13 23:42:17.920002 waagent[1983]: 2025-05-13T23:42:17.919960Z INFO EnvHandler ExtHandler Configure routes May 13 23:42:17.920229 waagent[1983]: 2025-05-13T23:42:17.920048Z INFO EnvHandler ExtHandler Gateway:None May 13 23:42:17.920370 waagent[1983]: 2025-05-13T23:42:17.920278Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread May 13 23:42:17.920524 waagent[1983]: 2025-05-13T23:42:17.920477Z INFO ExtHandler ExtHandler Start Extension Telemetry service. May 13 23:42:17.920567 waagent[1983]: 2025-05-13T23:42:17.920542Z INFO EnvHandler ExtHandler Routes:None May 13 23:42:17.921286 waagent[1983]: 2025-05-13T23:42:17.921212Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True May 13 23:42:17.921405 waagent[1983]: 2025-05-13T23:42:17.921355Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. May 13 23:42:17.921671 waagent[1983]: 2025-05-13T23:42:17.921629Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread May 13 23:42:17.928074 waagent[1983]: 2025-05-13T23:42:17.928027Z INFO ExtHandler ExtHandler May 13 23:42:17.928599 waagent[1983]: 2025-05-13T23:42:17.928558Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 2e4a6ec2-dc00-4c94-92fd-e7c1e1492cfe correlation 25b04704-d7ad-40cd-9523-c10596aad645 created: 2025-05-13T23:41:07.456253Z] May 13 23:42:17.929953 waagent[1983]: 2025-05-13T23:42:17.929912Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. May 13 23:42:17.930887 waagent[1983]: 2025-05-13T23:42:17.930855Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 2 ms] May 13 23:42:17.969618 waagent[1983]: 2025-05-13T23:42:17.969500Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.12.0.4 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 825FA904-4EFE-48B4-B8B8-BAEEA461E243;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0;UpdateMode: SelfUpdate;] May 13 23:42:17.983656 waagent[1983]: 2025-05-13T23:42:17.983592Z INFO MonitorHandler ExtHandler Network interfaces: May 13 23:42:17.983656 waagent[1983]: Executing ['ip', '-a', '-o', 'link']: May 13 23:42:17.983656 waagent[1983]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 May 13 23:42:17.983656 waagent[1983]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:f9:af:51 brd ff:ff:ff:ff:ff:ff May 13 23:42:17.983656 waagent[1983]: 3: enP17009s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:f9:af:51 brd ff:ff:ff:ff:ff:ff\ altname enP17009p0s2 May 13 23:42:17.983656 waagent[1983]: Executing ['ip', '-4', '-a', '-o', 'address']: May 13 23:42:17.983656 waagent[1983]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever May 13 23:42:17.983656 waagent[1983]: 2: eth0 inet 10.200.20.30/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever May 13 23:42:17.983656 waagent[1983]: Executing ['ip', '-6', '-a', '-o', 'address']: May 13 23:42:17.983656 waagent[1983]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever May 13 23:42:17.983656 waagent[1983]: 2: eth0 inet6 fe80::20d:3aff:fef9:af51/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever May 13 23:42:17.983656 waagent[1983]: 3: enP17009s1 inet6 fe80::20d:3aff:fef9:af51/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever May 13 23:42:18.024078 waagent[1983]: 2025-05-13T23:42:18.024009Z INFO EnvHandler ExtHandler Created firewall rules for the Azure Fabric: May 13 23:42:18.024078 waagent[1983]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) May 13 23:42:18.024078 waagent[1983]: pkts bytes target prot opt in out source destination May 13 23:42:18.024078 waagent[1983]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) May 13 23:42:18.024078 waagent[1983]: pkts bytes target prot opt in out source destination May 13 23:42:18.024078 waagent[1983]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) May 13 23:42:18.024078 waagent[1983]: pkts bytes target prot opt in out source destination May 13 23:42:18.024078 waagent[1983]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 May 13 23:42:18.024078 waagent[1983]: 4 416 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 May 13 23:42:18.024078 waagent[1983]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW May 13 23:42:18.028566 waagent[1983]: 2025-05-13T23:42:18.028224Z INFO EnvHandler ExtHandler Current Firewall rules: May 13 23:42:18.028566 waagent[1983]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) May 13 23:42:18.028566 waagent[1983]: pkts bytes target prot opt in out source destination May 13 23:42:18.028566 waagent[1983]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) May 13 23:42:18.028566 waagent[1983]: pkts bytes target prot opt in out source destination May 13 23:42:18.028566 waagent[1983]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) May 13 23:42:18.028566 waagent[1983]: pkts bytes target prot opt in out source destination May 13 23:42:18.028566 waagent[1983]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 May 13 23:42:18.028566 waagent[1983]: 14 1463 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 May 13 23:42:18.028566 waagent[1983]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW May 13 23:42:18.028566 waagent[1983]: 2025-05-13T23:42:18.028467Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 May 13 23:42:23.418775 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 13 23:42:23.420550 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:23.524162 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:23.527769 (kubelet)[2136]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 13 23:42:23.642100 kubelet[2136]: E0513 23:42:23.642028 2136 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 23:42:23.645226 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 23:42:23.645377 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 23:42:23.645659 systemd[1]: kubelet.service: Consumed 135ms CPU time, 98.4M memory peak. May 13 23:42:24.158483 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. May 13 23:42:24.159586 systemd[1]: Started sshd@0-10.200.20.30:22-10.200.16.10:54842.service - OpenSSH per-connection server daemon (10.200.16.10:54842). May 13 23:42:24.709652 sshd[2145]: Accepted publickey for core from 10.200.16.10 port 54842 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:24.710936 sshd-session[2145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:24.714950 systemd-logind[1744]: New session 3 of user core. May 13 23:42:24.725951 systemd[1]: Started session-3.scope - Session 3 of User core. May 13 23:42:25.133279 systemd[1]: Started sshd@1-10.200.20.30:22-10.200.16.10:54854.service - OpenSSH per-connection server daemon (10.200.16.10:54854). May 13 23:42:25.593387 sshd[2150]: Accepted publickey for core from 10.200.16.10 port 54854 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:25.594595 sshd-session[2150]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:25.599999 systemd-logind[1744]: New session 4 of user core. May 13 23:42:25.606996 systemd[1]: Started session-4.scope - Session 4 of User core. May 13 23:42:25.935738 sshd[2152]: Connection closed by 10.200.16.10 port 54854 May 13 23:42:25.936284 sshd-session[2150]: pam_unix(sshd:session): session closed for user core May 13 23:42:25.939545 systemd[1]: sshd@1-10.200.20.30:22-10.200.16.10:54854.service: Deactivated successfully. May 13 23:42:25.941052 systemd[1]: session-4.scope: Deactivated successfully. May 13 23:42:25.941764 systemd-logind[1744]: Session 4 logged out. Waiting for processes to exit. May 13 23:42:25.942745 systemd-logind[1744]: Removed session 4. May 13 23:42:26.029275 systemd[1]: Started sshd@2-10.200.20.30:22-10.200.16.10:54870.service - OpenSSH per-connection server daemon (10.200.16.10:54870). May 13 23:42:26.522483 sshd[2158]: Accepted publickey for core from 10.200.16.10 port 54870 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:26.523706 sshd-session[2158]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:26.529007 systemd-logind[1744]: New session 5 of user core. May 13 23:42:26.535962 systemd[1]: Started session-5.scope - Session 5 of User core. May 13 23:42:26.880934 sshd[2160]: Connection closed by 10.200.16.10 port 54870 May 13 23:42:26.881484 sshd-session[2158]: pam_unix(sshd:session): session closed for user core May 13 23:42:26.884766 systemd[1]: sshd@2-10.200.20.30:22-10.200.16.10:54870.service: Deactivated successfully. May 13 23:42:26.886279 systemd[1]: session-5.scope: Deactivated successfully. May 13 23:42:26.887003 systemd-logind[1744]: Session 5 logged out. Waiting for processes to exit. May 13 23:42:26.887927 systemd-logind[1744]: Removed session 5. May 13 23:42:26.963918 systemd[1]: Started sshd@3-10.200.20.30:22-10.200.16.10:54882.service - OpenSSH per-connection server daemon (10.200.16.10:54882). May 13 23:42:27.424544 sshd[2166]: Accepted publickey for core from 10.200.16.10 port 54882 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:27.425769 sshd-session[2166]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:27.430874 systemd-logind[1744]: New session 6 of user core. May 13 23:42:27.435952 systemd[1]: Started session-6.scope - Session 6 of User core. May 13 23:42:27.750343 sshd[2168]: Connection closed by 10.200.16.10 port 54882 May 13 23:42:27.750879 sshd-session[2166]: pam_unix(sshd:session): session closed for user core May 13 23:42:27.754136 systemd[1]: sshd@3-10.200.20.30:22-10.200.16.10:54882.service: Deactivated successfully. May 13 23:42:27.756522 systemd[1]: session-6.scope: Deactivated successfully. May 13 23:42:27.757620 systemd-logind[1744]: Session 6 logged out. Waiting for processes to exit. May 13 23:42:27.758393 systemd-logind[1744]: Removed session 6. May 13 23:42:27.838574 systemd[1]: Started sshd@4-10.200.20.30:22-10.200.16.10:54888.service - OpenSSH per-connection server daemon (10.200.16.10:54888). May 13 23:42:28.327756 sshd[2174]: Accepted publickey for core from 10.200.16.10 port 54888 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:28.331166 sshd-session[2174]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:28.335995 systemd-logind[1744]: New session 7 of user core. May 13 23:42:28.346010 systemd[1]: Started session-7.scope - Session 7 of User core. May 13 23:42:28.667737 sudo[2177]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 May 13 23:42:28.668439 sudo[2177]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 13 23:42:28.692874 sudo[2177]: pam_unix(sudo:session): session closed for user root May 13 23:42:28.784197 sshd[2176]: Connection closed by 10.200.16.10 port 54888 May 13 23:42:28.784924 sshd-session[2174]: pam_unix(sshd:session): session closed for user core May 13 23:42:28.788486 systemd-logind[1744]: Session 7 logged out. Waiting for processes to exit. May 13 23:42:28.789324 systemd[1]: sshd@4-10.200.20.30:22-10.200.16.10:54888.service: Deactivated successfully. May 13 23:42:28.791130 systemd[1]: session-7.scope: Deactivated successfully. May 13 23:42:28.792333 systemd-logind[1744]: Removed session 7. May 13 23:42:28.874259 systemd[1]: Started sshd@5-10.200.20.30:22-10.200.16.10:50916.service - OpenSSH per-connection server daemon (10.200.16.10:50916). May 13 23:42:29.361674 sshd[2183]: Accepted publickey for core from 10.200.16.10 port 50916 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:29.363012 sshd-session[2183]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:29.368452 systemd-logind[1744]: New session 8 of user core. May 13 23:42:29.374032 systemd[1]: Started session-8.scope - Session 8 of User core. May 13 23:42:29.632470 sudo[2187]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules May 13 23:42:29.633133 sudo[2187]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 13 23:42:29.636337 sudo[2187]: pam_unix(sudo:session): session closed for user root May 13 23:42:29.641377 sudo[2186]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules May 13 23:42:29.641636 sudo[2186]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 13 23:42:29.650139 systemd[1]: Starting audit-rules.service - Load Audit Rules... May 13 23:42:29.681252 augenrules[2209]: No rules May 13 23:42:29.682660 systemd[1]: audit-rules.service: Deactivated successfully. May 13 23:42:29.682893 systemd[1]: Finished audit-rules.service - Load Audit Rules. May 13 23:42:29.685018 sudo[2186]: pam_unix(sudo:session): session closed for user root May 13 23:42:29.776848 sshd[2185]: Connection closed by 10.200.16.10 port 50916 May 13 23:42:29.777361 sshd-session[2183]: pam_unix(sshd:session): session closed for user core May 13 23:42:29.781305 systemd[1]: sshd@5-10.200.20.30:22-10.200.16.10:50916.service: Deactivated successfully. May 13 23:42:29.782779 systemd[1]: session-8.scope: Deactivated successfully. May 13 23:42:29.783619 systemd-logind[1744]: Session 8 logged out. Waiting for processes to exit. May 13 23:42:29.784543 systemd-logind[1744]: Removed session 8. May 13 23:42:29.866930 systemd[1]: Started sshd@6-10.200.20.30:22-10.200.16.10:50918.service - OpenSSH per-connection server daemon (10.200.16.10:50918). May 13 23:42:30.355474 sshd[2218]: Accepted publickey for core from 10.200.16.10 port 50918 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:42:30.356730 sshd-session[2218]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:42:30.361858 systemd-logind[1744]: New session 9 of user core. May 13 23:42:30.369063 systemd[1]: Started session-9.scope - Session 9 of User core. May 13 23:42:30.624358 sudo[2221]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 13 23:42:30.624794 sudo[2221]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) May 13 23:42:31.559942 systemd[1]: Starting docker.service - Docker Application Container Engine... May 13 23:42:31.570191 (dockerd)[2238]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU May 13 23:42:32.104854 dockerd[2238]: time="2025-05-13T23:42:32.104420440Z" level=info msg="Starting up" May 13 23:42:32.105838 dockerd[2238]: time="2025-05-13T23:42:32.105797840Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" May 13 23:42:32.260321 dockerd[2238]: time="2025-05-13T23:42:32.260248600Z" level=info msg="Loading containers: start." May 13 23:42:32.440838 kernel: Initializing XFRM netlink socket May 13 23:42:32.521132 systemd-networkd[1353]: docker0: Link UP May 13 23:42:32.568799 dockerd[2238]: time="2025-05-13T23:42:32.568755680Z" level=info msg="Loading containers: done." May 13 23:42:32.609421 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck4251677681-merged.mount: Deactivated successfully. May 13 23:42:32.637108 dockerd[2238]: time="2025-05-13T23:42:32.637032000Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 13 23:42:32.637530 dockerd[2238]: time="2025-05-13T23:42:32.637265520Z" level=info msg="Docker daemon" commit=c710b88579fcb5e0d53f96dcae976d79323b9166 containerd-snapshotter=false storage-driver=overlay2 version=27.4.1 May 13 23:42:32.637530 dockerd[2238]: time="2025-05-13T23:42:32.637393880Z" level=info msg="Daemon has completed initialization" May 13 23:42:32.729564 dockerd[2238]: time="2025-05-13T23:42:32.728559120Z" level=info msg="API listen on /run/docker.sock" May 13 23:42:32.728647 systemd[1]: Started docker.service - Docker Application Container Engine. May 13 23:42:33.668853 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 13 23:42:33.670236 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:33.785166 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:33.795149 (kubelet)[2442]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 13 23:42:33.828710 kubelet[2442]: E0513 23:42:33.828658 2442 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 23:42:33.830939 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 23:42:33.831082 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 23:42:33.831370 systemd[1]: kubelet.service: Consumed 123ms CPU time, 96.3M memory peak. May 13 23:42:34.603584 containerd[1799]: time="2025-05-13T23:42:34.603502360Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\"" May 13 23:42:35.237464 chronyd[1723]: Selected source PHC0 May 13 23:42:35.557553 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1439713711.mount: Deactivated successfully. May 13 23:42:37.144945 containerd[1799]: time="2025-05-13T23:42:37.144877291Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:37.149546 containerd[1799]: time="2025-05-13T23:42:37.149492539Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.12: active requests=0, bytes read=29794150" May 13 23:42:37.153031 containerd[1799]: time="2025-05-13T23:42:37.152982585Z" level=info msg="ImageCreate event name:\"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:37.162160 containerd[1799]: time="2025-05-13T23:42:37.162101759Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:37.163152 containerd[1799]: time="2025-05-13T23:42:37.162982521Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.12\" with image id \"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.12\", repo digest \"registry.k8s.io/kube-apiserver@sha256:4878682f7a044274d42399a6316ef452c5411aafd4ad99cc57de7235ca490e4e\", size \"29790950\" in 2.559439721s" May 13 23:42:37.163152 containerd[1799]: time="2025-05-13T23:42:37.163019281Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.12\" returns image reference \"sha256:afbe230ec4abc2c9e87f7fbe7814bde21dbe30f03252c8861c4ca9510cb43ec6\"" May 13 23:42:37.179155 containerd[1799]: time="2025-05-13T23:42:37.179116027Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\"" May 13 23:42:38.688858 containerd[1799]: time="2025-05-13T23:42:38.688685055Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:38.694437 containerd[1799]: time="2025-05-13T23:42:38.694382464Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.12: active requests=0, bytes read=26855550" May 13 23:42:38.701317 containerd[1799]: time="2025-05-13T23:42:38.701260596Z" level=info msg="ImageCreate event name:\"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:38.707644 containerd[1799]: time="2025-05-13T23:42:38.707568166Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:38.708590 containerd[1799]: time="2025-05-13T23:42:38.708457328Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.12\" with image id \"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.12\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:3a36711d0409d565b370a18d0c19339e93d4f1b1f2b3fd382eb31c714c463b74\", size \"28297111\" in 1.529306141s" May 13 23:42:38.708590 containerd[1799]: time="2025-05-13T23:42:38.708496568Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.12\" returns image reference \"sha256:3df23260c56ff58d759f8a841c67846184e97ce81a269549ca8d14b36da14c14\"" May 13 23:42:38.724304 containerd[1799]: time="2025-05-13T23:42:38.724037313Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\"" May 13 23:42:40.305848 containerd[1799]: time="2025-05-13T23:42:40.305653462Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:40.308807 containerd[1799]: time="2025-05-13T23:42:40.308773907Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.12: active requests=0, bytes read=16263945" May 13 23:42:40.315676 containerd[1799]: time="2025-05-13T23:42:40.315644758Z" level=info msg="ImageCreate event name:\"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:40.323659 containerd[1799]: time="2025-05-13T23:42:40.323597971Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:40.324596 containerd[1799]: time="2025-05-13T23:42:40.324452413Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.12\" with image id \"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.12\", repo digest \"registry.k8s.io/kube-scheduler@sha256:521c843d01025be7d4e246ddee8cde74556eb9813c606d6db9f0f03236f6d029\", size \"17705524\" in 1.6003783s" May 13 23:42:40.324596 containerd[1799]: time="2025-05-13T23:42:40.324484133Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.12\" returns image reference \"sha256:fb0f5dac5fa74463b801d11598454c00462609b582d17052195012e5f682c2ba\"" May 13 23:42:40.343879 containerd[1799]: time="2025-05-13T23:42:40.343836325Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\"" May 13 23:42:41.667757 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4258790754.mount: Deactivated successfully. May 13 23:42:42.034899 containerd[1799]: time="2025-05-13T23:42:42.034848615Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.12\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:42.039170 containerd[1799]: time="2025-05-13T23:42:42.039106462Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.12: active requests=0, bytes read=25775705" May 13 23:42:42.046195 containerd[1799]: time="2025-05-13T23:42:42.046136554Z" level=info msg="ImageCreate event name:\"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:42.052434 containerd[1799]: time="2025-05-13T23:42:42.052365284Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:42.053047 containerd[1799]: time="2025-05-13T23:42:42.052877525Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.12\" with image id \"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\", repo tag \"registry.k8s.io/kube-proxy:v1.30.12\", repo digest \"registry.k8s.io/kube-proxy@sha256:ea8c7d5392acf6b0c11ebba78301e1a6c2dc6abcd7544102ed578e49d1c82f15\", size \"25774724\" in 1.70899632s" May 13 23:42:42.053047 containerd[1799]: time="2025-05-13T23:42:42.052910405Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.12\" returns image reference \"sha256:b4250a9efcae16f8d20358e204a159844e2b7e854edad08aee8791774acbdaed\"" May 13 23:42:42.072737 containerd[1799]: time="2025-05-13T23:42:42.072520958Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" May 13 23:42:42.782953 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount521721010.mount: Deactivated successfully. May 13 23:42:43.918770 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. May 13 23:42:43.923024 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:44.043666 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:44.053109 (kubelet)[2607]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 13 23:42:44.100494 kubelet[2607]: E0513 23:42:44.100229 2607 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 23:42:44.102714 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 23:42:44.102890 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 23:42:44.104181 systemd[1]: kubelet.service: Consumed 132ms CPU time, 96.4M memory peak. May 13 23:42:45.047411 containerd[1799]: time="2025-05-13T23:42:45.047102803Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:45.051834 containerd[1799]: time="2025-05-13T23:42:45.051759406Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485381" May 13 23:42:45.056702 containerd[1799]: time="2025-05-13T23:42:45.056645530Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:45.066705 containerd[1799]: time="2025-05-13T23:42:45.066561098Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:45.067190 containerd[1799]: time="2025-05-13T23:42:45.067157179Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 2.994595021s" May 13 23:42:45.067227 containerd[1799]: time="2025-05-13T23:42:45.067190539Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" May 13 23:42:45.082965 containerd[1799]: time="2025-05-13T23:42:45.082929911Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" May 13 23:42:45.905062 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1846884193.mount: Deactivated successfully. May 13 23:42:45.935169 containerd[1799]: time="2025-05-13T23:42:45.935120319Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:45.945387 containerd[1799]: time="2025-05-13T23:42:45.945166047Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268821" May 13 23:42:45.951347 containerd[1799]: time="2025-05-13T23:42:45.951315212Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:45.957197 containerd[1799]: time="2025-05-13T23:42:45.957093497Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:45.957986 containerd[1799]: time="2025-05-13T23:42:45.957845738Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 874.878267ms" May 13 23:42:45.957986 containerd[1799]: time="2025-05-13T23:42:45.957878738Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" May 13 23:42:45.974165 containerd[1799]: time="2025-05-13T23:42:45.974123271Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" May 13 23:42:46.822992 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount970446776.mount: Deactivated successfully. May 13 23:42:50.576894 containerd[1799]: time="2025-05-13T23:42:50.576847733Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:50.580345 containerd[1799]: time="2025-05-13T23:42:50.580083818Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=66191472" May 13 23:42:50.585126 containerd[1799]: time="2025-05-13T23:42:50.585078586Z" level=info msg="ImageCreate event name:\"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:50.591356 containerd[1799]: time="2025-05-13T23:42:50.591302636Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:42:50.592369 containerd[1799]: time="2025-05-13T23:42:50.592246717Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"66189079\" in 4.618083766s" May 13 23:42:50.592369 containerd[1799]: time="2025-05-13T23:42:50.592279277Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" May 13 23:42:54.168805 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. May 13 23:42:54.172308 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:54.409290 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:54.416128 (kubelet)[2765]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS May 13 23:42:54.461459 kubelet[2765]: E0513 23:42:54.461339 2765 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 13 23:42:54.464084 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 13 23:42:54.464610 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 13 23:42:54.465050 systemd[1]: kubelet.service: Consumed 133ms CPU time, 94.4M memory peak. May 13 23:42:55.781849 kernel: hv_balloon: Max. dynamic memory size: 4096 MB May 13 23:42:55.804462 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:55.804602 systemd[1]: kubelet.service: Consumed 133ms CPU time, 94.4M memory peak. May 13 23:42:55.806680 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:55.830886 systemd[1]: Reload requested from client PID 2779 ('systemctl') (unit session-9.scope)... May 13 23:42:55.830903 systemd[1]: Reloading... May 13 23:42:55.959850 zram_generator::config[2828]: No configuration found. May 13 23:42:56.066163 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 23:42:56.169422 systemd[1]: Reloading finished in 338 ms. May 13 23:42:56.231661 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:56.235849 systemd[1]: kubelet.service: Deactivated successfully. May 13 23:42:56.236092 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:56.236156 systemd[1]: kubelet.service: Consumed 96ms CPU time, 82.4M memory peak. May 13 23:42:56.238000 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:42:56.358850 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:42:56.368163 (kubelet)[2894]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 13 23:42:56.408074 kubelet[2894]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 23:42:56.408414 kubelet[2894]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 13 23:42:56.408466 kubelet[2894]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 23:42:56.409623 kubelet[2894]: I0513 23:42:56.409577 2894 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 13 23:42:56.468253 update_engine[1751]: I20250513 23:42:56.468174 1751 update_attempter.cc:509] Updating boot flags... May 13 23:42:56.769866 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 41 scanned by (udev-worker) (2916) May 13 23:42:57.234287 kubelet[2894]: I0513 23:42:57.234253 2894 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 13 23:42:57.234466 kubelet[2894]: I0513 23:42:57.234456 2894 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 13 23:42:57.234725 kubelet[2894]: I0513 23:42:57.234711 2894 server.go:927] "Client rotation is on, will bootstrap in background" May 13 23:42:57.252526 kubelet[2894]: E0513 23:42:57.252493 2894 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.20.30:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.253278 kubelet[2894]: I0513 23:42:57.253251 2894 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 13 23:42:57.261383 kubelet[2894]: I0513 23:42:57.261354 2894 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 13 23:42:57.261569 kubelet[2894]: I0513 23:42:57.261537 2894 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 13 23:42:57.261735 kubelet[2894]: I0513 23:42:57.261564 2894 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4284.0.0-n-f756a0c81e","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 13 23:42:57.261847 kubelet[2894]: I0513 23:42:57.261745 2894 topology_manager.go:138] "Creating topology manager with none policy" May 13 23:42:57.261847 kubelet[2894]: I0513 23:42:57.261754 2894 container_manager_linux.go:301] "Creating device plugin manager" May 13 23:42:57.261922 kubelet[2894]: I0513 23:42:57.261901 2894 state_mem.go:36] "Initialized new in-memory state store" May 13 23:42:57.263371 kubelet[2894]: I0513 23:42:57.263352 2894 kubelet.go:400] "Attempting to sync node with API server" May 13 23:42:57.263403 kubelet[2894]: I0513 23:42:57.263376 2894 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 13 23:42:57.263596 kubelet[2894]: I0513 23:42:57.263579 2894 kubelet.go:312] "Adding apiserver pod source" May 13 23:42:57.263622 kubelet[2894]: I0513 23:42:57.263602 2894 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 13 23:42:57.265489 kubelet[2894]: W0513 23:42:57.265369 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.30:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4284.0.0-n-f756a0c81e&limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.265489 kubelet[2894]: E0513 23:42:57.265432 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.20.30:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4284.0.0-n-f756a0c81e&limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.266989 kubelet[2894]: W0513 23:42:57.265738 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.30:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.266989 kubelet[2894]: E0513 23:42:57.265782 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.20.30:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.266989 kubelet[2894]: I0513 23:42:57.265877 2894 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" May 13 23:42:57.266989 kubelet[2894]: I0513 23:42:57.266039 2894 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 13 23:42:57.266989 kubelet[2894]: W0513 23:42:57.266082 2894 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 13 23:42:57.266989 kubelet[2894]: I0513 23:42:57.266582 2894 server.go:1264] "Started kubelet" May 13 23:42:57.271983 kubelet[2894]: I0513 23:42:57.271807 2894 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 13 23:42:57.272135 kubelet[2894]: E0513 23:42:57.272019 2894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.30:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.30:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4284.0.0-n-f756a0c81e.183f3ac01eb82c65 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4284.0.0-n-f756a0c81e,UID:ci-4284.0.0-n-f756a0c81e,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4284.0.0-n-f756a0c81e,},FirstTimestamp:2025-05-13 23:42:57.266560101 +0000 UTC m=+0.894223034,LastTimestamp:2025-05-13 23:42:57.266560101 +0000 UTC m=+0.894223034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4284.0.0-n-f756a0c81e,}" May 13 23:42:57.275215 kubelet[2894]: I0513 23:42:57.274528 2894 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 13 23:42:57.275432 kubelet[2894]: I0513 23:42:57.275398 2894 server.go:455] "Adding debug handlers to kubelet server" May 13 23:42:57.276285 kubelet[2894]: I0513 23:42:57.276223 2894 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 13 23:42:57.276467 kubelet[2894]: I0513 23:42:57.276441 2894 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 13 23:42:57.277691 kubelet[2894]: I0513 23:42:57.277657 2894 volume_manager.go:291] "Starting Kubelet Volume Manager" May 13 23:42:57.278325 kubelet[2894]: I0513 23:42:57.278303 2894 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 13 23:42:57.279460 kubelet[2894]: I0513 23:42:57.279427 2894 reconciler.go:26] "Reconciler: start to sync state" May 13 23:42:57.280422 kubelet[2894]: W0513 23:42:57.280358 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.30:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.280422 kubelet[2894]: E0513 23:42:57.280420 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.20.30:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.281574 kubelet[2894]: E0513 23:42:57.280937 2894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-n-f756a0c81e?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="200ms" May 13 23:42:57.281680 kubelet[2894]: E0513 23:42:57.281639 2894 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 13 23:42:57.283158 kubelet[2894]: I0513 23:42:57.283138 2894 factory.go:221] Registration of the containerd container factory successfully May 13 23:42:57.283249 kubelet[2894]: I0513 23:42:57.283239 2894 factory.go:221] Registration of the systemd container factory successfully May 13 23:42:57.283391 kubelet[2894]: I0513 23:42:57.283367 2894 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 13 23:42:57.299014 kubelet[2894]: I0513 23:42:57.298899 2894 cpu_manager.go:214] "Starting CPU manager" policy="none" May 13 23:42:57.299138 kubelet[2894]: I0513 23:42:57.299122 2894 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 13 23:42:57.299215 kubelet[2894]: I0513 23:42:57.299206 2894 state_mem.go:36] "Initialized new in-memory state store" May 13 23:42:57.308240 kubelet[2894]: I0513 23:42:57.308209 2894 policy_none.go:49] "None policy: Start" May 13 23:42:57.309197 kubelet[2894]: I0513 23:42:57.309112 2894 memory_manager.go:170] "Starting memorymanager" policy="None" May 13 23:42:57.309278 kubelet[2894]: I0513 23:42:57.309206 2894 state_mem.go:35] "Initializing new in-memory state store" May 13 23:42:57.319384 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. May 13 23:42:57.333715 kubelet[2894]: I0513 23:42:57.333500 2894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 13 23:42:57.335421 kubelet[2894]: I0513 23:42:57.335390 2894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 13 23:42:57.335513 kubelet[2894]: I0513 23:42:57.335431 2894 status_manager.go:217] "Starting to sync pod status with apiserver" May 13 23:42:57.335513 kubelet[2894]: I0513 23:42:57.335448 2894 kubelet.go:2337] "Starting kubelet main sync loop" May 13 23:42:57.335513 kubelet[2894]: E0513 23:42:57.335485 2894 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 13 23:42:57.336731 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. May 13 23:42:57.339040 kubelet[2894]: W0513 23:42:57.338959 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.30:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.339040 kubelet[2894]: E0513 23:42:57.338997 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.20.30:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:57.343733 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. May 13 23:42:57.353793 kubelet[2894]: I0513 23:42:57.353760 2894 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 13 23:42:57.353793 kubelet[2894]: I0513 23:42:57.353973 2894 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 13 23:42:57.353793 kubelet[2894]: I0513 23:42:57.354077 2894 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 13 23:42:57.355910 kubelet[2894]: E0513 23:42:57.355870 2894 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4284.0.0-n-f756a0c81e\" not found" May 13 23:42:57.380532 kubelet[2894]: I0513 23:42:57.380241 2894 kubelet_node_status.go:73] "Attempting to register node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.380771 kubelet[2894]: E0513 23:42:57.380748 2894 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.435953 kubelet[2894]: I0513 23:42:57.435912 2894 topology_manager.go:215] "Topology Admit Handler" podUID="3bee1d25c181715812c175aa23147e3b" podNamespace="kube-system" podName="kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.437935 kubelet[2894]: I0513 23:42:57.437654 2894 topology_manager.go:215] "Topology Admit Handler" podUID="73ec5123b8e39319d6ca407951132338" podNamespace="kube-system" podName="kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.439199 kubelet[2894]: I0513 23:42:57.439145 2894 topology_manager.go:215] "Topology Admit Handler" podUID="c8e183324f6a7a7a6779fb40e0a8a418" podNamespace="kube-system" podName="kube-scheduler-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.446982 systemd[1]: Created slice kubepods-burstable-pod73ec5123b8e39319d6ca407951132338.slice - libcontainer container kubepods-burstable-pod73ec5123b8e39319d6ca407951132338.slice. May 13 23:42:57.459167 systemd[1]: Created slice kubepods-burstable-pod3bee1d25c181715812c175aa23147e3b.slice - libcontainer container kubepods-burstable-pod3bee1d25c181715812c175aa23147e3b.slice. May 13 23:42:57.466675 systemd[1]: Created slice kubepods-burstable-podc8e183324f6a7a7a6779fb40e0a8a418.slice - libcontainer container kubepods-burstable-podc8e183324f6a7a7a6779fb40e0a8a418.slice. May 13 23:42:57.480037 kubelet[2894]: I0513 23:42:57.480006 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3bee1d25c181715812c175aa23147e3b-k8s-certs\") pod \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" (UID: \"3bee1d25c181715812c175aa23147e3b\") " pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480121 kubelet[2894]: I0513 23:42:57.480043 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3bee1d25c181715812c175aa23147e3b-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" (UID: \"3bee1d25c181715812c175aa23147e3b\") " pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480121 kubelet[2894]: I0513 23:42:57.480065 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-ca-certs\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480121 kubelet[2894]: I0513 23:42:57.480082 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-k8s-certs\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480121 kubelet[2894]: I0513 23:42:57.480099 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c8e183324f6a7a7a6779fb40e0a8a418-kubeconfig\") pod \"kube-scheduler-ci-4284.0.0-n-f756a0c81e\" (UID: \"c8e183324f6a7a7a6779fb40e0a8a418\") " pod="kube-system/kube-scheduler-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480121 kubelet[2894]: I0513 23:42:57.480114 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3bee1d25c181715812c175aa23147e3b-ca-certs\") pod \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" (UID: \"3bee1d25c181715812c175aa23147e3b\") " pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480233 kubelet[2894]: I0513 23:42:57.480128 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-flexvolume-dir\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480233 kubelet[2894]: I0513 23:42:57.480144 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-kubeconfig\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.480233 kubelet[2894]: I0513 23:42:57.480158 2894 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.481598 kubelet[2894]: E0513 23:42:57.481555 2894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-n-f756a0c81e?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="400ms" May 13 23:42:57.582807 kubelet[2894]: I0513 23:42:57.582711 2894 kubelet_node_status.go:73] "Attempting to register node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.583112 kubelet[2894]: E0513 23:42:57.583054 2894 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.757510 containerd[1799]: time="2025-05-13T23:42:57.757463931Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4284.0.0-n-f756a0c81e,Uid:73ec5123b8e39319d6ca407951132338,Namespace:kube-system,Attempt:0,}" May 13 23:42:57.765312 containerd[1799]: time="2025-05-13T23:42:57.765123944Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4284.0.0-n-f756a0c81e,Uid:3bee1d25c181715812c175aa23147e3b,Namespace:kube-system,Attempt:0,}" May 13 23:42:57.770021 containerd[1799]: time="2025-05-13T23:42:57.769894152Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4284.0.0-n-f756a0c81e,Uid:c8e183324f6a7a7a6779fb40e0a8a418,Namespace:kube-system,Attempt:0,}" May 13 23:42:57.882381 kubelet[2894]: E0513 23:42:57.882270 2894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-n-f756a0c81e?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="800ms" May 13 23:42:57.985139 kubelet[2894]: I0513 23:42:57.985053 2894 kubelet_node_status.go:73] "Attempting to register node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:57.985465 kubelet[2894]: E0513 23:42:57.985433 2894 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:58.382217 kubelet[2894]: W0513 23:42:58.382154 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.30:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.382217 kubelet[2894]: E0513 23:42:58.382195 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.20.30:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.458744 kubelet[2894]: W0513 23:42:58.458664 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.30:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4284.0.0-n-f756a0c81e&limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.458744 kubelet[2894]: E0513 23:42:58.458721 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.20.30:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4284.0.0-n-f756a0c81e&limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.549403 kubelet[2894]: W0513 23:42:58.549340 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.30:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.549403 kubelet[2894]: E0513 23:42:58.549404 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.20.30:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.604163 kubelet[2894]: W0513 23:42:58.604103 2894 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.30:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.604163 kubelet[2894]: E0513 23:42:58.604143 2894 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.20.30:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:58.682836 kubelet[2894]: E0513 23:42:58.682776 2894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4284.0.0-n-f756a0c81e?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="1.6s" May 13 23:42:58.779407 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount823977697.mount: Deactivated successfully. May 13 23:42:58.788211 kubelet[2894]: I0513 23:42:58.787798 2894 kubelet_node_status.go:73] "Attempting to register node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:58.788211 kubelet[2894]: E0513 23:42:58.788121 2894 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4284.0.0-n-f756a0c81e" May 13 23:42:58.833860 containerd[1799]: time="2025-05-13T23:42:58.833506272Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 13 23:42:58.852245 containerd[1799]: time="2025-05-13T23:42:58.852173864Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" May 13 23:42:58.858242 containerd[1799]: time="2025-05-13T23:42:58.857597953Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 13 23:42:58.864298 containerd[1799]: time="2025-05-13T23:42:58.864251844Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 13 23:42:58.879240 containerd[1799]: time="2025-05-13T23:42:58.879169269Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 13 23:42:58.883378 containerd[1799]: time="2025-05-13T23:42:58.883325756Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 13 23:42:58.886427 containerd[1799]: time="2025-05-13T23:42:58.886374282Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=0" May 13 23:42:58.892715 containerd[1799]: time="2025-05-13T23:42:58.892644652Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" May 13 23:42:58.893882 containerd[1799]: time="2025-05-13T23:42:58.893269373Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 1.092751769s" May 13 23:42:58.899039 containerd[1799]: time="2025-05-13T23:42:58.898998423Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 1.019148084s" May 13 23:42:58.908194 containerd[1799]: time="2025-05-13T23:42:58.908088318Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 1.034097269s" May 13 23:42:58.996428 containerd[1799]: time="2025-05-13T23:42:58.996307788Z" level=info msg="connecting to shim 8d64794d46e2a7273f6712bfc1a61157fcc1ac8ecf5f691378494a49f15ea7af" address="unix:///run/containerd/s/407f2120d223bdf8cf9324843abb0de9c357040e1c9bfae7cd0451452b87a2fb" namespace=k8s.io protocol=ttrpc version=3 May 13 23:42:59.018220 containerd[1799]: time="2025-05-13T23:42:59.017553344Z" level=info msg="connecting to shim 78ec8232a04d51fb53c0d4086a59da3d751e3418f5a6a69848bdcfc9bee8ce7c" address="unix:///run/containerd/s/944a7879b6f257119ae21b803e538c7106cf485db02a4a84133624e9d2617c41" namespace=k8s.io protocol=ttrpc version=3 May 13 23:42:59.026174 systemd[1]: Started cri-containerd-8d64794d46e2a7273f6712bfc1a61157fcc1ac8ecf5f691378494a49f15ea7af.scope - libcontainer container 8d64794d46e2a7273f6712bfc1a61157fcc1ac8ecf5f691378494a49f15ea7af. May 13 23:42:59.033289 containerd[1799]: time="2025-05-13T23:42:59.032858889Z" level=info msg="connecting to shim 207387820e313aa1ee7ee578abfc7a077c9dd66bf5005d2d8af9107e775207b7" address="unix:///run/containerd/s/dadb4d56e7a729ca3da25c75464646ec92566ff3555a22cff606a26ded50dcad" namespace=k8s.io protocol=ttrpc version=3 May 13 23:42:59.063006 systemd[1]: Started cri-containerd-207387820e313aa1ee7ee578abfc7a077c9dd66bf5005d2d8af9107e775207b7.scope - libcontainer container 207387820e313aa1ee7ee578abfc7a077c9dd66bf5005d2d8af9107e775207b7. May 13 23:42:59.064119 systemd[1]: Started cri-containerd-78ec8232a04d51fb53c0d4086a59da3d751e3418f5a6a69848bdcfc9bee8ce7c.scope - libcontainer container 78ec8232a04d51fb53c0d4086a59da3d751e3418f5a6a69848bdcfc9bee8ce7c. May 13 23:42:59.088228 containerd[1799]: time="2025-05-13T23:42:59.088074783Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4284.0.0-n-f756a0c81e,Uid:73ec5123b8e39319d6ca407951132338,Namespace:kube-system,Attempt:0,} returns sandbox id \"8d64794d46e2a7273f6712bfc1a61157fcc1ac8ecf5f691378494a49f15ea7af\"" May 13 23:42:59.092838 containerd[1799]: time="2025-05-13T23:42:59.091569989Z" level=info msg="CreateContainer within sandbox \"8d64794d46e2a7273f6712bfc1a61157fcc1ac8ecf5f691378494a49f15ea7af\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 13 23:42:59.129245 containerd[1799]: time="2025-05-13T23:42:59.129205493Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4284.0.0-n-f756a0c81e,Uid:c8e183324f6a7a7a6779fb40e0a8a418,Namespace:kube-system,Attempt:0,} returns sandbox id \"207387820e313aa1ee7ee578abfc7a077c9dd66bf5005d2d8af9107e775207b7\"" May 13 23:42:59.131524 containerd[1799]: time="2025-05-13T23:42:59.131494136Z" level=info msg="CreateContainer within sandbox \"207387820e313aa1ee7ee578abfc7a077c9dd66bf5005d2d8af9107e775207b7\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 13 23:42:59.139647 containerd[1799]: time="2025-05-13T23:42:59.139610750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4284.0.0-n-f756a0c81e,Uid:3bee1d25c181715812c175aa23147e3b,Namespace:kube-system,Attempt:0,} returns sandbox id \"78ec8232a04d51fb53c0d4086a59da3d751e3418f5a6a69848bdcfc9bee8ce7c\"" May 13 23:42:59.142600 containerd[1799]: time="2025-05-13T23:42:59.142567035Z" level=info msg="CreateContainer within sandbox \"78ec8232a04d51fb53c0d4086a59da3d751e3418f5a6a69848bdcfc9bee8ce7c\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 13 23:42:59.153484 containerd[1799]: time="2025-05-13T23:42:59.153432014Z" level=info msg="Container 99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda: CDI devices from CRI Config.CDIDevices: []" May 13 23:42:59.241251 containerd[1799]: time="2025-05-13T23:42:59.241204002Z" level=info msg="CreateContainer within sandbox \"8d64794d46e2a7273f6712bfc1a61157fcc1ac8ecf5f691378494a49f15ea7af\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda\"" May 13 23:42:59.241874 containerd[1799]: time="2025-05-13T23:42:59.241851163Z" level=info msg="StartContainer for \"99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda\"" May 13 23:42:59.243591 containerd[1799]: time="2025-05-13T23:42:59.243506806Z" level=info msg="connecting to shim 99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda" address="unix:///run/containerd/s/407f2120d223bdf8cf9324843abb0de9c357040e1c9bfae7cd0451452b87a2fb" protocol=ttrpc version=3 May 13 23:42:59.247258 containerd[1799]: time="2025-05-13T23:42:59.246525771Z" level=info msg="Container 8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951: CDI devices from CRI Config.CDIDevices: []" May 13 23:42:59.260466 containerd[1799]: time="2025-05-13T23:42:59.260125194Z" level=info msg="Container 6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a: CDI devices from CRI Config.CDIDevices: []" May 13 23:42:59.264993 systemd[1]: Started cri-containerd-99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda.scope - libcontainer container 99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda. May 13 23:42:59.279534 containerd[1799]: time="2025-05-13T23:42:59.279410226Z" level=info msg="CreateContainer within sandbox \"207387820e313aa1ee7ee578abfc7a077c9dd66bf5005d2d8af9107e775207b7\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951\"" May 13 23:42:59.280419 containerd[1799]: time="2025-05-13T23:42:59.280393508Z" level=info msg="StartContainer for \"8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951\"" May 13 23:42:59.281839 containerd[1799]: time="2025-05-13T23:42:59.281653310Z" level=info msg="connecting to shim 8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951" address="unix:///run/containerd/s/dadb4d56e7a729ca3da25c75464646ec92566ff3555a22cff606a26ded50dcad" protocol=ttrpc version=3 May 13 23:42:59.292290 containerd[1799]: time="2025-05-13T23:42:59.292248447Z" level=info msg="CreateContainer within sandbox \"78ec8232a04d51fb53c0d4086a59da3d751e3418f5a6a69848bdcfc9bee8ce7c\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a\"" May 13 23:42:59.293353 containerd[1799]: time="2025-05-13T23:42:59.293329409Z" level=info msg="StartContainer for \"6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a\"" May 13 23:42:59.294757 containerd[1799]: time="2025-05-13T23:42:59.294522370Z" level=info msg="connecting to shim 6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a" address="unix:///run/containerd/s/944a7879b6f257119ae21b803e538c7106cf485db02a4a84133624e9d2617c41" protocol=ttrpc version=3 May 13 23:42:59.312135 systemd[1]: Started cri-containerd-8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951.scope - libcontainer container 8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951. May 13 23:42:59.329857 containerd[1799]: time="2025-05-13T23:42:59.328990746Z" level=info msg="StartContainer for \"99414b2843291f2894363007e1289dc65e6030fee6cbe7893440e4e1efbc3eda\" returns successfully" May 13 23:42:59.332151 systemd[1]: Started cri-containerd-6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a.scope - libcontainer container 6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a. May 13 23:42:59.391567 kubelet[2894]: E0513 23:42:59.390900 2894 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.20.30:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.20.30:6443: connect: connection refused May 13 23:42:59.415271 containerd[1799]: time="2025-05-13T23:42:59.415225126Z" level=info msg="StartContainer for \"8d0efbc73b23db6a19204af004f9ccae5ce03978aa19624af2f3a6a63f98b951\" returns successfully" May 13 23:42:59.415414 containerd[1799]: time="2025-05-13T23:42:59.415396046Z" level=info msg="StartContainer for \"6b2183f12ce0644f95d6660e4b06f957ea73ad3c7920ff4d3719fe90b08d4b4a\" returns successfully" May 13 23:43:00.392921 kubelet[2894]: I0513 23:43:00.392886 2894 kubelet_node_status.go:73] "Attempting to register node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:43:01.372002 kubelet[2894]: E0513 23:43:01.371960 2894 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4284.0.0-n-f756a0c81e\" not found" node="ci-4284.0.0-n-f756a0c81e" May 13 23:43:01.472767 kubelet[2894]: I0513 23:43:01.472627 2894 kubelet_node_status.go:76] "Successfully registered node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:43:02.269075 kubelet[2894]: I0513 23:43:02.269032 2894 apiserver.go:52] "Watching apiserver" May 13 23:43:02.278691 kubelet[2894]: I0513 23:43:02.278658 2894 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 13 23:43:02.395540 kubelet[2894]: W0513 23:43:02.395484 2894 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 13 23:43:02.396384 kubelet[2894]: W0513 23:43:02.396360 2894 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 13 23:43:03.813715 systemd[1]: Reload requested from client PID 3233 ('systemctl') (unit session-9.scope)... May 13 23:43:03.813735 systemd[1]: Reloading... May 13 23:43:03.913892 zram_generator::config[3283]: No configuration found. May 13 23:43:04.042442 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 13 23:43:04.162132 systemd[1]: Reloading finished in 348 ms. May 13 23:43:04.182496 kubelet[2894]: E0513 23:43:04.182336 2894 event.go:319] "Unable to write event (broadcaster is shut down)" event="&Event{ObjectMeta:{ci-4284.0.0-n-f756a0c81e.183f3ac01eb82c65 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4284.0.0-n-f756a0c81e,UID:ci-4284.0.0-n-f756a0c81e,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4284.0.0-n-f756a0c81e,},FirstTimestamp:2025-05-13 23:42:57.266560101 +0000 UTC m=+0.894223034,LastTimestamp:2025-05-13 23:42:57.266560101 +0000 UTC m=+0.894223034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4284.0.0-n-f756a0c81e,}" May 13 23:43:04.182663 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:43:04.196188 systemd[1]: kubelet.service: Deactivated successfully. May 13 23:43:04.196421 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:43:04.196478 systemd[1]: kubelet.service: Consumed 1.195s CPU time, 113.1M memory peak. May 13 23:43:04.201158 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... May 13 23:43:04.322615 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. May 13 23:43:04.330122 (kubelet)[3344]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS May 13 23:43:04.378376 kubelet[3344]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 23:43:04.378376 kubelet[3344]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 13 23:43:04.378376 kubelet[3344]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 13 23:43:04.378745 kubelet[3344]: I0513 23:43:04.378420 3344 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 13 23:43:04.384836 kubelet[3344]: I0513 23:43:04.384767 3344 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" May 13 23:43:04.384836 kubelet[3344]: I0513 23:43:04.384798 3344 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 13 23:43:04.385416 kubelet[3344]: I0513 23:43:04.385030 3344 server.go:927] "Client rotation is on, will bootstrap in background" May 13 23:43:04.386483 kubelet[3344]: I0513 23:43:04.386393 3344 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 13 23:43:04.387958 kubelet[3344]: I0513 23:43:04.387714 3344 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 13 23:43:04.398810 kubelet[3344]: I0513 23:43:04.398772 3344 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 13 23:43:04.399059 kubelet[3344]: I0513 23:43:04.399023 3344 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 13 23:43:04.399231 kubelet[3344]: I0513 23:43:04.399057 3344 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4284.0.0-n-f756a0c81e","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} May 13 23:43:04.399318 kubelet[3344]: I0513 23:43:04.399234 3344 topology_manager.go:138] "Creating topology manager with none policy" May 13 23:43:04.399318 kubelet[3344]: I0513 23:43:04.399244 3344 container_manager_linux.go:301] "Creating device plugin manager" May 13 23:43:04.399318 kubelet[3344]: I0513 23:43:04.399282 3344 state_mem.go:36] "Initialized new in-memory state store" May 13 23:43:04.399410 kubelet[3344]: I0513 23:43:04.399390 3344 kubelet.go:400] "Attempting to sync node with API server" May 13 23:43:04.399439 kubelet[3344]: I0513 23:43:04.399410 3344 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" May 13 23:43:04.399873 kubelet[3344]: I0513 23:43:04.399846 3344 kubelet.go:312] "Adding apiserver pod source" May 13 23:43:04.399920 kubelet[3344]: I0513 23:43:04.399879 3344 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 13 23:43:04.403541 kubelet[3344]: I0513 23:43:04.403508 3344 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v2.0.1" apiVersion="v1" May 13 23:43:04.404820 kubelet[3344]: I0513 23:43:04.404016 3344 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 13 23:43:04.404820 kubelet[3344]: I0513 23:43:04.404539 3344 server.go:1264] "Started kubelet" May 13 23:43:04.409531 kubelet[3344]: I0513 23:43:04.408426 3344 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.423959 3344 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.425095 3344 server.go:455] "Adding debug handlers to kubelet server" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.426997 3344 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.446376 3344 volume_manager.go:291] "Starting Kubelet Volume Manager" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.446474 3344 desired_state_of_world_populator.go:149] "Desired state populator starts to run" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.452261 3344 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 13 23:43:04.595733 kubelet[3344]: E0513 23:43:04.452600 3344 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.459515 3344 factory.go:221] Registration of the containerd container factory successfully May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.459531 3344 factory.go:221] Registration of the systemd container factory successfully May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.462855 3344 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.465433 3344 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.465467 3344 status_manager.go:217] "Starting to sync pod status with apiserver" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.465484 3344 kubelet.go:2337] "Starting kubelet main sync loop" May 13 23:43:04.595733 kubelet[3344]: E0513 23:43:04.465526 3344 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.507050 3344 cpu_manager.go:214] "Starting CPU manager" policy="none" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.507067 3344 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 13 23:43:04.595733 kubelet[3344]: I0513 23:43:04.507089 3344 state_mem.go:36] "Initialized new in-memory state store" May 13 23:43:04.596226 kubelet[3344]: I0513 23:43:04.549964 3344 kubelet_node_status.go:73] "Attempting to register node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.596226 kubelet[3344]: E0513 23:43:04.565791 3344 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 13 23:43:04.596226 kubelet[3344]: I0513 23:43:04.566779 3344 kubelet_node_status.go:112] "Node was previously registered" node="ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.596226 kubelet[3344]: I0513 23:43:04.595473 3344 kubelet_node_status.go:76] "Successfully registered node" node="ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.597621 kubelet[3344]: I0513 23:43:04.596436 3344 reconciler.go:26] "Reconciler: start to sync state" May 13 23:43:04.597621 kubelet[3344]: I0513 23:43:04.596669 3344 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 13 23:43:04.597621 kubelet[3344]: I0513 23:43:04.596682 3344 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 13 23:43:04.597621 kubelet[3344]: I0513 23:43:04.596713 3344 policy_none.go:49] "None policy: Start" May 13 23:43:04.598808 kubelet[3344]: I0513 23:43:04.598794 3344 memory_manager.go:170] "Starting memorymanager" policy="None" May 13 23:43:04.598994 kubelet[3344]: I0513 23:43:04.598937 3344 state_mem.go:35] "Initializing new in-memory state store" May 13 23:43:04.599319 kubelet[3344]: I0513 23:43:04.599191 3344 state_mem.go:75] "Updated machine memory state" May 13 23:43:04.600335 kubelet[3344]: I0513 23:43:04.600273 3344 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 13 23:43:04.607862 kubelet[3344]: I0513 23:43:04.605451 3344 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 13 23:43:04.607862 kubelet[3344]: I0513 23:43:04.605606 3344 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 13 23:43:04.607862 kubelet[3344]: I0513 23:43:04.605704 3344 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 13 23:43:04.766981 kubelet[3344]: I0513 23:43:04.766928 3344 topology_manager.go:215] "Topology Admit Handler" podUID="c8e183324f6a7a7a6779fb40e0a8a418" podNamespace="kube-system" podName="kube-scheduler-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.767144 kubelet[3344]: I0513 23:43:04.767074 3344 topology_manager.go:215] "Topology Admit Handler" podUID="3bee1d25c181715812c175aa23147e3b" podNamespace="kube-system" podName="kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.767144 kubelet[3344]: I0513 23:43:04.767116 3344 topology_manager.go:215] "Topology Admit Handler" podUID="73ec5123b8e39319d6ca407951132338" podNamespace="kube-system" podName="kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.787009 kubelet[3344]: W0513 23:43:04.786852 3344 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 13 23:43:04.787009 kubelet[3344]: E0513 23:43:04.786926 3344 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" already exists" pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.787009 kubelet[3344]: W0513 23:43:04.786938 3344 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 13 23:43:04.787937 kubelet[3344]: W0513 23:43:04.787806 3344 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] May 13 23:43:04.787937 kubelet[3344]: E0513 23:43:04.787872 3344 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4284.0.0-n-f756a0c81e\" already exists" pod="kube-system/kube-scheduler-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.828496 sudo[3379]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 13 23:43:04.828766 sudo[3379]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) May 13 23:43:04.899199 kubelet[3344]: I0513 23:43:04.899057 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-ca-certs\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899199 kubelet[3344]: I0513 23:43:04.899102 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-flexvolume-dir\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899199 kubelet[3344]: I0513 23:43:04.899138 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899199 kubelet[3344]: I0513 23:43:04.899159 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/c8e183324f6a7a7a6779fb40e0a8a418-kubeconfig\") pod \"kube-scheduler-ci-4284.0.0-n-f756a0c81e\" (UID: \"c8e183324f6a7a7a6779fb40e0a8a418\") " pod="kube-system/kube-scheduler-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899199 kubelet[3344]: I0513 23:43:04.899177 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3bee1d25c181715812c175aa23147e3b-ca-certs\") pod \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" (UID: \"3bee1d25c181715812c175aa23147e3b\") " pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899513 kubelet[3344]: I0513 23:43:04.899206 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3bee1d25c181715812c175aa23147e3b-k8s-certs\") pod \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" (UID: \"3bee1d25c181715812c175aa23147e3b\") " pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899513 kubelet[3344]: I0513 23:43:04.899222 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3bee1d25c181715812c175aa23147e3b-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4284.0.0-n-f756a0c81e\" (UID: \"3bee1d25c181715812c175aa23147e3b\") " pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899513 kubelet[3344]: I0513 23:43:04.899236 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-k8s-certs\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:43:04.899513 kubelet[3344]: I0513 23:43:04.899251 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/73ec5123b8e39319d6ca407951132338-kubeconfig\") pod \"kube-controller-manager-ci-4284.0.0-n-f756a0c81e\" (UID: \"73ec5123b8e39319d6ca407951132338\") " pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" May 13 23:43:05.289370 sudo[3379]: pam_unix(sudo:session): session closed for user root May 13 23:43:05.400849 kubelet[3344]: I0513 23:43:05.400592 3344 apiserver.go:52] "Watching apiserver" May 13 23:43:05.447691 kubelet[3344]: I0513 23:43:05.447616 3344 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" May 13 23:43:05.460331 kubelet[3344]: I0513 23:43:05.460185 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4284.0.0-n-f756a0c81e" podStartSLOduration=1.460168778 podStartE2EDuration="1.460168778s" podCreationTimestamp="2025-05-13 23:43:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:43:05.460108538 +0000 UTC m=+1.124975340" watchObservedRunningTime="2025-05-13 23:43:05.460168778 +0000 UTC m=+1.125035580" May 13 23:43:05.498696 kubelet[3344]: I0513 23:43:05.497451 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4284.0.0-n-f756a0c81e" podStartSLOduration=3.4974336790000002 podStartE2EDuration="3.497433679s" podCreationTimestamp="2025-05-13 23:43:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:43:05.476438285 +0000 UTC m=+1.141305087" watchObservedRunningTime="2025-05-13 23:43:05.497433679 +0000 UTC m=+1.162300481" May 13 23:43:05.513892 kubelet[3344]: I0513 23:43:05.513626 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4284.0.0-n-f756a0c81e" podStartSLOduration=3.513608825 podStartE2EDuration="3.513608825s" podCreationTimestamp="2025-05-13 23:43:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:43:05.497777159 +0000 UTC m=+1.162643961" watchObservedRunningTime="2025-05-13 23:43:05.513608825 +0000 UTC m=+1.178475587" May 13 23:43:07.108584 sudo[2221]: pam_unix(sudo:session): session closed for user root May 13 23:43:07.200769 sshd[2220]: Connection closed by 10.200.16.10 port 50918 May 13 23:43:07.201115 sshd-session[2218]: pam_unix(sshd:session): session closed for user core May 13 23:43:07.206554 systemd-logind[1744]: Session 9 logged out. Waiting for processes to exit. May 13 23:43:07.206735 systemd[1]: sshd@6-10.200.20.30:22-10.200.16.10:50918.service: Deactivated successfully. May 13 23:43:07.210260 systemd[1]: session-9.scope: Deactivated successfully. May 13 23:43:07.211353 systemd[1]: session-9.scope: Consumed 7.141s CPU time, 282.7M memory peak. May 13 23:43:07.214075 systemd-logind[1744]: Removed session 9. May 13 23:43:17.374454 kubelet[3344]: I0513 23:43:17.374421 3344 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 13 23:43:17.375426 containerd[1799]: time="2025-05-13T23:43:17.375323849Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 13 23:43:17.376188 kubelet[3344]: I0513 23:43:17.375531 3344 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 13 23:43:18.340712 kubelet[3344]: I0513 23:43:18.340660 3344 topology_manager.go:215] "Topology Admit Handler" podUID="815743fd-a18b-4aaf-993d-3b7577b9dd4f" podNamespace="kube-system" podName="cilium-operator-599987898-r7bdk" May 13 23:43:18.349681 systemd[1]: Created slice kubepods-besteffort-pod815743fd_a18b_4aaf_993d_3b7577b9dd4f.slice - libcontainer container kubepods-besteffort-pod815743fd_a18b_4aaf_993d_3b7577b9dd4f.slice. May 13 23:43:18.476797 kubelet[3344]: I0513 23:43:18.476755 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/815743fd-a18b-4aaf-993d-3b7577b9dd4f-cilium-config-path\") pod \"cilium-operator-599987898-r7bdk\" (UID: \"815743fd-a18b-4aaf-993d-3b7577b9dd4f\") " pod="kube-system/cilium-operator-599987898-r7bdk" May 13 23:43:18.476797 kubelet[3344]: I0513 23:43:18.476799 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f5nn\" (UniqueName: \"kubernetes.io/projected/815743fd-a18b-4aaf-993d-3b7577b9dd4f-kube-api-access-2f5nn\") pod \"cilium-operator-599987898-r7bdk\" (UID: \"815743fd-a18b-4aaf-993d-3b7577b9dd4f\") " pod="kube-system/cilium-operator-599987898-r7bdk" May 13 23:43:18.532807 kubelet[3344]: I0513 23:43:18.532664 3344 topology_manager.go:215] "Topology Admit Handler" podUID="7660b3cc-589f-405f-856c-3ccb4aed108a" podNamespace="kube-system" podName="kube-proxy-s2wp5" May 13 23:43:18.534863 kubelet[3344]: I0513 23:43:18.534638 3344 topology_manager.go:215] "Topology Admit Handler" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" podNamespace="kube-system" podName="cilium-jtz2z" May 13 23:43:18.545214 systemd[1]: Created slice kubepods-besteffort-pod7660b3cc_589f_405f_856c_3ccb4aed108a.slice - libcontainer container kubepods-besteffort-pod7660b3cc_589f_405f_856c_3ccb4aed108a.slice. May 13 23:43:18.553005 systemd[1]: Created slice kubepods-burstable-podca8a8273_b2db_4a64_922e_a6200979faee.slice - libcontainer container kubepods-burstable-podca8a8273_b2db_4a64_922e_a6200979faee.slice. May 13 23:43:18.657442 containerd[1799]: time="2025-05-13T23:43:18.657125429Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-r7bdk,Uid:815743fd-a18b-4aaf-993d-3b7577b9dd4f,Namespace:kube-system,Attempt:0,}" May 13 23:43:18.679847 kubelet[3344]: I0513 23:43:18.677941 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-cgroup\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.679847 kubelet[3344]: I0513 23:43:18.677989 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-kernel\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.679847 kubelet[3344]: I0513 23:43:18.678014 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7660b3cc-589f-405f-856c-3ccb4aed108a-lib-modules\") pod \"kube-proxy-s2wp5\" (UID: \"7660b3cc-589f-405f-856c-3ccb4aed108a\") " pod="kube-system/kube-proxy-s2wp5" May 13 23:43:18.679847 kubelet[3344]: I0513 23:43:18.678041 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjf44\" (UniqueName: \"kubernetes.io/projected/7660b3cc-589f-405f-856c-3ccb4aed108a-kube-api-access-rjf44\") pod \"kube-proxy-s2wp5\" (UID: \"7660b3cc-589f-405f-856c-3ccb4aed108a\") " pod="kube-system/kube-proxy-s2wp5" May 13 23:43:18.679847 kubelet[3344]: I0513 23:43:18.678064 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-run\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.679847 kubelet[3344]: I0513 23:43:18.678084 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-bpf-maps\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680138 kubelet[3344]: I0513 23:43:18.678102 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-etc-cni-netd\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680138 kubelet[3344]: I0513 23:43:18.678122 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-net\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680138 kubelet[3344]: I0513 23:43:18.678138 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-xtables-lock\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680138 kubelet[3344]: I0513 23:43:18.678157 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-config-path\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680138 kubelet[3344]: I0513 23:43:18.678178 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-hubble-tls\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680138 kubelet[3344]: I0513 23:43:18.678197 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/7660b3cc-589f-405f-856c-3ccb4aed108a-xtables-lock\") pod \"kube-proxy-s2wp5\" (UID: \"7660b3cc-589f-405f-856c-3ccb4aed108a\") " pod="kube-system/kube-proxy-s2wp5" May 13 23:43:18.680268 kubelet[3344]: I0513 23:43:18.678216 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-hostproc\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680268 kubelet[3344]: I0513 23:43:18.678232 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ca8a8273-b2db-4a64-922e-a6200979faee-clustermesh-secrets\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680268 kubelet[3344]: I0513 23:43:18.678253 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/7660b3cc-589f-405f-856c-3ccb4aed108a-kube-proxy\") pod \"kube-proxy-s2wp5\" (UID: \"7660b3cc-589f-405f-856c-3ccb4aed108a\") " pod="kube-system/kube-proxy-s2wp5" May 13 23:43:18.680268 kubelet[3344]: I0513 23:43:18.678272 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-lib-modules\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680268 kubelet[3344]: I0513 23:43:18.678290 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cni-path\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.680268 kubelet[3344]: I0513 23:43:18.678312 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkjs2\" (UniqueName: \"kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-kube-api-access-bkjs2\") pod \"cilium-jtz2z\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " pod="kube-system/cilium-jtz2z" May 13 23:43:18.765292 containerd[1799]: time="2025-05-13T23:43:18.764793460Z" level=info msg="connecting to shim f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e" address="unix:///run/containerd/s/2ba2c3739628916c84bfa229306b218dc765249846a84391ec831be8d5048e59" namespace=k8s.io protocol=ttrpc version=3 May 13 23:43:18.787465 systemd[1]: Started cri-containerd-f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e.scope - libcontainer container f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e. May 13 23:43:18.844632 containerd[1799]: time="2025-05-13T23:43:18.844583230Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-r7bdk,Uid:815743fd-a18b-4aaf-993d-3b7577b9dd4f,Namespace:kube-system,Attempt:0,} returns sandbox id \"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\"" May 13 23:43:18.847130 containerd[1799]: time="2025-05-13T23:43:18.847091196Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 13 23:43:18.853702 containerd[1799]: time="2025-05-13T23:43:18.853651530Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-s2wp5,Uid:7660b3cc-589f-405f-856c-3ccb4aed108a,Namespace:kube-system,Attempt:0,}" May 13 23:43:18.862919 containerd[1799]: time="2025-05-13T23:43:18.862806629Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jtz2z,Uid:ca8a8273-b2db-4a64-922e-a6200979faee,Namespace:kube-system,Attempt:0,}" May 13 23:43:18.997925 containerd[1799]: time="2025-05-13T23:43:18.997874198Z" level=info msg="connecting to shim 0c1b7d9084bf65a480bc269a2ac9ce9a9a8ad107c0fc97b1e89a3bb3999302e6" address="unix:///run/containerd/s/148592215ec6a389458cc912acad4785dcab2c3156643efc1dc4c9283f1cc811" namespace=k8s.io protocol=ttrpc version=3 May 13 23:43:19.016956 containerd[1799]: time="2025-05-13T23:43:19.016087037Z" level=info msg="connecting to shim 75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb" address="unix:///run/containerd/s/1ae4c7d50d628953eaa40a240cd95e556119175b6292c0e7b7dc6c84bacff4f5" namespace=k8s.io protocol=ttrpc version=3 May 13 23:43:19.020047 systemd[1]: Started cri-containerd-0c1b7d9084bf65a480bc269a2ac9ce9a9a8ad107c0fc97b1e89a3bb3999302e6.scope - libcontainer container 0c1b7d9084bf65a480bc269a2ac9ce9a9a8ad107c0fc97b1e89a3bb3999302e6. May 13 23:43:19.045990 systemd[1]: Started cri-containerd-75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb.scope - libcontainer container 75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb. May 13 23:43:19.077180 containerd[1799]: time="2025-05-13T23:43:19.077135888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-s2wp5,Uid:7660b3cc-589f-405f-856c-3ccb4aed108a,Namespace:kube-system,Attempt:0,} returns sandbox id \"0c1b7d9084bf65a480bc269a2ac9ce9a9a8ad107c0fc97b1e89a3bb3999302e6\"" May 13 23:43:19.080810 containerd[1799]: time="2025-05-13T23:43:19.080761575Z" level=info msg="CreateContainer within sandbox \"0c1b7d9084bf65a480bc269a2ac9ce9a9a8ad107c0fc97b1e89a3bb3999302e6\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 13 23:43:19.086542 containerd[1799]: time="2025-05-13T23:43:19.086510988Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jtz2z,Uid:ca8a8273-b2db-4a64-922e-a6200979faee,Namespace:kube-system,Attempt:0,} returns sandbox id \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\"" May 13 23:43:19.133434 containerd[1799]: time="2025-05-13T23:43:19.133387008Z" level=info msg="Container 4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:19.163052 containerd[1799]: time="2025-05-13T23:43:19.163004751Z" level=info msg="CreateContainer within sandbox \"0c1b7d9084bf65a480bc269a2ac9ce9a9a8ad107c0fc97b1e89a3bb3999302e6\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15\"" May 13 23:43:19.164517 containerd[1799]: time="2025-05-13T23:43:19.164300314Z" level=info msg="StartContainer for \"4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15\"" May 13 23:43:19.166539 containerd[1799]: time="2025-05-13T23:43:19.166506559Z" level=info msg="connecting to shim 4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15" address="unix:///run/containerd/s/148592215ec6a389458cc912acad4785dcab2c3156643efc1dc4c9283f1cc811" protocol=ttrpc version=3 May 13 23:43:19.182988 systemd[1]: Started cri-containerd-4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15.scope - libcontainer container 4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15. May 13 23:43:19.231346 containerd[1799]: time="2025-05-13T23:43:19.231300897Z" level=info msg="StartContainer for \"4aa76a4ab17d0164185f63f475b5adeccc81db9718cd5704620a7f49aa1a6b15\" returns successfully" May 13 23:43:20.767646 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount694978394.mount: Deactivated successfully. May 13 23:43:21.285092 containerd[1799]: time="2025-05-13T23:43:21.284389363Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:43:21.292504 containerd[1799]: time="2025-05-13T23:43:21.292448937Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" May 13 23:43:21.297890 containerd[1799]: time="2025-05-13T23:43:21.297859827Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:43:21.299623 containerd[1799]: time="2025-05-13T23:43:21.299591790Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 2.452459714s" May 13 23:43:21.299708 containerd[1799]: time="2025-05-13T23:43:21.299625710Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" May 13 23:43:21.300859 containerd[1799]: time="2025-05-13T23:43:21.300807192Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 13 23:43:21.302042 containerd[1799]: time="2025-05-13T23:43:21.301902634Z" level=info msg="CreateContainer within sandbox \"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 13 23:43:21.335972 containerd[1799]: time="2025-05-13T23:43:21.335924614Z" level=info msg="Container 2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:21.339852 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount604414428.mount: Deactivated successfully. May 13 23:43:21.368436 containerd[1799]: time="2025-05-13T23:43:21.367399510Z" level=info msg="CreateContainer within sandbox \"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\"" May 13 23:43:21.369350 containerd[1799]: time="2025-05-13T23:43:21.369235273Z" level=info msg="StartContainer for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\"" May 13 23:43:21.370810 containerd[1799]: time="2025-05-13T23:43:21.370784836Z" level=info msg="connecting to shim 2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f" address="unix:///run/containerd/s/2ba2c3739628916c84bfa229306b218dc765249846a84391ec831be8d5048e59" protocol=ttrpc version=3 May 13 23:43:21.390974 systemd[1]: Started cri-containerd-2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f.scope - libcontainer container 2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f. May 13 23:43:21.420354 containerd[1799]: time="2025-05-13T23:43:21.420311204Z" level=info msg="StartContainer for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" returns successfully" May 13 23:43:21.549729 kubelet[3344]: I0513 23:43:21.548487 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-s2wp5" podStartSLOduration=3.548470951 podStartE2EDuration="3.548470951s" podCreationTimestamp="2025-05-13 23:43:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:43:19.530514977 +0000 UTC m=+15.195381779" watchObservedRunningTime="2025-05-13 23:43:21.548470951 +0000 UTC m=+17.213337753" May 13 23:43:25.293317 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2406002818.mount: Deactivated successfully. May 13 23:43:27.323382 containerd[1799]: time="2025-05-13T23:43:27.323337792Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:43:27.327509 containerd[1799]: time="2025-05-13T23:43:27.327463680Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" May 13 23:43:27.334381 containerd[1799]: time="2025-05-13T23:43:27.334334332Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" May 13 23:43:27.335844 containerd[1799]: time="2025-05-13T23:43:27.335715255Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 6.034748583s" May 13 23:43:27.335844 containerd[1799]: time="2025-05-13T23:43:27.335749015Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" May 13 23:43:27.338846 containerd[1799]: time="2025-05-13T23:43:27.338781220Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 13 23:43:27.367697 containerd[1799]: time="2025-05-13T23:43:27.366852472Z" level=info msg="Container 0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:27.384801 containerd[1799]: time="2025-05-13T23:43:27.384754225Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\"" May 13 23:43:27.385409 containerd[1799]: time="2025-05-13T23:43:27.385373546Z" level=info msg="StartContainer for \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\"" May 13 23:43:27.386849 containerd[1799]: time="2025-05-13T23:43:27.386609549Z" level=info msg="connecting to shim 0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304" address="unix:///run/containerd/s/1ae4c7d50d628953eaa40a240cd95e556119175b6292c0e7b7dc6c84bacff4f5" protocol=ttrpc version=3 May 13 23:43:27.411981 systemd[1]: Started cri-containerd-0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304.scope - libcontainer container 0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304. May 13 23:43:27.451751 containerd[1799]: time="2025-05-13T23:43:27.451416748Z" level=info msg="StartContainer for \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" returns successfully" May 13 23:43:27.458118 systemd[1]: cri-containerd-0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304.scope: Deactivated successfully. May 13 23:43:27.461166 containerd[1799]: time="2025-05-13T23:43:27.461129086Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" id:\"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" pid:3799 exited_at:{seconds:1747179807 nanos:460362725}" May 13 23:43:27.461274 containerd[1799]: time="2025-05-13T23:43:27.461258206Z" level=info msg="received exit event container_id:\"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" id:\"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" pid:3799 exited_at:{seconds:1747179807 nanos:460362725}" May 13 23:43:27.480671 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304-rootfs.mount: Deactivated successfully. May 13 23:43:27.555175 kubelet[3344]: I0513 23:43:27.555104 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-r7bdk" podStartSLOduration=7.100750821 podStartE2EDuration="9.555088659s" podCreationTimestamp="2025-05-13 23:43:18 +0000 UTC" firstStartedPulling="2025-05-13 23:43:18.846195234 +0000 UTC m=+14.511061996" lastFinishedPulling="2025-05-13 23:43:21.300533032 +0000 UTC m=+16.965399834" observedRunningTime="2025-05-13 23:43:21.560621412 +0000 UTC m=+17.225488214" watchObservedRunningTime="2025-05-13 23:43:27.555088659 +0000 UTC m=+23.219955421" May 13 23:43:29.546343 containerd[1799]: time="2025-05-13T23:43:29.546295090Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 13 23:43:29.583780 containerd[1799]: time="2025-05-13T23:43:29.582552957Z" level=info msg="Container 0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:29.587118 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2497348073.mount: Deactivated successfully. May 13 23:43:29.601614 containerd[1799]: time="2025-05-13T23:43:29.601499992Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\"" May 13 23:43:29.602168 containerd[1799]: time="2025-05-13T23:43:29.602124993Z" level=info msg="StartContainer for \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\"" May 13 23:43:29.603498 containerd[1799]: time="2025-05-13T23:43:29.603474556Z" level=info msg="connecting to shim 0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f" address="unix:///run/containerd/s/1ae4c7d50d628953eaa40a240cd95e556119175b6292c0e7b7dc6c84bacff4f5" protocol=ttrpc version=3 May 13 23:43:29.624974 systemd[1]: Started cri-containerd-0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f.scope - libcontainer container 0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f. May 13 23:43:29.654740 containerd[1799]: time="2025-05-13T23:43:29.654630210Z" level=info msg="StartContainer for \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" returns successfully" May 13 23:43:29.663112 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 13 23:43:29.663454 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. May 13 23:43:29.663711 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... May 13 23:43:29.667149 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... May 13 23:43:29.668893 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 13 23:43:29.669253 systemd[1]: cri-containerd-0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f.scope: Deactivated successfully. May 13 23:43:29.672997 containerd[1799]: time="2025-05-13T23:43:29.672786564Z" level=info msg="received exit event container_id:\"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" id:\"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" pid:3844 exited_at:{seconds:1747179809 nanos:671963402}" May 13 23:43:29.673103 containerd[1799]: time="2025-05-13T23:43:29.673034244Z" level=info msg="TaskExit event in podsandbox handler container_id:\"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" id:\"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" pid:3844 exited_at:{seconds:1747179809 nanos:671963402}" May 13 23:43:29.692976 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. May 13 23:43:30.548603 containerd[1799]: time="2025-05-13T23:43:30.548547418Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 13 23:43:30.582029 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f-rootfs.mount: Deactivated successfully. May 13 23:43:30.595465 containerd[1799]: time="2025-05-13T23:43:30.592926460Z" level=info msg="Container a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:30.615177 containerd[1799]: time="2025-05-13T23:43:30.615138701Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\"" May 13 23:43:30.615836 containerd[1799]: time="2025-05-13T23:43:30.615782542Z" level=info msg="StartContainer for \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\"" May 13 23:43:30.618250 containerd[1799]: time="2025-05-13T23:43:30.618217947Z" level=info msg="connecting to shim a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49" address="unix:///run/containerd/s/1ae4c7d50d628953eaa40a240cd95e556119175b6292c0e7b7dc6c84bacff4f5" protocol=ttrpc version=3 May 13 23:43:30.637965 systemd[1]: Started cri-containerd-a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49.scope - libcontainer container a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49. May 13 23:43:30.673105 systemd[1]: cri-containerd-a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49.scope: Deactivated successfully. May 13 23:43:30.674870 containerd[1799]: time="2025-05-13T23:43:30.674798571Z" level=info msg="TaskExit event in podsandbox handler container_id:\"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" id:\"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" pid:3890 exited_at:{seconds:1747179810 nanos:674580331}" May 13 23:43:30.676943 containerd[1799]: time="2025-05-13T23:43:30.676851095Z" level=info msg="received exit event container_id:\"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" id:\"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" pid:3890 exited_at:{seconds:1747179810 nanos:674580331}" May 13 23:43:30.683362 containerd[1799]: time="2025-05-13T23:43:30.683320027Z" level=info msg="StartContainer for \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" returns successfully" May 13 23:43:30.701278 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49-rootfs.mount: Deactivated successfully. May 13 23:43:31.552931 containerd[1799]: time="2025-05-13T23:43:31.552892167Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 13 23:43:31.588308 containerd[1799]: time="2025-05-13T23:43:31.587330970Z" level=info msg="Container ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:31.611212 containerd[1799]: time="2025-05-13T23:43:31.611165988Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\"" May 13 23:43:31.612651 containerd[1799]: time="2025-05-13T23:43:31.612439111Z" level=info msg="StartContainer for \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\"" May 13 23:43:31.613495 containerd[1799]: time="2025-05-13T23:43:31.613464913Z" level=info msg="connecting to shim ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9" address="unix:///run/containerd/s/1ae4c7d50d628953eaa40a240cd95e556119175b6292c0e7b7dc6c84bacff4f5" protocol=ttrpc version=3 May 13 23:43:31.629975 systemd[1]: Started cri-containerd-ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9.scope - libcontainer container ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9. May 13 23:43:31.654339 systemd[1]: cri-containerd-ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9.scope: Deactivated successfully. May 13 23:43:31.655549 containerd[1799]: time="2025-05-13T23:43:31.655407655Z" level=info msg="TaskExit event in podsandbox handler container_id:\"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" id:\"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" pid:3930 exited_at:{seconds:1747179811 nanos:654433212}" May 13 23:43:31.664200 containerd[1799]: time="2025-05-13T23:43:31.664075596Z" level=info msg="received exit event container_id:\"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" id:\"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" pid:3930 exited_at:{seconds:1747179811 nanos:654433212}" May 13 23:43:31.667801 containerd[1799]: time="2025-05-13T23:43:31.666977283Z" level=info msg="StartContainer for \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" returns successfully" May 13 23:43:31.683692 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9-rootfs.mount: Deactivated successfully. May 13 23:43:32.560833 containerd[1799]: time="2025-05-13T23:43:32.560286281Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 13 23:43:32.600136 containerd[1799]: time="2025-05-13T23:43:32.600092897Z" level=info msg="Container 187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:32.604728 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount84813773.mount: Deactivated successfully. May 13 23:43:32.626270 containerd[1799]: time="2025-05-13T23:43:32.626225120Z" level=info msg="CreateContainer within sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\"" May 13 23:43:32.628216 containerd[1799]: time="2025-05-13T23:43:32.627443683Z" level=info msg="StartContainer for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\"" May 13 23:43:32.628412 containerd[1799]: time="2025-05-13T23:43:32.628379405Z" level=info msg="connecting to shim 187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3" address="unix:///run/containerd/s/1ae4c7d50d628953eaa40a240cd95e556119175b6292c0e7b7dc6c84bacff4f5" protocol=ttrpc version=3 May 13 23:43:32.649989 systemd[1]: Started cri-containerd-187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3.scope - libcontainer container 187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3. May 13 23:43:32.686537 containerd[1799]: time="2025-05-13T23:43:32.686461546Z" level=info msg="StartContainer for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" returns successfully" May 13 23:43:32.743889 containerd[1799]: time="2025-05-13T23:43:32.743785764Z" level=info msg="TaskExit event in podsandbox handler container_id:\"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" id:\"bdea08f6d17cf5c07b42e31e6f5b398781bff300bb0ead31ecfb97b3b43a4115\" pid:3997 exited_at:{seconds:1747179812 nanos:743513523}" May 13 23:43:32.846380 kubelet[3344]: I0513 23:43:32.845448 3344 kubelet_node_status.go:497] "Fast updating node status as it just became ready" May 13 23:43:32.885402 kubelet[3344]: I0513 23:43:32.884370 3344 topology_manager.go:215] "Topology Admit Handler" podUID="f1495c6c-2079-4ea1-bf3a-a24048d8c525" podNamespace="kube-system" podName="coredns-7db6d8ff4d-2zldt" May 13 23:43:32.896865 kubelet[3344]: I0513 23:43:32.893454 3344 topology_manager.go:215] "Topology Admit Handler" podUID="fe34d60e-3093-44e8-ad4c-bc48701a180d" podNamespace="kube-system" podName="coredns-7db6d8ff4d-mzlsw" May 13 23:43:32.895149 systemd[1]: Created slice kubepods-burstable-podf1495c6c_2079_4ea1_bf3a_a24048d8c525.slice - libcontainer container kubepods-burstable-podf1495c6c_2079_4ea1_bf3a_a24048d8c525.slice. May 13 23:43:32.905972 systemd[1]: Created slice kubepods-burstable-podfe34d60e_3093_44e8_ad4c_bc48701a180d.slice - libcontainer container kubepods-burstable-podfe34d60e_3093_44e8_ad4c_bc48701a180d.slice. May 13 23:43:32.972001 kubelet[3344]: I0513 23:43:32.971685 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r446\" (UniqueName: \"kubernetes.io/projected/fe34d60e-3093-44e8-ad4c-bc48701a180d-kube-api-access-7r446\") pod \"coredns-7db6d8ff4d-mzlsw\" (UID: \"fe34d60e-3093-44e8-ad4c-bc48701a180d\") " pod="kube-system/coredns-7db6d8ff4d-mzlsw" May 13 23:43:32.972001 kubelet[3344]: I0513 23:43:32.971801 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f1495c6c-2079-4ea1-bf3a-a24048d8c525-config-volume\") pod \"coredns-7db6d8ff4d-2zldt\" (UID: \"f1495c6c-2079-4ea1-bf3a-a24048d8c525\") " pod="kube-system/coredns-7db6d8ff4d-2zldt" May 13 23:43:32.972001 kubelet[3344]: I0513 23:43:32.971865 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wrzh\" (UniqueName: \"kubernetes.io/projected/f1495c6c-2079-4ea1-bf3a-a24048d8c525-kube-api-access-8wrzh\") pod \"coredns-7db6d8ff4d-2zldt\" (UID: \"f1495c6c-2079-4ea1-bf3a-a24048d8c525\") " pod="kube-system/coredns-7db6d8ff4d-2zldt" May 13 23:43:32.972001 kubelet[3344]: I0513 23:43:32.971889 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fe34d60e-3093-44e8-ad4c-bc48701a180d-config-volume\") pod \"coredns-7db6d8ff4d-mzlsw\" (UID: \"fe34d60e-3093-44e8-ad4c-bc48701a180d\") " pod="kube-system/coredns-7db6d8ff4d-mzlsw" May 13 23:43:33.200607 containerd[1799]: time="2025-05-13T23:43:33.200556467Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-2zldt,Uid:f1495c6c-2079-4ea1-bf3a-a24048d8c525,Namespace:kube-system,Attempt:0,}" May 13 23:43:33.210848 containerd[1799]: time="2025-05-13T23:43:33.210471291Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-mzlsw,Uid:fe34d60e-3093-44e8-ad4c-bc48701a180d,Namespace:kube-system,Attempt:0,}" May 13 23:43:33.581328 kubelet[3344]: I0513 23:43:33.580868 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-jtz2z" podStartSLOduration=7.332179801 podStartE2EDuration="15.580853626s" podCreationTimestamp="2025-05-13 23:43:18 +0000 UTC" firstStartedPulling="2025-05-13 23:43:19.087991511 +0000 UTC m=+14.752858273" lastFinishedPulling="2025-05-13 23:43:27.336665296 +0000 UTC m=+23.001532098" observedRunningTime="2025-05-13 23:43:33.580537745 +0000 UTC m=+29.245404547" watchObservedRunningTime="2025-05-13 23:43:33.580853626 +0000 UTC m=+29.245720428" May 13 23:43:34.769520 systemd-networkd[1353]: cilium_host: Link UP May 13 23:43:34.769656 systemd-networkd[1353]: cilium_net: Link UP May 13 23:43:34.769777 systemd-networkd[1353]: cilium_net: Gained carrier May 13 23:43:34.769914 systemd-networkd[1353]: cilium_host: Gained carrier May 13 23:43:34.899501 systemd-networkd[1353]: cilium_vxlan: Link UP May 13 23:43:34.899513 systemd-networkd[1353]: cilium_vxlan: Gained carrier May 13 23:43:35.132096 kernel: NET: Registered PF_ALG protocol family May 13 23:43:35.352962 systemd-networkd[1353]: cilium_net: Gained IPv6LL May 13 23:43:35.607980 systemd-networkd[1353]: cilium_host: Gained IPv6LL May 13 23:43:35.825288 systemd-networkd[1353]: lxc_health: Link UP May 13 23:43:35.837934 systemd-networkd[1353]: lxc_health: Gained carrier May 13 23:43:36.234710 systemd-networkd[1353]: lxcfd6f1aafd804: Link UP May 13 23:43:36.244878 kernel: eth0: renamed from tmp8a38f May 13 23:43:36.253101 systemd-networkd[1353]: lxcfd6f1aafd804: Gained carrier May 13 23:43:36.263433 systemd-networkd[1353]: lxcc041663a7551: Link UP May 13 23:43:36.272904 kernel: eth0: renamed from tmpa6faf May 13 23:43:36.277465 systemd-networkd[1353]: lxcc041663a7551: Gained carrier May 13 23:43:36.696973 systemd-networkd[1353]: cilium_vxlan: Gained IPv6LL May 13 23:43:37.016956 systemd-networkd[1353]: lxc_health: Gained IPv6LL May 13 23:43:37.528956 systemd-networkd[1353]: lxcc041663a7551: Gained IPv6LL May 13 23:43:37.848933 systemd-networkd[1353]: lxcfd6f1aafd804: Gained IPv6LL May 13 23:43:39.831196 containerd[1799]: time="2025-05-13T23:43:39.831130231Z" level=info msg="connecting to shim 8a38f6de771b13e613fd1c589d8383c3fb51b31a6fde198c65b88be129a9a365" address="unix:///run/containerd/s/2ce9e0033a99fd04fab543d56d6450372c78da031574f51f7950a6c54b727dd3" namespace=k8s.io protocol=ttrpc version=3 May 13 23:43:39.861592 containerd[1799]: time="2025-05-13T23:43:39.861553172Z" level=info msg="connecting to shim a6faf000a184ec0b7f8a592cac265a06526160ce1adf742bc7243fb1d6b47709" address="unix:///run/containerd/s/5927ab290f0aa9adf4b51dd394d0e75792d0790da902582f5aec4813f22eed9a" namespace=k8s.io protocol=ttrpc version=3 May 13 23:43:39.863995 systemd[1]: Started cri-containerd-8a38f6de771b13e613fd1c589d8383c3fb51b31a6fde198c65b88be129a9a365.scope - libcontainer container 8a38f6de771b13e613fd1c589d8383c3fb51b31a6fde198c65b88be129a9a365. May 13 23:43:39.895020 systemd[1]: Started cri-containerd-a6faf000a184ec0b7f8a592cac265a06526160ce1adf742bc7243fb1d6b47709.scope - libcontainer container a6faf000a184ec0b7f8a592cac265a06526160ce1adf742bc7243fb1d6b47709. May 13 23:43:39.923688 containerd[1799]: time="2025-05-13T23:43:39.923360495Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-2zldt,Uid:f1495c6c-2079-4ea1-bf3a-a24048d8c525,Namespace:kube-system,Attempt:0,} returns sandbox id \"8a38f6de771b13e613fd1c589d8383c3fb51b31a6fde198c65b88be129a9a365\"" May 13 23:43:39.943633 containerd[1799]: time="2025-05-13T23:43:39.943558775Z" level=info msg="CreateContainer within sandbox \"8a38f6de771b13e613fd1c589d8383c3fb51b31a6fde198c65b88be129a9a365\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 13 23:43:39.976581 containerd[1799]: time="2025-05-13T23:43:39.976537560Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-mzlsw,Uid:fe34d60e-3093-44e8-ad4c-bc48701a180d,Namespace:kube-system,Attempt:0,} returns sandbox id \"a6faf000a184ec0b7f8a592cac265a06526160ce1adf742bc7243fb1d6b47709\"" May 13 23:43:39.979774 containerd[1799]: time="2025-05-13T23:43:39.979488286Z" level=info msg="CreateContainer within sandbox \"a6faf000a184ec0b7f8a592cac265a06526160ce1adf742bc7243fb1d6b47709\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 13 23:43:39.987568 containerd[1799]: time="2025-05-13T23:43:39.987523542Z" level=info msg="Container 240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:40.032042 containerd[1799]: time="2025-05-13T23:43:40.031983471Z" level=info msg="CreateContainer within sandbox \"8a38f6de771b13e613fd1c589d8383c3fb51b31a6fde198c65b88be129a9a365\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f\"" May 13 23:43:40.034140 containerd[1799]: time="2025-05-13T23:43:40.033973155Z" level=info msg="StartContainer for \"240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f\"" May 13 23:43:40.036386 containerd[1799]: time="2025-05-13T23:43:40.036358959Z" level=info msg="connecting to shim 240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f" address="unix:///run/containerd/s/2ce9e0033a99fd04fab543d56d6450372c78da031574f51f7950a6c54b727dd3" protocol=ttrpc version=3 May 13 23:43:40.040786 containerd[1799]: time="2025-05-13T23:43:40.039706766Z" level=info msg="Container 9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af: CDI devices from CRI Config.CDIDevices: []" May 13 23:43:40.054984 systemd[1]: Started cri-containerd-240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f.scope - libcontainer container 240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f. May 13 23:43:40.073354 containerd[1799]: time="2025-05-13T23:43:40.073142952Z" level=info msg="CreateContainer within sandbox \"a6faf000a184ec0b7f8a592cac265a06526160ce1adf742bc7243fb1d6b47709\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af\"" May 13 23:43:40.076869 containerd[1799]: time="2025-05-13T23:43:40.074594995Z" level=info msg="StartContainer for \"9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af\"" May 13 23:43:40.076869 containerd[1799]: time="2025-05-13T23:43:40.075413277Z" level=info msg="connecting to shim 9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af" address="unix:///run/containerd/s/5927ab290f0aa9adf4b51dd394d0e75792d0790da902582f5aec4813f22eed9a" protocol=ttrpc version=3 May 13 23:43:40.094335 containerd[1799]: time="2025-05-13T23:43:40.093420713Z" level=info msg="StartContainer for \"240b5ec0b44e6e8ae79f1ced0b894adec7457fb81ae8c81832a001cf0a068d8f\" returns successfully" May 13 23:43:40.102175 systemd[1]: Started cri-containerd-9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af.scope - libcontainer container 9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af. May 13 23:43:40.150363 containerd[1799]: time="2025-05-13T23:43:40.150321106Z" level=info msg="StartContainer for \"9229e96d8ffb7b79a423deef84c805fdcaffc4ac5167767901db36ad7e67c0af\" returns successfully" May 13 23:43:40.598289 kubelet[3344]: I0513 23:43:40.597930 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-2zldt" podStartSLOduration=22.597912635 podStartE2EDuration="22.597912635s" podCreationTimestamp="2025-05-13 23:43:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:43:40.597295674 +0000 UTC m=+36.262162476" watchObservedRunningTime="2025-05-13 23:43:40.597912635 +0000 UTC m=+36.262779477" May 13 23:43:40.636744 kubelet[3344]: I0513 23:43:40.636675 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-mzlsw" podStartSLOduration=22.636348072 podStartE2EDuration="22.636348072s" podCreationTimestamp="2025-05-13 23:43:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:43:40.63563763 +0000 UTC m=+36.300504432" watchObservedRunningTime="2025-05-13 23:43:40.636348072 +0000 UTC m=+36.301214874" May 13 23:43:46.104699 kubelet[3344]: I0513 23:43:46.104538 3344 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" May 13 23:44:45.306853 systemd[1]: Started sshd@7-10.200.20.30:22-10.200.16.10:50340.service - OpenSSH per-connection server daemon (10.200.16.10:50340). May 13 23:44:45.802465 sshd[4655]: Accepted publickey for core from 10.200.16.10 port 50340 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:44:45.803841 sshd-session[4655]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:44:45.808939 systemd-logind[1744]: New session 10 of user core. May 13 23:44:45.815972 systemd[1]: Started session-10.scope - Session 10 of User core. May 13 23:44:46.240607 sshd[4657]: Connection closed by 10.200.16.10 port 50340 May 13 23:44:46.241183 sshd-session[4655]: pam_unix(sshd:session): session closed for user core May 13 23:44:46.244610 systemd[1]: sshd@7-10.200.20.30:22-10.200.16.10:50340.service: Deactivated successfully. May 13 23:44:46.246637 systemd[1]: session-10.scope: Deactivated successfully. May 13 23:44:46.247524 systemd-logind[1744]: Session 10 logged out. Waiting for processes to exit. May 13 23:44:46.248391 systemd-logind[1744]: Removed session 10. May 13 23:44:51.332170 systemd[1]: Started sshd@8-10.200.20.30:22-10.200.16.10:39610.service - OpenSSH per-connection server daemon (10.200.16.10:39610). May 13 23:44:51.821589 sshd[4674]: Accepted publickey for core from 10.200.16.10 port 39610 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:44:51.822784 sshd-session[4674]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:44:51.828031 systemd-logind[1744]: New session 11 of user core. May 13 23:44:51.838027 systemd[1]: Started session-11.scope - Session 11 of User core. May 13 23:44:52.243678 sshd[4676]: Connection closed by 10.200.16.10 port 39610 May 13 23:44:52.242846 sshd-session[4674]: pam_unix(sshd:session): session closed for user core May 13 23:44:52.246244 systemd[1]: sshd@8-10.200.20.30:22-10.200.16.10:39610.service: Deactivated successfully. May 13 23:44:52.248127 systemd[1]: session-11.scope: Deactivated successfully. May 13 23:44:52.248884 systemd-logind[1744]: Session 11 logged out. Waiting for processes to exit. May 13 23:44:52.250247 systemd-logind[1744]: Removed session 11. May 13 23:44:57.328105 systemd[1]: Started sshd@9-10.200.20.30:22-10.200.16.10:39616.service - OpenSSH per-connection server daemon (10.200.16.10:39616). May 13 23:44:57.783317 sshd[4689]: Accepted publickey for core from 10.200.16.10 port 39616 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:44:57.784583 sshd-session[4689]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:44:57.789360 systemd-logind[1744]: New session 12 of user core. May 13 23:44:57.792201 systemd[1]: Started session-12.scope - Session 12 of User core. May 13 23:44:58.165441 sshd[4691]: Connection closed by 10.200.16.10 port 39616 May 13 23:44:58.166275 sshd-session[4689]: pam_unix(sshd:session): session closed for user core May 13 23:44:58.169395 systemd[1]: sshd@9-10.200.20.30:22-10.200.16.10:39616.service: Deactivated successfully. May 13 23:44:58.172063 systemd[1]: session-12.scope: Deactivated successfully. May 13 23:44:58.173004 systemd-logind[1744]: Session 12 logged out. Waiting for processes to exit. May 13 23:44:58.174807 systemd-logind[1744]: Removed session 12. May 13 23:45:03.257467 systemd[1]: Started sshd@10-10.200.20.30:22-10.200.16.10:35252.service - OpenSSH per-connection server daemon (10.200.16.10:35252). May 13 23:45:03.752283 sshd[4704]: Accepted publickey for core from 10.200.16.10 port 35252 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:03.753602 sshd-session[4704]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:03.757595 systemd-logind[1744]: New session 13 of user core. May 13 23:45:03.764024 systemd[1]: Started session-13.scope - Session 13 of User core. May 13 23:45:04.173630 sshd[4706]: Connection closed by 10.200.16.10 port 35252 May 13 23:45:04.174186 sshd-session[4704]: pam_unix(sshd:session): session closed for user core May 13 23:45:04.177575 systemd-logind[1744]: Session 13 logged out. Waiting for processes to exit. May 13 23:45:04.178339 systemd[1]: sshd@10-10.200.20.30:22-10.200.16.10:35252.service: Deactivated successfully. May 13 23:45:04.180709 systemd[1]: session-13.scope: Deactivated successfully. May 13 23:45:04.182152 systemd-logind[1744]: Removed session 13. May 13 23:45:05.466373 update_engine[1751]: I20250513 23:45:05.465930 1751 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs May 13 23:45:05.466373 update_engine[1751]: I20250513 23:45:05.465981 1751 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs May 13 23:45:05.466373 update_engine[1751]: I20250513 23:45:05.466140 1751 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466526 1751 omaha_request_params.cc:62] Current group set to alpha May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466614 1751 update_attempter.cc:499] Already updated boot flags. Skipping. May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466622 1751 update_attempter.cc:643] Scheduling an action processor start. May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466637 1751 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466666 1751 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466710 1751 omaha_request_action.cc:271] Posting an Omaha request to disabled May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466717 1751 omaha_request_action.cc:272] Request: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: May 13 23:45:05.466731 update_engine[1751]: I20250513 23:45:05.466723 1751 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 13 23:45:05.467631 locksmithd[1849]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 May 13 23:45:05.467925 update_engine[1751]: I20250513 23:45:05.467890 1751 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 13 23:45:05.468272 update_engine[1751]: I20250513 23:45:05.468235 1751 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 13 23:45:05.650458 update_engine[1751]: E20250513 23:45:05.650398 1751 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 13 23:45:05.650594 update_engine[1751]: I20250513 23:45:05.650506 1751 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 May 13 23:45:09.270684 systemd[1]: Started sshd@11-10.200.20.30:22-10.200.16.10:51174.service - OpenSSH per-connection server daemon (10.200.16.10:51174). May 13 23:45:09.728314 sshd[4720]: Accepted publickey for core from 10.200.16.10 port 51174 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:09.729597 sshd-session[4720]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:09.735116 systemd-logind[1744]: New session 14 of user core. May 13 23:45:09.742038 systemd[1]: Started session-14.scope - Session 14 of User core. May 13 23:45:10.112565 sshd[4722]: Connection closed by 10.200.16.10 port 51174 May 13 23:45:10.113050 sshd-session[4720]: pam_unix(sshd:session): session closed for user core May 13 23:45:10.116560 systemd-logind[1744]: Session 14 logged out. Waiting for processes to exit. May 13 23:45:10.117167 systemd[1]: sshd@11-10.200.20.30:22-10.200.16.10:51174.service: Deactivated successfully. May 13 23:45:10.119475 systemd[1]: session-14.scope: Deactivated successfully. May 13 23:45:10.120681 systemd-logind[1744]: Removed session 14. May 13 23:45:15.201196 systemd[1]: Started sshd@12-10.200.20.30:22-10.200.16.10:51190.service - OpenSSH per-connection server daemon (10.200.16.10:51190). May 13 23:45:15.470221 update_engine[1751]: I20250513 23:45:15.470077 1751 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 13 23:45:15.470538 update_engine[1751]: I20250513 23:45:15.470431 1751 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 13 23:45:15.470720 update_engine[1751]: I20250513 23:45:15.470682 1751 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 13 23:45:15.509652 update_engine[1751]: E20250513 23:45:15.509600 1751 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 13 23:45:15.509760 update_engine[1751]: I20250513 23:45:15.509680 1751 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 May 13 23:45:15.690144 sshd[4734]: Accepted publickey for core from 10.200.16.10 port 51190 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:15.691461 sshd-session[4734]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:15.695918 systemd-logind[1744]: New session 15 of user core. May 13 23:45:15.700972 systemd[1]: Started session-15.scope - Session 15 of User core. May 13 23:45:16.110777 sshd[4736]: Connection closed by 10.200.16.10 port 51190 May 13 23:45:16.111341 sshd-session[4734]: pam_unix(sshd:session): session closed for user core May 13 23:45:16.115081 systemd-logind[1744]: Session 15 logged out. Waiting for processes to exit. May 13 23:45:16.115372 systemd[1]: sshd@12-10.200.20.30:22-10.200.16.10:51190.service: Deactivated successfully. May 13 23:45:16.117643 systemd[1]: session-15.scope: Deactivated successfully. May 13 23:45:16.118860 systemd-logind[1744]: Removed session 15. May 13 23:45:21.201260 systemd[1]: Started sshd@13-10.200.20.30:22-10.200.16.10:33984.service - OpenSSH per-connection server daemon (10.200.16.10:33984). May 13 23:45:21.700426 sshd[4751]: Accepted publickey for core from 10.200.16.10 port 33984 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:21.701737 sshd-session[4751]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:21.706424 systemd-logind[1744]: New session 16 of user core. May 13 23:45:21.712988 systemd[1]: Started session-16.scope - Session 16 of User core. May 13 23:45:22.126147 sshd[4753]: Connection closed by 10.200.16.10 port 33984 May 13 23:45:22.126672 sshd-session[4751]: pam_unix(sshd:session): session closed for user core May 13 23:45:22.130037 systemd[1]: sshd@13-10.200.20.30:22-10.200.16.10:33984.service: Deactivated successfully. May 13 23:45:22.131636 systemd[1]: session-16.scope: Deactivated successfully. May 13 23:45:22.133372 systemd-logind[1744]: Session 16 logged out. Waiting for processes to exit. May 13 23:45:22.134883 systemd-logind[1744]: Removed session 16. May 13 23:45:22.213501 systemd[1]: Started sshd@14-10.200.20.30:22-10.200.16.10:33992.service - OpenSSH per-connection server daemon (10.200.16.10:33992). May 13 23:45:22.706155 sshd[4766]: Accepted publickey for core from 10.200.16.10 port 33992 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:22.707496 sshd-session[4766]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:22.711964 systemd-logind[1744]: New session 17 of user core. May 13 23:45:22.716953 systemd[1]: Started session-17.scope - Session 17 of User core. May 13 23:45:23.166955 sshd[4768]: Connection closed by 10.200.16.10 port 33992 May 13 23:45:23.167640 sshd-session[4766]: pam_unix(sshd:session): session closed for user core May 13 23:45:23.171100 systemd[1]: sshd@14-10.200.20.30:22-10.200.16.10:33992.service: Deactivated successfully. May 13 23:45:23.173322 systemd[1]: session-17.scope: Deactivated successfully. May 13 23:45:23.174769 systemd-logind[1744]: Session 17 logged out. Waiting for processes to exit. May 13 23:45:23.175950 systemd-logind[1744]: Removed session 17. May 13 23:45:23.255440 systemd[1]: Started sshd@15-10.200.20.30:22-10.200.16.10:33996.service - OpenSSH per-connection server daemon (10.200.16.10:33996). May 13 23:45:23.747965 sshd[4778]: Accepted publickey for core from 10.200.16.10 port 33996 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:23.750311 sshd-session[4778]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:23.754409 systemd-logind[1744]: New session 18 of user core. May 13 23:45:23.760980 systemd[1]: Started session-18.scope - Session 18 of User core. May 13 23:45:24.169502 sshd[4780]: Connection closed by 10.200.16.10 port 33996 May 13 23:45:24.170049 sshd-session[4778]: pam_unix(sshd:session): session closed for user core May 13 23:45:24.173451 systemd[1]: sshd@15-10.200.20.30:22-10.200.16.10:33996.service: Deactivated successfully. May 13 23:45:24.175064 systemd[1]: session-18.scope: Deactivated successfully. May 13 23:45:24.175915 systemd-logind[1744]: Session 18 logged out. Waiting for processes to exit. May 13 23:45:24.177024 systemd-logind[1744]: Removed session 18. May 13 23:45:25.465595 update_engine[1751]: I20250513 23:45:25.465178 1751 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 13 23:45:25.465595 update_engine[1751]: I20250513 23:45:25.465435 1751 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 13 23:45:25.465986 update_engine[1751]: I20250513 23:45:25.465674 1751 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 13 23:45:25.582068 update_engine[1751]: E20250513 23:45:25.582011 1751 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 13 23:45:25.582195 update_engine[1751]: I20250513 23:45:25.582099 1751 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 May 13 23:45:29.261313 systemd[1]: Started sshd@16-10.200.20.30:22-10.200.16.10:53668.service - OpenSSH per-connection server daemon (10.200.16.10:53668). May 13 23:45:29.751073 sshd[4792]: Accepted publickey for core from 10.200.16.10 port 53668 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:29.752429 sshd-session[4792]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:29.757954 systemd-logind[1744]: New session 19 of user core. May 13 23:45:29.766957 systemd[1]: Started session-19.scope - Session 19 of User core. May 13 23:45:30.167140 sshd[4794]: Connection closed by 10.200.16.10 port 53668 May 13 23:45:30.167740 sshd-session[4792]: pam_unix(sshd:session): session closed for user core May 13 23:45:30.171555 systemd[1]: sshd@16-10.200.20.30:22-10.200.16.10:53668.service: Deactivated successfully. May 13 23:45:30.174230 systemd[1]: session-19.scope: Deactivated successfully. May 13 23:45:30.175685 systemd-logind[1744]: Session 19 logged out. Waiting for processes to exit. May 13 23:45:30.176698 systemd-logind[1744]: Removed session 19. May 13 23:45:35.263409 systemd[1]: Started sshd@17-10.200.20.30:22-10.200.16.10:53678.service - OpenSSH per-connection server daemon (10.200.16.10:53678). May 13 23:45:35.466712 update_engine[1751]: I20250513 23:45:35.466650 1751 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 13 23:45:35.467119 update_engine[1751]: I20250513 23:45:35.466912 1751 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 13 23:45:35.467157 update_engine[1751]: I20250513 23:45:35.467139 1751 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 13 23:45:35.494250 update_engine[1751]: E20250513 23:45:35.494208 1751 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 13 23:45:35.494336 update_engine[1751]: I20250513 23:45:35.494274 1751 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded May 13 23:45:35.494336 update_engine[1751]: I20250513 23:45:35.494283 1751 omaha_request_action.cc:617] Omaha request response: May 13 23:45:35.494383 update_engine[1751]: E20250513 23:45:35.494369 1751 omaha_request_action.cc:636] Omaha request network transfer failed. May 13 23:45:35.494432 update_engine[1751]: I20250513 23:45:35.494385 1751 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. May 13 23:45:35.494432 update_engine[1751]: I20250513 23:45:35.494424 1751 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 13 23:45:35.494481 update_engine[1751]: I20250513 23:45:35.494431 1751 update_attempter.cc:306] Processing Done. May 13 23:45:35.494481 update_engine[1751]: E20250513 23:45:35.494446 1751 update_attempter.cc:619] Update failed. May 13 23:45:35.494481 update_engine[1751]: I20250513 23:45:35.494451 1751 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse May 13 23:45:35.494481 update_engine[1751]: I20250513 23:45:35.494455 1751 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) May 13 23:45:35.494481 update_engine[1751]: I20250513 23:45:35.494462 1751 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. May 13 23:45:35.494589 update_engine[1751]: I20250513 23:45:35.494519 1751 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction May 13 23:45:35.494589 update_engine[1751]: I20250513 23:45:35.494542 1751 omaha_request_action.cc:271] Posting an Omaha request to disabled May 13 23:45:35.494589 update_engine[1751]: I20250513 23:45:35.494547 1751 omaha_request_action.cc:272] Request: May 13 23:45:35.494589 update_engine[1751]: May 13 23:45:35.494589 update_engine[1751]: May 13 23:45:35.494589 update_engine[1751]: May 13 23:45:35.494589 update_engine[1751]: May 13 23:45:35.494589 update_engine[1751]: May 13 23:45:35.494589 update_engine[1751]: May 13 23:45:35.494589 update_engine[1751]: I20250513 23:45:35.494553 1751 libcurl_http_fetcher.cc:47] Starting/Resuming transfer May 13 23:45:35.494766 update_engine[1751]: I20250513 23:45:35.494678 1751 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP May 13 23:45:35.495045 update_engine[1751]: I20250513 23:45:35.494889 1751 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. May 13 23:45:35.495108 locksmithd[1849]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 May 13 23:45:35.607249 update_engine[1751]: E20250513 23:45:35.607125 1751 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607203 1751 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607210 1751 omaha_request_action.cc:617] Omaha request response: May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607218 1751 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607223 1751 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607227 1751 update_attempter.cc:306] Processing Done. May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607234 1751 update_attempter.cc:310] Error event sent. May 13 23:45:35.607249 update_engine[1751]: I20250513 23:45:35.607244 1751 update_check_scheduler.cc:74] Next update check in 42m53s May 13 23:45:35.607631 locksmithd[1849]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 May 13 23:45:35.749023 sshd[4806]: Accepted publickey for core from 10.200.16.10 port 53678 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:35.750355 sshd-session[4806]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:35.755010 systemd-logind[1744]: New session 20 of user core. May 13 23:45:35.760963 systemd[1]: Started session-20.scope - Session 20 of User core. May 13 23:45:36.167672 sshd[4808]: Connection closed by 10.200.16.10 port 53678 May 13 23:45:36.168249 sshd-session[4806]: pam_unix(sshd:session): session closed for user core May 13 23:45:36.171683 systemd[1]: sshd@17-10.200.20.30:22-10.200.16.10:53678.service: Deactivated successfully. May 13 23:45:36.173890 systemd[1]: session-20.scope: Deactivated successfully. May 13 23:45:36.174994 systemd-logind[1744]: Session 20 logged out. Waiting for processes to exit. May 13 23:45:36.175839 systemd-logind[1744]: Removed session 20. May 13 23:45:36.265452 systemd[1]: Started sshd@18-10.200.20.30:22-10.200.16.10:53680.service - OpenSSH per-connection server daemon (10.200.16.10:53680). May 13 23:45:36.750005 sshd[4820]: Accepted publickey for core from 10.200.16.10 port 53680 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:36.751297 sshd-session[4820]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:36.756023 systemd-logind[1744]: New session 21 of user core. May 13 23:45:36.761956 systemd[1]: Started session-21.scope - Session 21 of User core. May 13 23:45:37.212847 sshd[4822]: Connection closed by 10.200.16.10 port 53680 May 13 23:45:37.213473 sshd-session[4820]: pam_unix(sshd:session): session closed for user core May 13 23:45:37.216265 systemd-logind[1744]: Session 21 logged out. Waiting for processes to exit. May 13 23:45:37.218116 systemd[1]: sshd@18-10.200.20.30:22-10.200.16.10:53680.service: Deactivated successfully. May 13 23:45:37.220099 systemd[1]: session-21.scope: Deactivated successfully. May 13 23:45:37.220858 systemd-logind[1744]: Removed session 21. May 13 23:45:37.298055 systemd[1]: Started sshd@19-10.200.20.30:22-10.200.16.10:53684.service - OpenSSH per-connection server daemon (10.200.16.10:53684). May 13 23:45:37.763170 sshd[4832]: Accepted publickey for core from 10.200.16.10 port 53684 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:37.764514 sshd-session[4832]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:37.769007 systemd-logind[1744]: New session 22 of user core. May 13 23:45:37.776955 systemd[1]: Started session-22.scope - Session 22 of User core. May 13 23:45:39.621847 sshd[4834]: Connection closed by 10.200.16.10 port 53684 May 13 23:45:39.622332 sshd-session[4832]: pam_unix(sshd:session): session closed for user core May 13 23:45:39.625666 systemd[1]: sshd@19-10.200.20.30:22-10.200.16.10:53684.service: Deactivated successfully. May 13 23:45:39.628351 systemd[1]: session-22.scope: Deactivated successfully. May 13 23:45:39.630207 systemd-logind[1744]: Session 22 logged out. Waiting for processes to exit. May 13 23:45:39.631611 systemd-logind[1744]: Removed session 22. May 13 23:45:39.733384 systemd[1]: Started sshd@20-10.200.20.30:22-10.200.16.10:44890.service - OpenSSH per-connection server daemon (10.200.16.10:44890). May 13 23:45:40.228489 sshd[4851]: Accepted publickey for core from 10.200.16.10 port 44890 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:40.230021 sshd-session[4851]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:40.234320 systemd-logind[1744]: New session 23 of user core. May 13 23:45:40.239962 systemd[1]: Started session-23.scope - Session 23 of User core. May 13 23:45:40.759682 sshd[4853]: Connection closed by 10.200.16.10 port 44890 May 13 23:45:40.760066 sshd-session[4851]: pam_unix(sshd:session): session closed for user core May 13 23:45:40.763868 systemd[1]: sshd@20-10.200.20.30:22-10.200.16.10:44890.service: Deactivated successfully. May 13 23:45:40.765844 systemd[1]: session-23.scope: Deactivated successfully. May 13 23:45:40.766717 systemd-logind[1744]: Session 23 logged out. Waiting for processes to exit. May 13 23:45:40.767665 systemd-logind[1744]: Removed session 23. May 13 23:45:40.847234 systemd[1]: Started sshd@21-10.200.20.30:22-10.200.16.10:44892.service - OpenSSH per-connection server daemon (10.200.16.10:44892). May 13 23:45:41.334984 sshd[4863]: Accepted publickey for core from 10.200.16.10 port 44892 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:41.336317 sshd-session[4863]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:41.342051 systemd-logind[1744]: New session 24 of user core. May 13 23:45:41.344960 systemd[1]: Started session-24.scope - Session 24 of User core. May 13 23:45:41.751906 sshd[4865]: Connection closed by 10.200.16.10 port 44892 May 13 23:45:41.752471 sshd-session[4863]: pam_unix(sshd:session): session closed for user core May 13 23:45:41.755629 systemd[1]: sshd@21-10.200.20.30:22-10.200.16.10:44892.service: Deactivated successfully. May 13 23:45:41.757492 systemd[1]: session-24.scope: Deactivated successfully. May 13 23:45:41.758297 systemd-logind[1744]: Session 24 logged out. Waiting for processes to exit. May 13 23:45:41.759189 systemd-logind[1744]: Removed session 24. May 13 23:45:46.841017 systemd[1]: Started sshd@22-10.200.20.30:22-10.200.16.10:44896.service - OpenSSH per-connection server daemon (10.200.16.10:44896). May 13 23:45:47.335197 sshd[4876]: Accepted publickey for core from 10.200.16.10 port 44896 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:47.336530 sshd-session[4876]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:47.340750 systemd-logind[1744]: New session 25 of user core. May 13 23:45:47.345948 systemd[1]: Started session-25.scope - Session 25 of User core. May 13 23:45:47.757790 sshd[4878]: Connection closed by 10.200.16.10 port 44896 May 13 23:45:47.758030 sshd-session[4876]: pam_unix(sshd:session): session closed for user core May 13 23:45:47.761165 systemd[1]: sshd@22-10.200.20.30:22-10.200.16.10:44896.service: Deactivated successfully. May 13 23:45:47.762695 systemd[1]: session-25.scope: Deactivated successfully. May 13 23:45:47.763792 systemd-logind[1744]: Session 25 logged out. Waiting for processes to exit. May 13 23:45:47.764801 systemd-logind[1744]: Removed session 25. May 13 23:45:52.841472 systemd[1]: Started sshd@23-10.200.20.30:22-10.200.16.10:52834.service - OpenSSH per-connection server daemon (10.200.16.10:52834). May 13 23:45:53.308178 sshd[4894]: Accepted publickey for core from 10.200.16.10 port 52834 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:53.309422 sshd-session[4894]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:53.313610 systemd-logind[1744]: New session 26 of user core. May 13 23:45:53.323951 systemd[1]: Started session-26.scope - Session 26 of User core. May 13 23:45:53.707240 sshd[4896]: Connection closed by 10.200.16.10 port 52834 May 13 23:45:53.707859 sshd-session[4894]: pam_unix(sshd:session): session closed for user core May 13 23:45:53.711263 systemd[1]: sshd@23-10.200.20.30:22-10.200.16.10:52834.service: Deactivated successfully. May 13 23:45:53.713100 systemd[1]: session-26.scope: Deactivated successfully. May 13 23:45:53.713991 systemd-logind[1744]: Session 26 logged out. Waiting for processes to exit. May 13 23:45:53.715060 systemd-logind[1744]: Removed session 26. May 13 23:45:58.796763 systemd[1]: Started sshd@24-10.200.20.30:22-10.200.16.10:41248.service - OpenSSH per-connection server daemon (10.200.16.10:41248). May 13 23:45:59.259370 sshd[4908]: Accepted publickey for core from 10.200.16.10 port 41248 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:45:59.260661 sshd-session[4908]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:45:59.265240 systemd-logind[1744]: New session 27 of user core. May 13 23:45:59.272969 systemd[1]: Started session-27.scope - Session 27 of User core. May 13 23:45:59.660982 sshd[4911]: Connection closed by 10.200.16.10 port 41248 May 13 23:45:59.661646 sshd-session[4908]: pam_unix(sshd:session): session closed for user core May 13 23:45:59.664866 systemd-logind[1744]: Session 27 logged out. Waiting for processes to exit. May 13 23:45:59.665525 systemd[1]: sshd@24-10.200.20.30:22-10.200.16.10:41248.service: Deactivated successfully. May 13 23:45:59.667743 systemd[1]: session-27.scope: Deactivated successfully. May 13 23:45:59.669006 systemd-logind[1744]: Removed session 27. May 13 23:46:04.749926 systemd[1]: Started sshd@25-10.200.20.30:22-10.200.16.10:41254.service - OpenSSH per-connection server daemon (10.200.16.10:41254). May 13 23:46:05.246882 sshd[4924]: Accepted publickey for core from 10.200.16.10 port 41254 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:46:05.248197 sshd-session[4924]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:46:05.252271 systemd-logind[1744]: New session 28 of user core. May 13 23:46:05.258943 systemd[1]: Started session-28.scope - Session 28 of User core. May 13 23:46:05.667988 sshd[4927]: Connection closed by 10.200.16.10 port 41254 May 13 23:46:05.668679 sshd-session[4924]: pam_unix(sshd:session): session closed for user core May 13 23:46:05.671985 systemd-logind[1744]: Session 28 logged out. Waiting for processes to exit. May 13 23:46:05.672566 systemd[1]: sshd@25-10.200.20.30:22-10.200.16.10:41254.service: Deactivated successfully. May 13 23:46:05.674798 systemd[1]: session-28.scope: Deactivated successfully. May 13 23:46:05.676523 systemd-logind[1744]: Removed session 28. May 13 23:46:05.754277 systemd[1]: Started sshd@26-10.200.20.30:22-10.200.16.10:41256.service - OpenSSH per-connection server daemon (10.200.16.10:41256). May 13 23:46:06.219408 sshd[4939]: Accepted publickey for core from 10.200.16.10 port 41256 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:46:06.220680 sshd-session[4939]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:46:06.225875 systemd-logind[1744]: New session 29 of user core. May 13 23:46:06.231984 systemd[1]: Started session-29.scope - Session 29 of User core. May 13 23:46:09.085700 containerd[1799]: time="2025-05-13T23:46:09.085636074Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 13 23:46:09.089394 containerd[1799]: time="2025-05-13T23:46:09.089293881Z" level=info msg="TaskExit event in podsandbox handler container_id:\"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" id:\"e62b489cdcb52d0483f9eebdad787f56ece245f23e9c0bbbd02895b3ba8a3857\" pid:4959 exited_at:{seconds:1747179969 nanos:88980920}" May 13 23:46:09.092299 containerd[1799]: time="2025-05-13T23:46:09.092272406Z" level=info msg="StopContainer for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" with timeout 2 (s)" May 13 23:46:09.092591 containerd[1799]: time="2025-05-13T23:46:09.092558727Z" level=info msg="Stop container \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" with signal terminated" May 13 23:46:09.099598 systemd-networkd[1353]: lxc_health: Link DOWN May 13 23:46:09.100476 systemd-networkd[1353]: lxc_health: Lost carrier May 13 23:46:09.110591 systemd[1]: cri-containerd-187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3.scope: Deactivated successfully. May 13 23:46:09.111184 systemd[1]: cri-containerd-187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3.scope: Consumed 6.163s CPU time, 123.9M memory peak, 136K read from disk, 12.9M written to disk. May 13 23:46:09.111877 containerd[1799]: time="2025-05-13T23:46:09.111205523Z" level=info msg="received exit event container_id:\"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" id:\"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" pid:3967 exited_at:{seconds:1747179969 nanos:110606321}" May 13 23:46:09.111877 containerd[1799]: time="2025-05-13T23:46:09.111748804Z" level=info msg="TaskExit event in podsandbox handler container_id:\"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" id:\"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" pid:3967 exited_at:{seconds:1747179969 nanos:110606321}" May 13 23:46:09.130704 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3-rootfs.mount: Deactivated successfully. May 13 23:46:09.150846 containerd[1799]: time="2025-05-13T23:46:09.150787318Z" level=info msg="StopContainer for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" with timeout 30 (s)" May 13 23:46:09.151239 containerd[1799]: time="2025-05-13T23:46:09.151218239Z" level=info msg="Stop container \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" with signal terminated" May 13 23:46:09.161493 systemd[1]: cri-containerd-2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f.scope: Deactivated successfully. May 13 23:46:09.164596 containerd[1799]: time="2025-05-13T23:46:09.164368224Z" level=info msg="received exit event container_id:\"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" id:\"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" pid:3737 exited_at:{seconds:1747179969 nanos:163793023}" May 13 23:46:09.164909 containerd[1799]: time="2025-05-13T23:46:09.164741905Z" level=info msg="TaskExit event in podsandbox handler container_id:\"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" id:\"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" pid:3737 exited_at:{seconds:1747179969 nanos:163793023}" May 13 23:46:09.182492 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f-rootfs.mount: Deactivated successfully. May 13 23:46:09.255350 containerd[1799]: time="2025-05-13T23:46:09.255311239Z" level=info msg="StopContainer for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" returns successfully" May 13 23:46:09.257285 containerd[1799]: time="2025-05-13T23:46:09.256853122Z" level=info msg="StopPodSandbox for \"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\"" May 13 23:46:09.257285 containerd[1799]: time="2025-05-13T23:46:09.256924682Z" level=info msg="Container to stop \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 13 23:46:09.257861 containerd[1799]: time="2025-05-13T23:46:09.257830363Z" level=info msg="StopContainer for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" returns successfully" May 13 23:46:09.258273 containerd[1799]: time="2025-05-13T23:46:09.258250284Z" level=info msg="StopPodSandbox for \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\"" May 13 23:46:09.258420 containerd[1799]: time="2025-05-13T23:46:09.258401805Z" level=info msg="Container to stop \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 13 23:46:09.258555 containerd[1799]: time="2025-05-13T23:46:09.258483165Z" level=info msg="Container to stop \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 13 23:46:09.258555 containerd[1799]: time="2025-05-13T23:46:09.258498845Z" level=info msg="Container to stop \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 13 23:46:09.258555 containerd[1799]: time="2025-05-13T23:46:09.258508445Z" level=info msg="Container to stop \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 13 23:46:09.258555 containerd[1799]: time="2025-05-13T23:46:09.258517485Z" level=info msg="Container to stop \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 13 23:46:09.263546 systemd[1]: cri-containerd-f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e.scope: Deactivated successfully. May 13 23:46:09.265180 systemd[1]: cri-containerd-75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb.scope: Deactivated successfully. May 13 23:46:09.273336 containerd[1799]: time="2025-05-13T23:46:09.273298633Z" level=info msg="TaskExit event in podsandbox handler container_id:\"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" id:\"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" pid:3450 exit_status:137 exited_at:{seconds:1747179969 nanos:273059073}" May 13 23:46:09.293924 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb-rootfs.mount: Deactivated successfully. May 13 23:46:09.305629 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e-rootfs.mount: Deactivated successfully. May 13 23:46:09.326417 containerd[1799]: time="2025-05-13T23:46:09.326267975Z" level=info msg="shim disconnected" id=75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb namespace=k8s.io May 13 23:46:09.326557 containerd[1799]: time="2025-05-13T23:46:09.326408335Z" level=warning msg="cleaning up after shim disconnected" id=75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb namespace=k8s.io May 13 23:46:09.326557 containerd[1799]: time="2025-05-13T23:46:09.326440855Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 13 23:46:09.327258 containerd[1799]: time="2025-05-13T23:46:09.327231376Z" level=info msg="TaskExit event in podsandbox handler container_id:\"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" id:\"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" pid:3537 exit_status:137 exited_at:{seconds:1747179969 nanos:273701274}" May 13 23:46:09.327848 containerd[1799]: time="2025-05-13T23:46:09.327722977Z" level=info msg="received exit event sandbox_id:\"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" exit_status:137 exited_at:{seconds:1747179969 nanos:273701274}" May 13 23:46:09.328481 containerd[1799]: time="2025-05-13T23:46:09.328453859Z" level=info msg="shim disconnected" id=f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e namespace=k8s.io May 13 23:46:09.328630 containerd[1799]: time="2025-05-13T23:46:09.328558379Z" level=warning msg="cleaning up after shim disconnected" id=f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e namespace=k8s.io May 13 23:46:09.328630 containerd[1799]: time="2025-05-13T23:46:09.328586659Z" level=info msg="cleaning up dead shim" namespace=k8s.io May 13 23:46:09.329376 containerd[1799]: time="2025-05-13T23:46:09.329306220Z" level=info msg="received exit event sandbox_id:\"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" exit_status:137 exited_at:{seconds:1747179969 nanos:273059073}" May 13 23:46:09.331784 containerd[1799]: time="2025-05-13T23:46:09.331425144Z" level=info msg="TearDown network for sandbox \"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" successfully" May 13 23:46:09.331784 containerd[1799]: time="2025-05-13T23:46:09.331745265Z" level=info msg="StopPodSandbox for \"f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e\" returns successfully" May 13 23:46:09.332866 containerd[1799]: time="2025-05-13T23:46:09.332779787Z" level=info msg="TearDown network for sandbox \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" successfully" May 13 23:46:09.333024 containerd[1799]: time="2025-05-13T23:46:09.332937987Z" level=info msg="StopPodSandbox for \"75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb\" returns successfully" May 13 23:46:09.334401 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-75d3b55e0f00cc1bb413415c8ee250f5b40a6b3e9168cf0ca4256abc27da5ceb-shm.mount: Deactivated successfully. May 13 23:46:09.334502 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f2b4595876d1d5d259eedd662e518f0f140fd03309a1b01e6e73f131219de07e-shm.mount: Deactivated successfully. May 13 23:46:09.410728 kubelet[3344]: I0513 23:46:09.410158 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkjs2\" (UniqueName: \"kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-kube-api-access-bkjs2\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.410728 kubelet[3344]: I0513 23:46:09.410195 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-etc-cni-netd\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.410728 kubelet[3344]: I0513 23:46:09.410214 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-bpf-maps\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.410728 kubelet[3344]: I0513 23:46:09.410228 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-kernel\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.410728 kubelet[3344]: I0513 23:46:09.410246 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2f5nn\" (UniqueName: \"kubernetes.io/projected/815743fd-a18b-4aaf-993d-3b7577b9dd4f-kube-api-access-2f5nn\") pod \"815743fd-a18b-4aaf-993d-3b7577b9dd4f\" (UID: \"815743fd-a18b-4aaf-993d-3b7577b9dd4f\") " May 13 23:46:09.410728 kubelet[3344]: I0513 23:46:09.410262 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-xtables-lock\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411183 kubelet[3344]: I0513 23:46:09.410276 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-lib-modules\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411183 kubelet[3344]: I0513 23:46:09.410291 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-hubble-tls\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411183 kubelet[3344]: I0513 23:46:09.410305 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-net\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411183 kubelet[3344]: I0513 23:46:09.410319 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cni-path\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411183 kubelet[3344]: I0513 23:46:09.410332 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-hostproc\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411183 kubelet[3344]: I0513 23:46:09.410348 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-run\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411318 kubelet[3344]: I0513 23:46:09.410365 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/815743fd-a18b-4aaf-993d-3b7577b9dd4f-cilium-config-path\") pod \"815743fd-a18b-4aaf-993d-3b7577b9dd4f\" (UID: \"815743fd-a18b-4aaf-993d-3b7577b9dd4f\") " May 13 23:46:09.411318 kubelet[3344]: I0513 23:46:09.410381 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-config-path\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411318 kubelet[3344]: I0513 23:46:09.410397 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-cgroup\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.411318 kubelet[3344]: I0513 23:46:09.410416 3344 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ca8a8273-b2db-4a64-922e-a6200979faee-clustermesh-secrets\") pod \"ca8a8273-b2db-4a64-922e-a6200979faee\" (UID: \"ca8a8273-b2db-4a64-922e-a6200979faee\") " May 13 23:46:09.415607 kubelet[3344]: I0513 23:46:09.414799 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.415607 kubelet[3344]: I0513 23:46:09.414867 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cni-path" (OuterVolumeSpecName: "cni-path") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.415607 kubelet[3344]: I0513 23:46:09.414884 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-hostproc" (OuterVolumeSpecName: "hostproc") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.415607 kubelet[3344]: I0513 23:46:09.414899 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.415607 kubelet[3344]: I0513 23:46:09.415279 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.416617 kubelet[3344]: I0513 23:46:09.415310 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.416617 kubelet[3344]: I0513 23:46:09.415342 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.416914 kubelet[3344]: I0513 23:46:09.416847 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca8a8273-b2db-4a64-922e-a6200979faee-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 13 23:46:09.418619 kubelet[3344]: I0513 23:46:09.418492 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.418619 kubelet[3344]: I0513 23:46:09.418534 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.418749 kubelet[3344]: I0513 23:46:09.418727 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/815743fd-a18b-4aaf-993d-3b7577b9dd4f-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "815743fd-a18b-4aaf-993d-3b7577b9dd4f" (UID: "815743fd-a18b-4aaf-993d-3b7577b9dd4f"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 13 23:46:09.418838 kubelet[3344]: I0513 23:46:09.418800 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 13 23:46:09.419832 kubelet[3344]: I0513 23:46:09.419766 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-kube-api-access-bkjs2" (OuterVolumeSpecName: "kube-api-access-bkjs2") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "kube-api-access-bkjs2". PluginName "kubernetes.io/projected", VolumeGidValue "" May 13 23:46:09.420144 kubelet[3344]: I0513 23:46:09.420115 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/815743fd-a18b-4aaf-993d-3b7577b9dd4f-kube-api-access-2f5nn" (OuterVolumeSpecName: "kube-api-access-2f5nn") pod "815743fd-a18b-4aaf-993d-3b7577b9dd4f" (UID: "815743fd-a18b-4aaf-993d-3b7577b9dd4f"). InnerVolumeSpecName "kube-api-access-2f5nn". PluginName "kubernetes.io/projected", VolumeGidValue "" May 13 23:46:09.420189 kubelet[3344]: I0513 23:46:09.420159 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 13 23:46:09.420942 kubelet[3344]: I0513 23:46:09.420910 3344 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "ca8a8273-b2db-4a64-922e-a6200979faee" (UID: "ca8a8273-b2db-4a64-922e-a6200979faee"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511442 3344 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-hubble-tls\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511474 3344 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-net\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511484 3344 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cni-path\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511493 3344 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-hostproc\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511502 3344 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-run\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511511 3344 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/815743fd-a18b-4aaf-993d-3b7577b9dd4f-cilium-config-path\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511518 3344 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-config-path\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.511632 kubelet[3344]: I0513 23:46:09.511526 3344 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-cilium-cgroup\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511542 3344 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ca8a8273-b2db-4a64-922e-a6200979faee-clustermesh-secrets\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511551 3344 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-bkjs2\" (UniqueName: \"kubernetes.io/projected/ca8a8273-b2db-4a64-922e-a6200979faee-kube-api-access-bkjs2\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511560 3344 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-etc-cni-netd\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511568 3344 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-bpf-maps\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511575 3344 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-host-proc-sys-kernel\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511587 3344 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-2f5nn\" (UniqueName: \"kubernetes.io/projected/815743fd-a18b-4aaf-993d-3b7577b9dd4f-kube-api-access-2f5nn\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511595 3344 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-xtables-lock\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.512016 kubelet[3344]: I0513 23:46:09.511603 3344 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ca8a8273-b2db-4a64-922e-a6200979faee-lib-modules\") on node \"ci-4284.0.0-n-f756a0c81e\" DevicePath \"\"" May 13 23:46:09.642972 kubelet[3344]: E0513 23:46:09.642930 3344 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 13 23:46:09.845606 kubelet[3344]: I0513 23:46:09.845236 3344 scope.go:117] "RemoveContainer" containerID="187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3" May 13 23:46:09.848757 containerd[1799]: time="2025-05-13T23:46:09.848714216Z" level=info msg="RemoveContainer for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\"" May 13 23:46:09.854202 systemd[1]: Removed slice kubepods-burstable-podca8a8273_b2db_4a64_922e_a6200979faee.slice - libcontainer container kubepods-burstable-podca8a8273_b2db_4a64_922e_a6200979faee.slice. May 13 23:46:09.854390 systemd[1]: kubepods-burstable-podca8a8273_b2db_4a64_922e_a6200979faee.slice: Consumed 6.238s CPU time, 124.3M memory peak, 136K read from disk, 12.9M written to disk. May 13 23:46:09.857726 systemd[1]: Removed slice kubepods-besteffort-pod815743fd_a18b_4aaf_993d_3b7577b9dd4f.slice - libcontainer container kubepods-besteffort-pod815743fd_a18b_4aaf_993d_3b7577b9dd4f.slice. May 13 23:46:09.862614 containerd[1799]: time="2025-05-13T23:46:09.862470322Z" level=info msg="RemoveContainer for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" returns successfully" May 13 23:46:09.862871 kubelet[3344]: I0513 23:46:09.862850 3344 scope.go:117] "RemoveContainer" containerID="ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9" May 13 23:46:09.865580 containerd[1799]: time="2025-05-13T23:46:09.865529688Z" level=info msg="RemoveContainer for \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\"" May 13 23:46:09.886677 containerd[1799]: time="2025-05-13T23:46:09.886617488Z" level=info msg="RemoveContainer for \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" returns successfully" May 13 23:46:09.886996 kubelet[3344]: I0513 23:46:09.886965 3344 scope.go:117] "RemoveContainer" containerID="a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49" May 13 23:46:09.889495 containerd[1799]: time="2025-05-13T23:46:09.889421734Z" level=info msg="RemoveContainer for \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\"" May 13 23:46:09.920861 containerd[1799]: time="2025-05-13T23:46:09.920826634Z" level=info msg="RemoveContainer for \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" returns successfully" May 13 23:46:09.921066 kubelet[3344]: I0513 23:46:09.921037 3344 scope.go:117] "RemoveContainer" containerID="0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f" May 13 23:46:09.922938 containerd[1799]: time="2025-05-13T23:46:09.922907638Z" level=info msg="RemoveContainer for \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\"" May 13 23:46:09.937398 containerd[1799]: time="2025-05-13T23:46:09.937344585Z" level=info msg="RemoveContainer for \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" returns successfully" May 13 23:46:09.937702 kubelet[3344]: I0513 23:46:09.937618 3344 scope.go:117] "RemoveContainer" containerID="0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304" May 13 23:46:09.939726 containerd[1799]: time="2025-05-13T23:46:09.939233669Z" level=info msg="RemoveContainer for \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\"" May 13 23:46:09.950375 containerd[1799]: time="2025-05-13T23:46:09.950273770Z" level=info msg="RemoveContainer for \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" returns successfully" May 13 23:46:09.950617 kubelet[3344]: I0513 23:46:09.950475 3344 scope.go:117] "RemoveContainer" containerID="187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3" May 13 23:46:09.950694 containerd[1799]: time="2025-05-13T23:46:09.950654211Z" level=error msg="ContainerStatus for \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\": not found" May 13 23:46:09.951266 kubelet[3344]: E0513 23:46:09.951242 3344 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\": not found" containerID="187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3" May 13 23:46:09.951427 kubelet[3344]: I0513 23:46:09.951274 3344 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3"} err="failed to get container status \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\": rpc error: code = NotFound desc = an error occurred when try to find container \"187af95222616429c69fb91a8fb30b4b40818809ba3833e787e9afa08db81cc3\": not found" May 13 23:46:09.951427 kubelet[3344]: I0513 23:46:09.951346 3344 scope.go:117] "RemoveContainer" containerID="ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9" May 13 23:46:09.951593 containerd[1799]: time="2025-05-13T23:46:09.951544413Z" level=error msg="ContainerStatus for \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\": not found" May 13 23:46:09.952139 kubelet[3344]: E0513 23:46:09.951708 3344 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\": not found" containerID="ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9" May 13 23:46:09.952139 kubelet[3344]: I0513 23:46:09.951734 3344 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9"} err="failed to get container status \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\": rpc error: code = NotFound desc = an error occurred when try to find container \"ccea6bcf15cb9902230140f9ef9f3f5b993148722e89f65a9617e5b46cfd8fe9\": not found" May 13 23:46:09.952139 kubelet[3344]: I0513 23:46:09.951750 3344 scope.go:117] "RemoveContainer" containerID="a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49" May 13 23:46:09.952314 containerd[1799]: time="2025-05-13T23:46:09.952077254Z" level=error msg="ContainerStatus for \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\": not found" May 13 23:46:09.952554 kubelet[3344]: E0513 23:46:09.952437 3344 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\": not found" containerID="a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49" May 13 23:46:09.952554 kubelet[3344]: I0513 23:46:09.952464 3344 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49"} err="failed to get container status \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\": rpc error: code = NotFound desc = an error occurred when try to find container \"a69dfd02eb419a84cbb3df6209dc080e4d3395067c3594ba211a9e7b05c23f49\": not found" May 13 23:46:09.952554 kubelet[3344]: I0513 23:46:09.952484 3344 scope.go:117] "RemoveContainer" containerID="0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f" May 13 23:46:09.952809 containerd[1799]: time="2025-05-13T23:46:09.952773775Z" level=error msg="ContainerStatus for \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\": not found" May 13 23:46:09.952953 kubelet[3344]: E0513 23:46:09.952928 3344 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\": not found" containerID="0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f" May 13 23:46:09.953000 kubelet[3344]: I0513 23:46:09.952960 3344 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f"} err="failed to get container status \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\": rpc error: code = NotFound desc = an error occurred when try to find container \"0c9674019144d95031b6463b5fccbc458e7239b309dd6140ebfa1313f9f1597f\": not found" May 13 23:46:09.953000 kubelet[3344]: I0513 23:46:09.952982 3344 scope.go:117] "RemoveContainer" containerID="0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304" May 13 23:46:09.953167 containerd[1799]: time="2025-05-13T23:46:09.953136536Z" level=error msg="ContainerStatus for \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\": not found" May 13 23:46:09.953413 kubelet[3344]: E0513 23:46:09.953307 3344 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\": not found" containerID="0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304" May 13 23:46:09.953413 kubelet[3344]: I0513 23:46:09.953333 3344 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304"} err="failed to get container status \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\": rpc error: code = NotFound desc = an error occurred when try to find container \"0b913e924df9dd4f99f46888a1204274afc6acd103c1f46f11a6f6742a2e4304\": not found" May 13 23:46:09.953413 kubelet[3344]: I0513 23:46:09.953351 3344 scope.go:117] "RemoveContainer" containerID="2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f" May 13 23:46:09.955074 containerd[1799]: time="2025-05-13T23:46:09.955042019Z" level=info msg="RemoveContainer for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\"" May 13 23:46:09.969163 containerd[1799]: time="2025-05-13T23:46:09.969130726Z" level=info msg="RemoveContainer for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" returns successfully" May 13 23:46:09.969392 kubelet[3344]: I0513 23:46:09.969369 3344 scope.go:117] "RemoveContainer" containerID="2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f" May 13 23:46:09.969658 containerd[1799]: time="2025-05-13T23:46:09.969629487Z" level=error msg="ContainerStatus for \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\": not found" May 13 23:46:09.970001 kubelet[3344]: E0513 23:46:09.969935 3344 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\": not found" containerID="2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f" May 13 23:46:09.970001 kubelet[3344]: I0513 23:46:09.969961 3344 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f"} err="failed to get container status \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\": rpc error: code = NotFound desc = an error occurred when try to find container \"2ab6f8bcb2575487818f76e0a551e0490a585c2ca4f64e58599435add730e63f\": not found" May 13 23:46:10.131398 systemd[1]: var-lib-kubelet-pods-ca8a8273\x2db2db\x2d4a64\x2d922e\x2da6200979faee-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbkjs2.mount: Deactivated successfully. May 13 23:46:10.131852 systemd[1]: var-lib-kubelet-pods-ca8a8273\x2db2db\x2d4a64\x2d922e\x2da6200979faee-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 13 23:46:10.131911 systemd[1]: var-lib-kubelet-pods-ca8a8273\x2db2db\x2d4a64\x2d922e\x2da6200979faee-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 13 23:46:10.131970 systemd[1]: var-lib-kubelet-pods-815743fd\x2da18b\x2d4aaf\x2d993d\x2d3b7577b9dd4f-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d2f5nn.mount: Deactivated successfully. May 13 23:46:10.468477 kubelet[3344]: I0513 23:46:10.468441 3344 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="815743fd-a18b-4aaf-993d-3b7577b9dd4f" path="/var/lib/kubelet/pods/815743fd-a18b-4aaf-993d-3b7577b9dd4f/volumes" May 13 23:46:10.468956 kubelet[3344]: I0513 23:46:10.468854 3344 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" path="/var/lib/kubelet/pods/ca8a8273-b2db-4a64-922e-a6200979faee/volumes" May 13 23:46:11.042351 sshd[4941]: Connection closed by 10.200.16.10 port 41256 May 13 23:46:11.042965 sshd-session[4939]: pam_unix(sshd:session): session closed for user core May 13 23:46:11.045862 systemd-logind[1744]: Session 29 logged out. Waiting for processes to exit. May 13 23:46:11.046788 systemd[1]: sshd@26-10.200.20.30:22-10.200.16.10:41256.service: Deactivated successfully. May 13 23:46:11.049103 systemd[1]: session-29.scope: Deactivated successfully. May 13 23:46:11.049283 systemd[1]: session-29.scope: Consumed 1.899s CPU time, 25.6M memory peak. May 13 23:46:11.050789 systemd-logind[1744]: Removed session 29. May 13 23:46:11.123479 systemd[1]: Started sshd@27-10.200.20.30:22-10.200.16.10:51118.service - OpenSSH per-connection server daemon (10.200.16.10:51118). May 13 23:46:11.465925 kubelet[3344]: E0513 23:46:11.465854 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-mzlsw" podUID="fe34d60e-3093-44e8-ad4c-bc48701a180d" May 13 23:46:11.578062 sshd[5088]: Accepted publickey for core from 10.200.16.10 port 51118 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:46:11.579333 sshd-session[5088]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:46:11.583996 systemd-logind[1744]: New session 30 of user core. May 13 23:46:11.592955 systemd[1]: Started session-30.scope - Session 30 of User core. May 13 23:46:13.465982 kubelet[3344]: E0513 23:46:13.465809 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-mzlsw" podUID="fe34d60e-3093-44e8-ad4c-bc48701a180d" May 13 23:46:13.465982 kubelet[3344]: E0513 23:46:13.465938 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-2zldt" podUID="f1495c6c-2079-4ea1-bf3a-a24048d8c525" May 13 23:46:13.750750 sshd[5091]: Connection closed by 10.200.16.10 port 51118 May 13 23:46:13.750246 sshd-session[5088]: pam_unix(sshd:session): session closed for user core May 13 23:46:13.753356 systemd-logind[1744]: Session 30 logged out. Waiting for processes to exit. May 13 23:46:13.753605 systemd[1]: sshd@27-10.200.20.30:22-10.200.16.10:51118.service: Deactivated successfully. May 13 23:46:13.755762 systemd[1]: session-30.scope: Deactivated successfully. May 13 23:46:13.755988 systemd[1]: session-30.scope: Consumed 1.449s CPU time, 24M memory peak. May 13 23:46:13.757559 systemd-logind[1744]: Removed session 30. May 13 23:46:13.792625 kubelet[3344]: I0513 23:46:13.790262 3344 topology_manager.go:215] "Topology Admit Handler" podUID="cd4a439a-d22f-41d2-8b6b-74f3b4ea2004" podNamespace="kube-system" podName="cilium-kvcmc" May 13 23:46:13.792625 kubelet[3344]: E0513 23:46:13.790320 3344 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" containerName="apply-sysctl-overwrites" May 13 23:46:13.792625 kubelet[3344]: E0513 23:46:13.790329 3344 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" containerName="cilium-agent" May 13 23:46:13.792625 kubelet[3344]: E0513 23:46:13.790334 3344 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" containerName="mount-cgroup" May 13 23:46:13.792625 kubelet[3344]: E0513 23:46:13.790341 3344 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="815743fd-a18b-4aaf-993d-3b7577b9dd4f" containerName="cilium-operator" May 13 23:46:13.792625 kubelet[3344]: E0513 23:46:13.790346 3344 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" containerName="mount-bpf-fs" May 13 23:46:13.792625 kubelet[3344]: E0513 23:46:13.790354 3344 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" containerName="clean-cilium-state" May 13 23:46:13.792625 kubelet[3344]: I0513 23:46:13.790374 3344 memory_manager.go:354] "RemoveStaleState removing state" podUID="815743fd-a18b-4aaf-993d-3b7577b9dd4f" containerName="cilium-operator" May 13 23:46:13.792625 kubelet[3344]: I0513 23:46:13.790380 3344 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8a8273-b2db-4a64-922e-a6200979faee" containerName="cilium-agent" May 13 23:46:13.802482 systemd[1]: Created slice kubepods-burstable-podcd4a439a_d22f_41d2_8b6b_74f3b4ea2004.slice - libcontainer container kubepods-burstable-podcd4a439a_d22f_41d2_8b6b_74f3b4ea2004.slice. May 13 23:46:13.835039 kubelet[3344]: I0513 23:46:13.834979 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-cilium-config-path\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835162 kubelet[3344]: I0513 23:46:13.835052 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-clustermesh-secrets\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835162 kubelet[3344]: I0513 23:46:13.835087 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-cilium-ipsec-secrets\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835162 kubelet[3344]: I0513 23:46:13.835106 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-etc-cni-netd\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835162 kubelet[3344]: I0513 23:46:13.835124 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-xtables-lock\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835162 kubelet[3344]: I0513 23:46:13.835140 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-cni-path\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835162 kubelet[3344]: I0513 23:46:13.835155 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-host-proc-sys-net\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835308 kubelet[3344]: I0513 23:46:13.835169 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg9h4\" (UniqueName: \"kubernetes.io/projected/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-kube-api-access-xg9h4\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835308 kubelet[3344]: I0513 23:46:13.835186 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-cilium-cgroup\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835308 kubelet[3344]: I0513 23:46:13.835200 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-cilium-run\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835308 kubelet[3344]: I0513 23:46:13.835214 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-lib-modules\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835308 kubelet[3344]: I0513 23:46:13.835229 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-bpf-maps\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835308 kubelet[3344]: I0513 23:46:13.835243 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-hostproc\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835435 kubelet[3344]: I0513 23:46:13.835256 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-host-proc-sys-kernel\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.835435 kubelet[3344]: I0513 23:46:13.835274 3344 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/cd4a439a-d22f-41d2-8b6b-74f3b4ea2004-hubble-tls\") pod \"cilium-kvcmc\" (UID: \"cd4a439a-d22f-41d2-8b6b-74f3b4ea2004\") " pod="kube-system/cilium-kvcmc" May 13 23:46:13.844327 systemd[1]: Started sshd@28-10.200.20.30:22-10.200.16.10:51124.service - OpenSSH per-connection server daemon (10.200.16.10:51124). May 13 23:46:14.109767 containerd[1799]: time="2025-05-13T23:46:14.109627937Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kvcmc,Uid:cd4a439a-d22f-41d2-8b6b-74f3b4ea2004,Namespace:kube-system,Attempt:0,}" May 13 23:46:14.179094 containerd[1799]: time="2025-05-13T23:46:14.178766350Z" level=info msg="connecting to shim f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa" address="unix:///run/containerd/s/e8aa6bef3bd594dac75e8ca21bf8576f58137e78d2f5704f5869df002608cff7" namespace=k8s.io protocol=ttrpc version=3 May 13 23:46:14.202958 systemd[1]: Started cri-containerd-f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa.scope - libcontainer container f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa. May 13 23:46:14.229772 containerd[1799]: time="2025-05-13T23:46:14.229729567Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-kvcmc,Uid:cd4a439a-d22f-41d2-8b6b-74f3b4ea2004,Namespace:kube-system,Attempt:0,} returns sandbox id \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\"" May 13 23:46:14.233247 containerd[1799]: time="2025-05-13T23:46:14.233201294Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 13 23:46:14.261793 containerd[1799]: time="2025-05-13T23:46:14.261757949Z" level=info msg="Container c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf: CDI devices from CRI Config.CDIDevices: []" May 13 23:46:14.291232 containerd[1799]: time="2025-05-13T23:46:14.291192125Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\"" May 13 23:46:14.292049 containerd[1799]: time="2025-05-13T23:46:14.291650926Z" level=info msg="StartContainer for \"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\"" May 13 23:46:14.292685 containerd[1799]: time="2025-05-13T23:46:14.292652168Z" level=info msg="connecting to shim c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf" address="unix:///run/containerd/s/e8aa6bef3bd594dac75e8ca21bf8576f58137e78d2f5704f5869df002608cff7" protocol=ttrpc version=3 May 13 23:46:14.310977 systemd[1]: Started cri-containerd-c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf.scope - libcontainer container c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf. May 13 23:46:14.313615 sshd[5102]: Accepted publickey for core from 10.200.16.10 port 51124 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:46:14.315305 sshd-session[5102]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:46:14.321639 systemd-logind[1744]: New session 31 of user core. May 13 23:46:14.326967 systemd[1]: Started session-31.scope - Session 31 of User core. May 13 23:46:14.355919 systemd[1]: cri-containerd-c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf.scope: Deactivated successfully. May 13 23:46:14.359101 containerd[1799]: time="2025-05-13T23:46:14.358142653Z" level=info msg="StartContainer for \"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\" returns successfully" May 13 23:46:14.359432 containerd[1799]: time="2025-05-13T23:46:14.359363456Z" level=info msg="received exit event container_id:\"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\" id:\"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\" pid:5168 exited_at:{seconds:1747179974 nanos:358952615}" May 13 23:46:14.359481 containerd[1799]: time="2025-05-13T23:46:14.359439816Z" level=info msg="TaskExit event in podsandbox handler container_id:\"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\" id:\"c0d3d04feef741fedcc5f7baa9e8f9c032e281a4893b5967f71c05db477965cf\" pid:5168 exited_at:{seconds:1747179974 nanos:358952615}" May 13 23:46:14.635373 sshd[5175]: Connection closed by 10.200.16.10 port 51124 May 13 23:46:14.636246 sshd-session[5102]: pam_unix(sshd:session): session closed for user core May 13 23:46:14.640751 systemd[1]: sshd@28-10.200.20.30:22-10.200.16.10:51124.service: Deactivated successfully. May 13 23:46:14.643619 systemd[1]: session-31.scope: Deactivated successfully. May 13 23:46:14.644515 kubelet[3344]: E0513 23:46:14.644242 3344 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 13 23:46:14.645272 systemd-logind[1744]: Session 31 logged out. Waiting for processes to exit. May 13 23:46:14.646553 systemd-logind[1744]: Removed session 31. May 13 23:46:14.730439 systemd[1]: Started sshd@29-10.200.20.30:22-10.200.16.10:51132.service - OpenSSH per-connection server daemon (10.200.16.10:51132). May 13 23:46:14.869224 containerd[1799]: time="2025-05-13T23:46:14.868742431Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 13 23:46:14.897650 containerd[1799]: time="2025-05-13T23:46:14.897309486Z" level=info msg="Container 16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb: CDI devices from CRI Config.CDIDevices: []" May 13 23:46:14.925306 containerd[1799]: time="2025-05-13T23:46:14.925190659Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\"" May 13 23:46:14.926300 containerd[1799]: time="2025-05-13T23:46:14.926150101Z" level=info msg="StartContainer for \"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\"" May 13 23:46:14.928132 containerd[1799]: time="2025-05-13T23:46:14.928100105Z" level=info msg="connecting to shim 16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb" address="unix:///run/containerd/s/e8aa6bef3bd594dac75e8ca21bf8576f58137e78d2f5704f5869df002608cff7" protocol=ttrpc version=3 May 13 23:46:14.949972 systemd[1]: Started cri-containerd-16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb.scope - libcontainer container 16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb. May 13 23:46:14.978420 containerd[1799]: time="2025-05-13T23:46:14.978309441Z" level=info msg="StartContainer for \"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\" returns successfully" May 13 23:46:14.983038 systemd[1]: cri-containerd-16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb.scope: Deactivated successfully. May 13 23:46:14.983470 containerd[1799]: time="2025-05-13T23:46:14.983426211Z" level=info msg="received exit event container_id:\"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\" id:\"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\" pid:5221 exited_at:{seconds:1747179974 nanos:983123010}" May 13 23:46:14.984879 containerd[1799]: time="2025-05-13T23:46:14.984724893Z" level=info msg="TaskExit event in podsandbox handler container_id:\"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\" id:\"16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb\" pid:5221 exited_at:{seconds:1747179974 nanos:983123010}" May 13 23:46:15.001981 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-16a57298295a3b032dadb804f6345cf5973fce106284cbd0fa149067c37465fb-rootfs.mount: Deactivated successfully. May 13 23:46:15.221549 sshd[5207]: Accepted publickey for core from 10.200.16.10 port 51132 ssh2: RSA SHA256:vkfaD5ZBcZpTdQVgl7gjxJv9L2x8eoUpkC37aWFhQ2A May 13 23:46:15.222900 sshd-session[5207]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) May 13 23:46:15.227664 systemd-logind[1744]: New session 32 of user core. May 13 23:46:15.232980 systemd[1]: Started session-32.scope - Session 32 of User core. May 13 23:46:15.466102 kubelet[3344]: E0513 23:46:15.465929 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-2zldt" podUID="f1495c6c-2079-4ea1-bf3a-a24048d8c525" May 13 23:46:15.466102 kubelet[3344]: E0513 23:46:15.466055 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-mzlsw" podUID="fe34d60e-3093-44e8-ad4c-bc48701a180d" May 13 23:46:15.874393 containerd[1799]: time="2025-05-13T23:46:15.874255237Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 13 23:46:15.907380 containerd[1799]: time="2025-05-13T23:46:15.906483419Z" level=info msg="Container 7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa: CDI devices from CRI Config.CDIDevices: []" May 13 23:46:15.929039 containerd[1799]: time="2025-05-13T23:46:15.929003982Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\"" May 13 23:46:15.929871 containerd[1799]: time="2025-05-13T23:46:15.929736103Z" level=info msg="StartContainer for \"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\"" May 13 23:46:15.931487 containerd[1799]: time="2025-05-13T23:46:15.931294466Z" level=info msg="connecting to shim 7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa" address="unix:///run/containerd/s/e8aa6bef3bd594dac75e8ca21bf8576f58137e78d2f5704f5869df002608cff7" protocol=ttrpc version=3 May 13 23:46:15.952029 systemd[1]: Started cri-containerd-7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa.scope - libcontainer container 7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa. May 13 23:46:15.982752 systemd[1]: cri-containerd-7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa.scope: Deactivated successfully. May 13 23:46:15.987137 containerd[1799]: time="2025-05-13T23:46:15.987007613Z" level=info msg="received exit event container_id:\"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\" id:\"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\" pid:5271 exited_at:{seconds:1747179975 nanos:985290090}" May 13 23:46:15.987137 containerd[1799]: time="2025-05-13T23:46:15.987129933Z" level=info msg="TaskExit event in podsandbox handler container_id:\"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\" id:\"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\" pid:5271 exited_at:{seconds:1747179975 nanos:985290090}" May 13 23:46:15.996092 containerd[1799]: time="2025-05-13T23:46:15.996055270Z" level=info msg="StartContainer for \"7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa\" returns successfully" May 13 23:46:16.006394 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7dd2c1b41fddab27fded2f77c14c1416b504e9e3263a0296769c14f71c3685fa-rootfs.mount: Deactivated successfully. May 13 23:46:16.881359 containerd[1799]: time="2025-05-13T23:46:16.881304686Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 13 23:46:16.914369 containerd[1799]: time="2025-05-13T23:46:16.913718748Z" level=info msg="Container 3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5: CDI devices from CRI Config.CDIDevices: []" May 13 23:46:16.935029 containerd[1799]: time="2025-05-13T23:46:16.934990909Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\"" May 13 23:46:16.935653 containerd[1799]: time="2025-05-13T23:46:16.935626750Z" level=info msg="StartContainer for \"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\"" May 13 23:46:16.936593 containerd[1799]: time="2025-05-13T23:46:16.936566512Z" level=info msg="connecting to shim 3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5" address="unix:///run/containerd/s/e8aa6bef3bd594dac75e8ca21bf8576f58137e78d2f5704f5869df002608cff7" protocol=ttrpc version=3 May 13 23:46:16.958170 systemd[1]: Started cri-containerd-3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5.scope - libcontainer container 3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5. May 13 23:46:17.034292 systemd[1]: cri-containerd-3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5.scope: Deactivated successfully. May 13 23:46:17.038008 containerd[1799]: time="2025-05-13T23:46:17.036497943Z" level=info msg="TaskExit event in podsandbox handler container_id:\"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\" id:\"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\" pid:5309 exited_at:{seconds:1747179977 nanos:36198863}" May 13 23:46:17.039694 containerd[1799]: time="2025-05-13T23:46:17.039665869Z" level=info msg="received exit event container_id:\"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\" id:\"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\" pid:5309 exited_at:{seconds:1747179977 nanos:36198863}" May 13 23:46:17.045842 containerd[1799]: time="2025-05-13T23:46:17.044130478Z" level=info msg="StartContainer for \"3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5\" returns successfully" May 13 23:46:17.065059 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3f7854dd93f3beb30c3da3212fa5cc857247a91a3ef5d81c12c556b0d0766cb5-rootfs.mount: Deactivated successfully. May 13 23:46:17.466632 kubelet[3344]: E0513 23:46:17.466582 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-mzlsw" podUID="fe34d60e-3093-44e8-ad4c-bc48701a180d" May 13 23:46:17.467130 kubelet[3344]: E0513 23:46:17.466984 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-2zldt" podUID="f1495c6c-2079-4ea1-bf3a-a24048d8c525" May 13 23:46:17.888917 containerd[1799]: time="2025-05-13T23:46:17.888475015Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 13 23:46:17.917854 containerd[1799]: time="2025-05-13T23:46:17.917460430Z" level=info msg="Container bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731: CDI devices from CRI Config.CDIDevices: []" May 13 23:46:17.941837 containerd[1799]: time="2025-05-13T23:46:17.941786717Z" level=info msg="CreateContainer within sandbox \"f1ff9c79bdface06c301ddece319473eb7f140119b7e01de0bb34375fa0e4dfa\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\"" May 13 23:46:17.942614 containerd[1799]: time="2025-05-13T23:46:17.942582278Z" level=info msg="StartContainer for \"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\"" May 13 23:46:17.943551 containerd[1799]: time="2025-05-13T23:46:17.943518760Z" level=info msg="connecting to shim bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731" address="unix:///run/containerd/s/e8aa6bef3bd594dac75e8ca21bf8576f58137e78d2f5704f5869df002608cff7" protocol=ttrpc version=3 May 13 23:46:17.961969 systemd[1]: Started cri-containerd-bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731.scope - libcontainer container bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731. May 13 23:46:18.007570 containerd[1799]: time="2025-05-13T23:46:18.007527843Z" level=info msg="StartContainer for \"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" returns successfully" May 13 23:46:18.056218 containerd[1799]: time="2025-05-13T23:46:18.056099456Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" id:\"b549229c39fa5f4ecd954803ff59c3c35d7eab9ff93caccd9c6d46a6ab11fba1\" pid:5376 exited_at:{seconds:1747179978 nanos:55642175}" May 13 23:46:18.321053 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) May 13 23:46:18.914123 kubelet[3344]: I0513 23:46:18.913759 3344 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-kvcmc" podStartSLOduration=5.913742099 podStartE2EDuration="5.913742099s" podCreationTimestamp="2025-05-13 23:46:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-13 23:46:18.913380018 +0000 UTC m=+194.578246820" watchObservedRunningTime="2025-05-13 23:46:18.913742099 +0000 UTC m=+194.578608901" May 13 23:46:19.294571 kubelet[3344]: I0513 23:46:19.294522 3344 setters.go:580] "Node became not ready" node="ci-4284.0.0-n-f756a0c81e" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-13T23:46:19Z","lastTransitionTime":"2025-05-13T23:46:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 13 23:46:19.466917 kubelet[3344]: E0513 23:46:19.466746 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-mzlsw" podUID="fe34d60e-3093-44e8-ad4c-bc48701a180d" May 13 23:46:19.467205 kubelet[3344]: E0513 23:46:19.467168 3344 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-7db6d8ff4d-2zldt" podUID="f1495c6c-2079-4ea1-bf3a-a24048d8c525" May 13 23:46:19.684042 containerd[1799]: time="2025-05-13T23:46:19.683994814Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" id:\"25fc78aa0de4eabe206673482f170c1a4a455a8fd194160a7ae7eb4127040ffc\" pid:5460 exit_status:1 exited_at:{seconds:1747179979 nanos:683490853}" May 13 23:46:20.928530 systemd-networkd[1353]: lxc_health: Link UP May 13 23:46:20.944453 systemd-networkd[1353]: lxc_health: Gained carrier May 13 23:46:21.813683 containerd[1799]: time="2025-05-13T23:46:21.813474500Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" id:\"d14cf2efdd210dfed83c853e8ac9c6b420e7490ea68c9da5f28ca63483f3eca2\" pid:5898 exited_at:{seconds:1747179981 nanos:812991339}" May 13 23:46:22.328002 systemd-networkd[1353]: lxc_health: Gained IPv6LL May 13 23:46:23.961371 containerd[1799]: time="2025-05-13T23:46:23.961316021Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" id:\"e52542f980bf30e59441c3d7684cb72cd1810f21100223849f5016580c95af75\" pid:5931 exited_at:{seconds:1747179983 nanos:959924339}" May 13 23:46:26.090219 containerd[1799]: time="2025-05-13T23:46:26.090176026Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" id:\"54f3cd82fbb86ad196667284d0ca0eff1175687506df4071677f8fb5cbda7c40\" pid:5971 exited_at:{seconds:1747179986 nanos:89798305}" May 13 23:46:28.210257 containerd[1799]: time="2025-05-13T23:46:28.210132321Z" level=info msg="TaskExit event in podsandbox handler container_id:\"bb90cc4f73a26d5ffca30fa3edec52f9b95e3da688107845d1740628bfe3a731\" id:\"ec7bdf8ac22ad57138c8f6347276240b06ae94b710a46e6a0fe1ed5a57f73c09\" pid:5994 exited_at:{seconds:1747179988 nanos:209773201}" May 13 23:46:28.326944 sshd[5252]: Connection closed by 10.200.16.10 port 51132 May 13 23:46:28.326851 sshd-session[5207]: pam_unix(sshd:session): session closed for user core May 13 23:46:28.329768 systemd-logind[1744]: Session 32 logged out. Waiting for processes to exit. May 13 23:46:28.329970 systemd[1]: sshd@29-10.200.20.30:22-10.200.16.10:51132.service: Deactivated successfully. May 13 23:46:28.332061 systemd[1]: session-32.scope: Deactivated successfully. May 13 23:46:28.334109 systemd-logind[1744]: Removed session 32.