Aug 13 07:08:33.271413 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Aug 13 07:08:33.271433 kernel: Linux version 6.6.100-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241221 p7) 14.2.1 20241221, GNU ld (Gentoo 2.43 p3) 2.43.1) #1 SMP PREEMPT Tue Aug 12 21:42:02 -00 2025 Aug 13 07:08:33.271441 kernel: KASLR enabled Aug 13 07:08:33.271447 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Aug 13 07:08:33.271454 kernel: printk: bootconsole [pl11] enabled Aug 13 07:08:33.271459 kernel: efi: EFI v2.7 by EDK II Aug 13 07:08:33.271466 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3f20f698 RNG=0x3fd5f998 MEMRESERVE=0x3e477598 Aug 13 07:08:33.271472 kernel: random: crng init done Aug 13 07:08:33.271478 kernel: secureboot: Secure boot disabled Aug 13 07:08:33.271483 kernel: ACPI: Early table checksum verification disabled Aug 13 07:08:33.271489 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Aug 13 07:08:33.271495 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271501 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271508 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Aug 13 07:08:33.271515 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271521 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271528 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271535 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271541 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271547 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271553 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Aug 13 07:08:33.271559 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Aug 13 07:08:33.271565 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Aug 13 07:08:33.271571 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Aug 13 07:08:33.271578 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Aug 13 07:08:33.271584 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Aug 13 07:08:33.271590 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Aug 13 07:08:33.271596 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Aug 13 07:08:33.271603 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Aug 13 07:08:33.271610 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Aug 13 07:08:33.271616 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Aug 13 07:08:33.271622 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Aug 13 07:08:33.271628 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Aug 13 07:08:33.271634 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Aug 13 07:08:33.271640 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Aug 13 07:08:33.271646 kernel: NUMA: NODE_DATA [mem 0x1bf7ef800-0x1bf7f4fff] Aug 13 07:08:33.271652 kernel: Zone ranges: Aug 13 07:08:33.271658 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Aug 13 07:08:33.271664 kernel: DMA32 empty Aug 13 07:08:33.271670 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Aug 13 07:08:33.271680 kernel: Movable zone start for each node Aug 13 07:08:33.271687 kernel: Early memory node ranges Aug 13 07:08:33.271693 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Aug 13 07:08:33.271700 kernel: node 0: [mem 0x0000000000824000-0x000000003e45ffff] Aug 13 07:08:33.271706 kernel: node 0: [mem 0x000000003e460000-0x000000003e46ffff] Aug 13 07:08:33.271714 kernel: node 0: [mem 0x000000003e470000-0x000000003e54ffff] Aug 13 07:08:33.271720 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Aug 13 07:08:33.271727 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Aug 13 07:08:33.271733 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Aug 13 07:08:33.271740 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Aug 13 07:08:33.271746 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Aug 13 07:08:33.271753 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Aug 13 07:08:33.271759 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Aug 13 07:08:33.271765 kernel: psci: probing for conduit method from ACPI. Aug 13 07:08:33.271772 kernel: psci: PSCIv1.1 detected in firmware. Aug 13 07:08:33.271778 kernel: psci: Using standard PSCI v0.2 function IDs Aug 13 07:08:33.271784 kernel: psci: MIGRATE_INFO_TYPE not supported. Aug 13 07:08:33.271792 kernel: psci: SMC Calling Convention v1.4 Aug 13 07:08:33.271799 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Aug 13 07:08:33.271805 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Aug 13 07:08:33.271812 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Aug 13 07:08:33.271818 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Aug 13 07:08:33.271825 kernel: pcpu-alloc: [0] 0 [0] 1 Aug 13 07:08:33.271831 kernel: Detected PIPT I-cache on CPU0 Aug 13 07:08:33.271837 kernel: CPU features: detected: GIC system register CPU interface Aug 13 07:08:33.271844 kernel: CPU features: detected: Hardware dirty bit management Aug 13 07:08:33.272953 kernel: CPU features: detected: Spectre-BHB Aug 13 07:08:33.272963 kernel: CPU features: kernel page table isolation forced ON by KASLR Aug 13 07:08:33.272974 kernel: CPU features: detected: Kernel page table isolation (KPTI) Aug 13 07:08:33.272981 kernel: CPU features: detected: ARM erratum 1418040 Aug 13 07:08:33.272988 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Aug 13 07:08:33.272994 kernel: CPU features: detected: SSBS not fully self-synchronizing Aug 13 07:08:33.273001 kernel: alternatives: applying boot alternatives Aug 13 07:08:33.273008 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=c44ba8b4c0c81c1bcadc13a1606b9de202ee4e4226c47e1c865eaa5fc436b169 Aug 13 07:08:33.273016 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Aug 13 07:08:33.273022 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Aug 13 07:08:33.273029 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Aug 13 07:08:33.273035 kernel: Fallback order for Node 0: 0 Aug 13 07:08:33.273042 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Aug 13 07:08:33.273050 kernel: Policy zone: Normal Aug 13 07:08:33.273057 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Aug 13 07:08:33.273063 kernel: software IO TLB: area num 2. Aug 13 07:08:33.273069 kernel: software IO TLB: mapped [mem 0x0000000036530000-0x000000003a530000] (64MB) Aug 13 07:08:33.273076 kernel: Memory: 3983528K/4194160K available (10368K kernel code, 2186K rwdata, 8104K rodata, 38400K init, 897K bss, 210632K reserved, 0K cma-reserved) Aug 13 07:08:33.273083 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Aug 13 07:08:33.273089 kernel: rcu: Preemptible hierarchical RCU implementation. Aug 13 07:08:33.273096 kernel: rcu: RCU event tracing is enabled. Aug 13 07:08:33.273103 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Aug 13 07:08:33.273110 kernel: Trampoline variant of Tasks RCU enabled. Aug 13 07:08:33.273116 kernel: Tracing variant of Tasks RCU enabled. Aug 13 07:08:33.273124 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Aug 13 07:08:33.273131 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Aug 13 07:08:33.273137 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Aug 13 07:08:33.273144 kernel: GICv3: 960 SPIs implemented Aug 13 07:08:33.273150 kernel: GICv3: 0 Extended SPIs implemented Aug 13 07:08:33.273157 kernel: Root IRQ handler: gic_handle_irq Aug 13 07:08:33.273163 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Aug 13 07:08:33.273169 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Aug 13 07:08:33.273176 kernel: ITS: No ITS available, not enabling LPIs Aug 13 07:08:33.273182 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Aug 13 07:08:33.273189 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 07:08:33.273195 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Aug 13 07:08:33.273204 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Aug 13 07:08:33.273210 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Aug 13 07:08:33.273217 kernel: Console: colour dummy device 80x25 Aug 13 07:08:33.273224 kernel: printk: console [tty1] enabled Aug 13 07:08:33.273230 kernel: ACPI: Core revision 20230628 Aug 13 07:08:33.273237 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Aug 13 07:08:33.273244 kernel: pid_max: default: 32768 minimum: 301 Aug 13 07:08:33.273250 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Aug 13 07:08:33.273257 kernel: landlock: Up and running. Aug 13 07:08:33.273265 kernel: SELinux: Initializing. Aug 13 07:08:33.273272 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 07:08:33.273278 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Aug 13 07:08:33.273285 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 07:08:33.273292 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Aug 13 07:08:33.273299 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 Aug 13 07:08:33.273306 kernel: Hyper-V: Host Build 10.0.22477.1619-1-0 Aug 13 07:08:33.273319 kernel: Hyper-V: enabling crash_kexec_post_notifiers Aug 13 07:08:33.273325 kernel: rcu: Hierarchical SRCU implementation. Aug 13 07:08:33.273333 kernel: rcu: Max phase no-delay instances is 400. Aug 13 07:08:33.273340 kernel: Remapping and enabling EFI services. Aug 13 07:08:33.273347 kernel: smp: Bringing up secondary CPUs ... Aug 13 07:08:33.273355 kernel: Detected PIPT I-cache on CPU1 Aug 13 07:08:33.273362 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Aug 13 07:08:33.273369 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Aug 13 07:08:33.273377 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Aug 13 07:08:33.273384 kernel: smp: Brought up 1 node, 2 CPUs Aug 13 07:08:33.273392 kernel: SMP: Total of 2 processors activated. Aug 13 07:08:33.273399 kernel: CPU features: detected: 32-bit EL0 Support Aug 13 07:08:33.273407 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Aug 13 07:08:33.273414 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Aug 13 07:08:33.273421 kernel: CPU features: detected: CRC32 instructions Aug 13 07:08:33.273428 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Aug 13 07:08:33.273435 kernel: CPU features: detected: LSE atomic instructions Aug 13 07:08:33.273442 kernel: CPU features: detected: Privileged Access Never Aug 13 07:08:33.273449 kernel: CPU: All CPU(s) started at EL1 Aug 13 07:08:33.273457 kernel: alternatives: applying system-wide alternatives Aug 13 07:08:33.273464 kernel: devtmpfs: initialized Aug 13 07:08:33.273472 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Aug 13 07:08:33.273479 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Aug 13 07:08:33.273486 kernel: pinctrl core: initialized pinctrl subsystem Aug 13 07:08:33.273493 kernel: SMBIOS 3.1.0 present. Aug 13 07:08:33.273500 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Aug 13 07:08:33.273507 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Aug 13 07:08:33.273514 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Aug 13 07:08:33.273522 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Aug 13 07:08:33.273530 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Aug 13 07:08:33.273537 kernel: audit: initializing netlink subsys (disabled) Aug 13 07:08:33.273544 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Aug 13 07:08:33.273551 kernel: thermal_sys: Registered thermal governor 'step_wise' Aug 13 07:08:33.273557 kernel: cpuidle: using governor menu Aug 13 07:08:33.273564 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Aug 13 07:08:33.273571 kernel: ASID allocator initialised with 32768 entries Aug 13 07:08:33.273578 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Aug 13 07:08:33.273587 kernel: Serial: AMBA PL011 UART driver Aug 13 07:08:33.273594 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Aug 13 07:08:33.273600 kernel: Modules: 0 pages in range for non-PLT usage Aug 13 07:08:33.273607 kernel: Modules: 509248 pages in range for PLT usage Aug 13 07:08:33.273614 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Aug 13 07:08:33.273621 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Aug 13 07:08:33.273628 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Aug 13 07:08:33.273635 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Aug 13 07:08:33.273642 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Aug 13 07:08:33.273651 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Aug 13 07:08:33.273658 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Aug 13 07:08:33.273665 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Aug 13 07:08:33.273672 kernel: ACPI: Added _OSI(Module Device) Aug 13 07:08:33.273679 kernel: ACPI: Added _OSI(Processor Device) Aug 13 07:08:33.273686 kernel: ACPI: Added _OSI(Processor Aggregator Device) Aug 13 07:08:33.273693 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Aug 13 07:08:33.273700 kernel: ACPI: Interpreter enabled Aug 13 07:08:33.273707 kernel: ACPI: Using GIC for interrupt routing Aug 13 07:08:33.273716 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Aug 13 07:08:33.273723 kernel: printk: console [ttyAMA0] enabled Aug 13 07:08:33.273730 kernel: printk: bootconsole [pl11] disabled Aug 13 07:08:33.273737 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Aug 13 07:08:33.273744 kernel: iommu: Default domain type: Translated Aug 13 07:08:33.273751 kernel: iommu: DMA domain TLB invalidation policy: strict mode Aug 13 07:08:33.273758 kernel: efivars: Registered efivars operations Aug 13 07:08:33.273765 kernel: vgaarb: loaded Aug 13 07:08:33.273772 kernel: clocksource: Switched to clocksource arch_sys_counter Aug 13 07:08:33.273781 kernel: VFS: Disk quotas dquot_6.6.0 Aug 13 07:08:33.273788 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Aug 13 07:08:33.273795 kernel: pnp: PnP ACPI init Aug 13 07:08:33.273802 kernel: pnp: PnP ACPI: found 0 devices Aug 13 07:08:33.273809 kernel: NET: Registered PF_INET protocol family Aug 13 07:08:33.273816 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Aug 13 07:08:33.273823 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Aug 13 07:08:33.273830 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Aug 13 07:08:33.273837 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Aug 13 07:08:33.273855 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Aug 13 07:08:33.273864 kernel: TCP: Hash tables configured (established 32768 bind 32768) Aug 13 07:08:33.273871 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 07:08:33.273879 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Aug 13 07:08:33.273885 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Aug 13 07:08:33.273893 kernel: PCI: CLS 0 bytes, default 64 Aug 13 07:08:33.273899 kernel: kvm [1]: HYP mode not available Aug 13 07:08:33.273907 kernel: Initialise system trusted keyrings Aug 13 07:08:33.273914 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Aug 13 07:08:33.273922 kernel: Key type asymmetric registered Aug 13 07:08:33.273929 kernel: Asymmetric key parser 'x509' registered Aug 13 07:08:33.273936 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Aug 13 07:08:33.273943 kernel: io scheduler mq-deadline registered Aug 13 07:08:33.273950 kernel: io scheduler kyber registered Aug 13 07:08:33.273957 kernel: io scheduler bfq registered Aug 13 07:08:33.273964 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Aug 13 07:08:33.273971 kernel: thunder_xcv, ver 1.0 Aug 13 07:08:33.273978 kernel: thunder_bgx, ver 1.0 Aug 13 07:08:33.273987 kernel: nicpf, ver 1.0 Aug 13 07:08:33.273994 kernel: nicvf, ver 1.0 Aug 13 07:08:33.274126 kernel: rtc-efi rtc-efi.0: registered as rtc0 Aug 13 07:08:33.274195 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-08-13T07:08:32 UTC (1755068912) Aug 13 07:08:33.274205 kernel: efifb: probing for efifb Aug 13 07:08:33.274213 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Aug 13 07:08:33.274220 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Aug 13 07:08:33.274227 kernel: efifb: scrolling: redraw Aug 13 07:08:33.274236 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Aug 13 07:08:33.274243 kernel: Console: switching to colour frame buffer device 128x48 Aug 13 07:08:33.274250 kernel: fb0: EFI VGA frame buffer device Aug 13 07:08:33.274257 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Aug 13 07:08:33.274264 kernel: hid: raw HID events driver (C) Jiri Kosina Aug 13 07:08:33.274271 kernel: No ACPI PMU IRQ for CPU0 Aug 13 07:08:33.274278 kernel: No ACPI PMU IRQ for CPU1 Aug 13 07:08:33.274285 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available Aug 13 07:08:33.274292 kernel: watchdog: Delayed init of the lockup detector failed: -19 Aug 13 07:08:33.274301 kernel: watchdog: Hard watchdog permanently disabled Aug 13 07:08:33.274308 kernel: NET: Registered PF_INET6 protocol family Aug 13 07:08:33.274315 kernel: Segment Routing with IPv6 Aug 13 07:08:33.274322 kernel: In-situ OAM (IOAM) with IPv6 Aug 13 07:08:33.274329 kernel: NET: Registered PF_PACKET protocol family Aug 13 07:08:33.274336 kernel: Key type dns_resolver registered Aug 13 07:08:33.274343 kernel: registered taskstats version 1 Aug 13 07:08:33.274349 kernel: Loading compiled-in X.509 certificates Aug 13 07:08:33.274357 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.100-flatcar: b805f03ae64b71ea1aa3cf76d07ec816116f6d0c' Aug 13 07:08:33.274365 kernel: Key type .fscrypt registered Aug 13 07:08:33.274372 kernel: Key type fscrypt-provisioning registered Aug 13 07:08:33.274379 kernel: ima: No TPM chip found, activating TPM-bypass! Aug 13 07:08:33.274386 kernel: ima: Allocated hash algorithm: sha1 Aug 13 07:08:33.274393 kernel: ima: No architecture policies found Aug 13 07:08:33.274400 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Aug 13 07:08:33.274407 kernel: clk: Disabling unused clocks Aug 13 07:08:33.274415 kernel: Freeing unused kernel memory: 38400K Aug 13 07:08:33.274422 kernel: Run /init as init process Aug 13 07:08:33.274430 kernel: with arguments: Aug 13 07:08:33.274437 kernel: /init Aug 13 07:08:33.274443 kernel: with environment: Aug 13 07:08:33.274450 kernel: HOME=/ Aug 13 07:08:33.274457 kernel: TERM=linux Aug 13 07:08:33.274464 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Aug 13 07:08:33.274472 systemd[1]: Successfully made /usr/ read-only. Aug 13 07:08:33.274482 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Aug 13 07:08:33.274492 systemd[1]: Detected virtualization microsoft. Aug 13 07:08:33.274499 systemd[1]: Detected architecture arm64. Aug 13 07:08:33.274506 systemd[1]: Running in initrd. Aug 13 07:08:33.274514 systemd[1]: No hostname configured, using default hostname. Aug 13 07:08:33.274522 systemd[1]: Hostname set to . Aug 13 07:08:33.274529 systemd[1]: Initializing machine ID from random generator. Aug 13 07:08:33.274537 systemd[1]: Queued start job for default target initrd.target. Aug 13 07:08:33.274544 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:08:33.274554 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:08:33.274562 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Aug 13 07:08:33.274570 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:08:33.274578 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Aug 13 07:08:33.274586 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Aug 13 07:08:33.274595 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Aug 13 07:08:33.274604 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Aug 13 07:08:33.274612 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:08:33.274619 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:08:33.274627 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:08:33.274634 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:08:33.274642 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:08:33.274650 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:08:33.274657 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:08:33.274665 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:08:33.274674 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Aug 13 07:08:33.274682 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Aug 13 07:08:33.274689 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:08:33.274697 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:08:33.274705 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:08:33.274712 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:08:33.274720 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Aug 13 07:08:33.274728 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:08:33.274735 systemd[1]: Finished network-cleanup.service - Network Cleanup. Aug 13 07:08:33.274745 systemd[1]: Starting systemd-fsck-usr.service... Aug 13 07:08:33.274752 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:08:33.274760 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:08:33.274783 systemd-journald[218]: Collecting audit messages is disabled. Aug 13 07:08:33.274803 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:33.274812 systemd-journald[218]: Journal started Aug 13 07:08:33.274829 systemd-journald[218]: Runtime Journal (/run/log/journal/e39424a895ff4cc29cf2d247cd1ba8dc) is 8M, max 78.5M, 70.5M free. Aug 13 07:08:33.278892 systemd-modules-load[220]: Inserted module 'overlay' Aug 13 07:08:33.293797 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:08:33.294369 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Aug 13 07:08:33.301406 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:08:33.333134 systemd[1]: Finished systemd-fsck-usr.service. Aug 13 07:08:33.353016 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Aug 13 07:08:33.353039 kernel: Bridge firewalling registered Aug 13 07:08:33.348431 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:33.352239 systemd-modules-load[220]: Inserted module 'br_netfilter' Aug 13 07:08:33.360870 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:08:33.381088 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:33.403001 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:08:33.410999 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Aug 13 07:08:33.441000 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:08:33.449773 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:33.463273 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:08:33.469174 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Aug 13 07:08:33.484461 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:08:33.513040 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Aug 13 07:08:33.520959 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:08:33.538734 dracut-cmdline[252]: dracut-dracut-053 Aug 13 07:08:33.538734 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=c44ba8b4c0c81c1bcadc13a1606b9de202ee4e4226c47e1c865eaa5fc436b169 Aug 13 07:08:33.543287 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:08:33.587072 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:08:33.621513 systemd-resolved[253]: Positive Trust Anchors: Aug 13 07:08:33.625937 systemd-resolved[253]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:08:33.625970 systemd-resolved[253]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:08:33.628592 systemd-resolved[253]: Defaulting to hostname 'linux'. Aug 13 07:08:33.629416 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:08:33.636024 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:08:33.727872 kernel: SCSI subsystem initialized Aug 13 07:08:33.735867 kernel: Loading iSCSI transport class v2.0-870. Aug 13 07:08:33.745875 kernel: iscsi: registered transport (tcp) Aug 13 07:08:33.762613 kernel: iscsi: registered transport (qla4xxx) Aug 13 07:08:33.762630 kernel: QLogic iSCSI HBA Driver Aug 13 07:08:33.793967 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Aug 13 07:08:33.807043 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Aug 13 07:08:33.839442 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Aug 13 07:08:33.839493 kernel: device-mapper: uevent: version 1.0.3 Aug 13 07:08:33.846076 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Aug 13 07:08:33.893873 kernel: raid6: neonx8 gen() 15789 MB/s Aug 13 07:08:33.913865 kernel: raid6: neonx4 gen() 15842 MB/s Aug 13 07:08:33.933856 kernel: raid6: neonx2 gen() 13226 MB/s Aug 13 07:08:33.956857 kernel: raid6: neonx1 gen() 10523 MB/s Aug 13 07:08:33.976855 kernel: raid6: int64x8 gen() 6796 MB/s Aug 13 07:08:33.996855 kernel: raid6: int64x4 gen() 7357 MB/s Aug 13 07:08:34.017856 kernel: raid6: int64x2 gen() 6118 MB/s Aug 13 07:08:34.041561 kernel: raid6: int64x1 gen() 5059 MB/s Aug 13 07:08:34.041580 kernel: raid6: using algorithm neonx4 gen() 15842 MB/s Aug 13 07:08:34.065903 kernel: raid6: .... xor() 12392 MB/s, rmw enabled Aug 13 07:08:34.065919 kernel: raid6: using neon recovery algorithm Aug 13 07:08:34.076879 kernel: xor: measuring software checksum speed Aug 13 07:08:34.076893 kernel: 8regs : 21607 MB/sec Aug 13 07:08:34.080274 kernel: 32regs : 21664 MB/sec Aug 13 07:08:34.083636 kernel: arm64_neon : 27927 MB/sec Aug 13 07:08:34.087754 kernel: xor: using function: arm64_neon (27927 MB/sec) Aug 13 07:08:34.136871 kernel: Btrfs loaded, zoned=no, fsverity=no Aug 13 07:08:34.145978 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:08:34.161013 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:08:34.184791 systemd-udevd[438]: Using default interface naming scheme 'v255'. Aug 13 07:08:34.191599 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:08:34.216085 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Aug 13 07:08:34.233415 dracut-pre-trigger[455]: rd.md=0: removing MD RAID activation Aug 13 07:08:34.259508 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:08:34.274288 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:08:34.312890 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:08:34.333047 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Aug 13 07:08:34.350501 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Aug 13 07:08:34.361105 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:08:34.379007 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:08:34.393488 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:08:34.410905 kernel: hv_vmbus: Vmbus version:5.3 Aug 13 07:08:34.411979 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Aug 13 07:08:34.432985 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:08:34.483522 kernel: hv_vmbus: registering driver hid_hyperv Aug 13 07:08:34.483546 kernel: hv_vmbus: registering driver hyperv_keyboard Aug 13 07:08:34.483556 kernel: pps_core: LinuxPPS API ver. 1 registered Aug 13 07:08:34.483565 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Aug 13 07:08:34.483582 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input0 Aug 13 07:08:34.483591 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input1 Aug 13 07:08:34.483601 kernel: hv_vmbus: registering driver hv_storvsc Aug 13 07:08:34.433199 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:34.510953 kernel: scsi host1: storvsc_host_t Aug 13 07:08:34.511123 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Aug 13 07:08:34.511215 kernel: scsi host0: storvsc_host_t Aug 13 07:08:34.511236 kernel: hv_vmbus: registering driver hv_netvsc Aug 13 07:08:34.467706 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:34.529367 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Aug 13 07:08:34.522731 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:08:34.522916 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:34.563926 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Aug 13 07:08:34.541789 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:34.584625 kernel: PTP clock support registered Aug 13 07:08:34.584648 kernel: hv_utils: Registering HyperV Utility Driver Aug 13 07:08:34.574520 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:34.745663 kernel: hv_vmbus: registering driver hv_utils Aug 13 07:08:34.745694 kernel: hv_utils: Heartbeat IC version 3.0 Aug 13 07:08:34.745704 kernel: hv_utils: Shutdown IC version 3.2 Aug 13 07:08:34.745713 kernel: hv_utils: TimeSync IC version 4.0 Aug 13 07:08:34.745722 kernel: hv_netvsc 000d3afe-f8e0-000d-3afe-f8e0000d3afe eth0: VF slot 1 added Aug 13 07:08:34.595440 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:08:34.735498 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:34.776389 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Aug 13 07:08:34.776560 kernel: hv_vmbus: registering driver hv_pci Aug 13 07:08:34.776571 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Aug 13 07:08:34.735572 systemd-resolved[253]: Clock change detected. Flushing caches. Aug 13 07:08:34.818635 kernel: hv_pci c946019d-2713-49e9-aeaa-e69641b94389: PCI VMBus probing: Using version 0x10004 Aug 13 07:08:34.818816 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Aug 13 07:08:34.818935 kernel: hv_pci c946019d-2713-49e9-aeaa-e69641b94389: PCI host bridge to bus 2713:00 Aug 13 07:08:34.819011 kernel: pci_bus 2713:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Aug 13 07:08:34.819102 kernel: pci_bus 2713:00: No busn resource found for root bus, will use [bus 00-ff] Aug 13 07:08:34.786824 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:34.882482 kernel: pci 2713:00:02.0: [15b3:1018] type 00 class 0x020000 Aug 13 07:08:34.882537 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Aug 13 07:08:34.882699 kernel: pci 2713:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Aug 13 07:08:34.882715 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Aug 13 07:08:34.882800 kernel: pci 2713:00:02.0: enabling Extended Tags Aug 13 07:08:34.882815 kernel: sd 0:0:0:0: [sda] Write Protect is off Aug 13 07:08:34.882897 kernel: pci 2713:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 2713:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Aug 13 07:08:34.882915 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Aug 13 07:08:34.811979 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:08:34.913900 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Aug 13 07:08:34.918384 kernel: pci_bus 2713:00: busn_res: [bus 00-ff] end is updated to 00 Aug 13 07:08:34.918504 kernel: pci 2713:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Aug 13 07:08:34.918640 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Aug 13 07:08:34.812079 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:34.827766 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:08:34.950066 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Aug 13 07:08:34.827823 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:34.870494 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:34.925499 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:34.951385 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:34.989710 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Aug 13 07:08:35.015808 kernel: mlx5_core 2713:00:02.0: enabling device (0000 -> 0002) Aug 13 07:08:35.014065 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:35.034600 kernel: mlx5_core 2713:00:02.0: firmware version: 16.31.2424 Aug 13 07:08:35.306386 kernel: hv_netvsc 000d3afe-f8e0-000d-3afe-f8e0000d3afe eth0: VF registering: eth1 Aug 13 07:08:35.312309 kernel: mlx5_core 2713:00:02.0 eth1: joined to eth0 Aug 13 07:08:35.321352 kernel: mlx5_core 2713:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Aug 13 07:08:35.331294 kernel: mlx5_core 2713:00:02.0 enP10003s1: renamed from eth1 Aug 13 07:08:35.449413 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Aug 13 07:08:35.500513 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (495) Aug 13 07:08:35.517175 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Aug 13 07:08:35.542300 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Aug 13 07:08:35.636298 kernel: BTRFS: device fsid 66ef7c2c-768e-46b2-8baa-a2b24df44a90 devid 1 transid 42 /dev/sda3 scanned by (udev-worker) (489) Aug 13 07:08:35.651974 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Aug 13 07:08:35.659313 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Aug 13 07:08:35.691423 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Aug 13 07:08:35.713314 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Aug 13 07:08:35.721296 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Aug 13 07:08:36.729751 disk-uuid[612]: The operation has completed successfully. Aug 13 07:08:36.735062 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Aug 13 07:08:36.786917 systemd[1]: disk-uuid.service: Deactivated successfully. Aug 13 07:08:36.788301 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Aug 13 07:08:36.837408 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Aug 13 07:08:36.851311 sh[698]: Success Aug 13 07:08:36.880588 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Aug 13 07:08:37.059468 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Aug 13 07:08:37.082409 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Aug 13 07:08:37.092318 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Aug 13 07:08:37.123518 kernel: BTRFS info (device dm-0): first mount of filesystem 66ef7c2c-768e-46b2-8baa-a2b24df44a90 Aug 13 07:08:37.123558 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Aug 13 07:08:37.130442 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Aug 13 07:08:37.135516 kernel: BTRFS info (device dm-0): disabling log replay at mount time Aug 13 07:08:37.139754 kernel: BTRFS info (device dm-0): using free space tree Aug 13 07:08:37.403519 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Aug 13 07:08:37.408933 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Aug 13 07:08:37.429452 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Aug 13 07:08:37.435434 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Aug 13 07:08:37.473436 kernel: BTRFS info (device sda6): first mount of filesystem 5832a3b0-f866-4304-b935-a4d38424b8f9 Aug 13 07:08:37.473477 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Aug 13 07:08:37.477512 kernel: BTRFS info (device sda6): using free space tree Aug 13 07:08:37.498549 kernel: BTRFS info (device sda6): auto enabling async discard Aug 13 07:08:37.508362 kernel: BTRFS info (device sda6): last unmount of filesystem 5832a3b0-f866-4304-b935-a4d38424b8f9 Aug 13 07:08:37.519744 systemd[1]: Finished ignition-setup.service - Ignition (setup). Aug 13 07:08:37.534522 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Aug 13 07:08:37.570639 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:08:37.589426 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:08:37.617793 systemd-networkd[879]: lo: Link UP Aug 13 07:08:37.617807 systemd-networkd[879]: lo: Gained carrier Aug 13 07:08:37.619455 systemd-networkd[879]: Enumeration completed Aug 13 07:08:37.620995 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:08:37.627774 systemd-networkd[879]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:37.627777 systemd-networkd[879]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:08:37.628044 systemd[1]: Reached target network.target - Network. Aug 13 07:08:37.693299 kernel: mlx5_core 2713:00:02.0 enP10003s1: Link up Aug 13 07:08:37.771665 kernel: hv_netvsc 000d3afe-f8e0-000d-3afe-f8e0000d3afe eth0: Data path switched to VF: enP10003s1 Aug 13 07:08:37.771385 systemd-networkd[879]: enP10003s1: Link UP Aug 13 07:08:37.771461 systemd-networkd[879]: eth0: Link UP Aug 13 07:08:37.771551 systemd-networkd[879]: eth0: Gained carrier Aug 13 07:08:37.771559 systemd-networkd[879]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:37.782590 systemd-networkd[879]: enP10003s1: Gained carrier Aug 13 07:08:37.806322 systemd-networkd[879]: eth0: DHCPv4 address 10.200.20.42/24, gateway 10.200.20.1 acquired from 168.63.129.16 Aug 13 07:08:38.420813 ignition[838]: Ignition 2.20.0 Aug 13 07:08:38.420823 ignition[838]: Stage: fetch-offline Aug 13 07:08:38.422555 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:08:38.420859 ignition[838]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:38.437472 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Aug 13 07:08:38.420867 ignition[838]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:38.420967 ignition[838]: parsed url from cmdline: "" Aug 13 07:08:38.420970 ignition[838]: no config URL provided Aug 13 07:08:38.420975 ignition[838]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 07:08:38.420982 ignition[838]: no config at "/usr/lib/ignition/user.ign" Aug 13 07:08:38.420987 ignition[838]: failed to fetch config: resource requires networking Aug 13 07:08:38.421155 ignition[838]: Ignition finished successfully Aug 13 07:08:38.459968 ignition[888]: Ignition 2.20.0 Aug 13 07:08:38.459974 ignition[888]: Stage: fetch Aug 13 07:08:38.460190 ignition[888]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:38.460200 ignition[888]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:38.460309 ignition[888]: parsed url from cmdline: "" Aug 13 07:08:38.460313 ignition[888]: no config URL provided Aug 13 07:08:38.460318 ignition[888]: reading system config file "/usr/lib/ignition/user.ign" Aug 13 07:08:38.460325 ignition[888]: no config at "/usr/lib/ignition/user.ign" Aug 13 07:08:38.460349 ignition[888]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Aug 13 07:08:38.591312 ignition[888]: GET result: OK Aug 13 07:08:38.591375 ignition[888]: config has been read from IMDS userdata Aug 13 07:08:38.591417 ignition[888]: parsing config with SHA512: c04595e23e17ba59ab2b8660f1ce63e50c695e182aae806c022d5afa0f2b7cea1116bd53e8f0a56be84b31637283ababd8542af51f34b52992ef2bce397ba87b Aug 13 07:08:38.595406 unknown[888]: fetched base config from "system" Aug 13 07:08:38.595767 ignition[888]: fetch: fetch complete Aug 13 07:08:38.595413 unknown[888]: fetched base config from "system" Aug 13 07:08:38.595772 ignition[888]: fetch: fetch passed Aug 13 07:08:38.595418 unknown[888]: fetched user config from "azure" Aug 13 07:08:38.595811 ignition[888]: Ignition finished successfully Aug 13 07:08:38.599696 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Aug 13 07:08:38.621468 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Aug 13 07:08:38.642502 ignition[895]: Ignition 2.20.0 Aug 13 07:08:38.642517 ignition[895]: Stage: kargs Aug 13 07:08:38.642675 ignition[895]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:38.650306 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Aug 13 07:08:38.642689 ignition[895]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:38.643779 ignition[895]: kargs: kargs passed Aug 13 07:08:38.643833 ignition[895]: Ignition finished successfully Aug 13 07:08:38.675449 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Aug 13 07:08:38.690776 ignition[902]: Ignition 2.20.0 Aug 13 07:08:38.691317 ignition[902]: Stage: disks Aug 13 07:08:38.696303 systemd[1]: Finished ignition-disks.service - Ignition (disks). Aug 13 07:08:38.691488 ignition[902]: no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:38.701849 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Aug 13 07:08:38.691499 ignition[902]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:38.712014 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Aug 13 07:08:38.692398 ignition[902]: disks: disks passed Aug 13 07:08:38.722457 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:08:38.692441 ignition[902]: Ignition finished successfully Aug 13 07:08:38.732782 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:08:38.743492 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:08:38.771501 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Aug 13 07:08:38.837184 systemd-fsck[911]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Aug 13 07:08:38.844310 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Aug 13 07:08:38.864441 systemd[1]: Mounting sysroot.mount - /sysroot... Aug 13 07:08:38.920303 kernel: EXT4-fs (sda9): mounted filesystem 4e885a6c-f4f3-43a5-b152-e0e8bd6b099d r/w with ordered data mode. Quota mode: none. Aug 13 07:08:38.921347 systemd[1]: Mounted sysroot.mount - /sysroot. Aug 13 07:08:38.925764 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Aug 13 07:08:38.966348 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:08:38.973415 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Aug 13 07:08:38.985913 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Aug 13 07:08:38.992255 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Aug 13 07:08:38.992342 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:08:39.045819 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (922) Aug 13 07:08:39.045841 kernel: BTRFS info (device sda6): first mount of filesystem 5832a3b0-f866-4304-b935-a4d38424b8f9 Aug 13 07:08:39.045857 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Aug 13 07:08:39.013891 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Aug 13 07:08:39.054101 kernel: BTRFS info (device sda6): using free space tree Aug 13 07:08:39.046481 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Aug 13 07:08:39.071293 kernel: BTRFS info (device sda6): auto enabling async discard Aug 13 07:08:39.071790 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:08:39.357379 systemd-networkd[879]: eth0: Gained IPv6LL Aug 13 07:08:39.570490 coreos-metadata[924]: Aug 13 07:08:39.570 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Aug 13 07:08:39.580687 coreos-metadata[924]: Aug 13 07:08:39.580 INFO Fetch successful Aug 13 07:08:39.580687 coreos-metadata[924]: Aug 13 07:08:39.580 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Aug 13 07:08:39.595494 coreos-metadata[924]: Aug 13 07:08:39.592 INFO Fetch successful Aug 13 07:08:39.608091 coreos-metadata[924]: Aug 13 07:08:39.608 INFO wrote hostname ci-4230.2.2-a-ad5fc6e7e6 to /sysroot/etc/hostname Aug 13 07:08:39.617306 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 07:08:39.741888 initrd-setup-root[953]: cut: /sysroot/etc/passwd: No such file or directory Aug 13 07:08:39.765022 initrd-setup-root[960]: cut: /sysroot/etc/group: No such file or directory Aug 13 07:08:39.771703 initrd-setup-root[967]: cut: /sysroot/etc/shadow: No such file or directory Aug 13 07:08:39.779684 initrd-setup-root[974]: cut: /sysroot/etc/gshadow: No such file or directory Aug 13 07:08:40.580821 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Aug 13 07:08:40.594429 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Aug 13 07:08:40.606611 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Aug 13 07:08:40.619343 kernel: BTRFS info (device sda6): last unmount of filesystem 5832a3b0-f866-4304-b935-a4d38424b8f9 Aug 13 07:08:40.628221 systemd[1]: sysroot-oem.mount: Deactivated successfully. Aug 13 07:08:40.643591 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Aug 13 07:08:40.657150 ignition[1044]: INFO : Ignition 2.20.0 Aug 13 07:08:40.657150 ignition[1044]: INFO : Stage: mount Aug 13 07:08:40.670614 ignition[1044]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:40.670614 ignition[1044]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:40.670614 ignition[1044]: INFO : mount: mount passed Aug 13 07:08:40.670614 ignition[1044]: INFO : Ignition finished successfully Aug 13 07:08:40.661978 systemd[1]: Finished ignition-mount.service - Ignition (mount). Aug 13 07:08:40.686437 systemd[1]: Starting ignition-files.service - Ignition (files)... Aug 13 07:08:40.702465 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Aug 13 07:08:40.733556 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (1053) Aug 13 07:08:40.733591 kernel: BTRFS info (device sda6): first mount of filesystem 5832a3b0-f866-4304-b935-a4d38424b8f9 Aug 13 07:08:40.739185 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Aug 13 07:08:40.742942 kernel: BTRFS info (device sda6): using free space tree Aug 13 07:08:40.749307 kernel: BTRFS info (device sda6): auto enabling async discard Aug 13 07:08:40.750421 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Aug 13 07:08:40.775260 ignition[1071]: INFO : Ignition 2.20.0 Aug 13 07:08:40.775260 ignition[1071]: INFO : Stage: files Aug 13 07:08:40.782468 ignition[1071]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:40.782468 ignition[1071]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:40.782468 ignition[1071]: DEBUG : files: compiled without relabeling support, skipping Aug 13 07:08:40.799767 ignition[1071]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Aug 13 07:08:40.799767 ignition[1071]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Aug 13 07:08:40.817245 ignition[1071]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Aug 13 07:08:40.824457 ignition[1071]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Aug 13 07:08:40.824457 ignition[1071]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Aug 13 07:08:40.824457 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Aug 13 07:08:40.824457 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Aug 13 07:08:40.817665 unknown[1071]: wrote ssh authorized keys file for user: core Aug 13 07:08:40.995124 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Aug 13 07:08:41.568956 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Aug 13 07:08:41.568956 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Aug 13 07:08:41.588572 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Aug 13 07:08:41.767677 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Aug 13 07:08:41.845250 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 07:08:41.854225 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-arm64.raw: attempt #1 Aug 13 07:08:42.314334 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Aug 13 07:08:42.518356 ignition[1071]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-arm64.raw" Aug 13 07:08:42.529946 ignition[1071]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Aug 13 07:08:42.550748 ignition[1071]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Aug 13 07:08:42.562343 ignition[1071]: INFO : files: files passed Aug 13 07:08:42.562343 ignition[1071]: INFO : Ignition finished successfully Aug 13 07:08:42.561248 systemd[1]: Finished ignition-files.service - Ignition (files). Aug 13 07:08:42.590485 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Aug 13 07:08:42.604438 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Aug 13 07:08:42.637445 systemd[1]: ignition-quench.service: Deactivated successfully. Aug 13 07:08:42.637533 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Aug 13 07:08:42.672619 initrd-setup-root-after-ignition[1099]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:08:42.672619 initrd-setup-root-after-ignition[1099]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:08:42.695220 initrd-setup-root-after-ignition[1103]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Aug 13 07:08:42.673766 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:08:42.687692 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Aug 13 07:08:42.719451 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Aug 13 07:08:42.752832 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Aug 13 07:08:42.752942 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Aug 13 07:08:42.771229 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Aug 13 07:08:42.776856 systemd[1]: Reached target initrd.target - Initrd Default Target. Aug 13 07:08:42.787207 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Aug 13 07:08:42.805466 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Aug 13 07:08:42.819231 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:08:42.832697 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Aug 13 07:08:42.849147 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:08:42.855428 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:08:42.866966 systemd[1]: Stopped target timers.target - Timer Units. Aug 13 07:08:42.877096 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Aug 13 07:08:42.877209 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Aug 13 07:08:42.892249 systemd[1]: Stopped target initrd.target - Initrd Default Target. Aug 13 07:08:42.903452 systemd[1]: Stopped target basic.target - Basic System. Aug 13 07:08:42.912601 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Aug 13 07:08:42.922441 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Aug 13 07:08:42.933700 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Aug 13 07:08:42.945049 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Aug 13 07:08:42.955395 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Aug 13 07:08:42.966455 systemd[1]: Stopped target sysinit.target - System Initialization. Aug 13 07:08:42.977744 systemd[1]: Stopped target local-fs.target - Local File Systems. Aug 13 07:08:42.987837 systemd[1]: Stopped target swap.target - Swaps. Aug 13 07:08:42.997159 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Aug 13 07:08:42.997346 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Aug 13 07:08:43.012189 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:08:43.023129 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:08:43.034145 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Aug 13 07:08:43.038303 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:08:43.046185 systemd[1]: dracut-initqueue.service: Deactivated successfully. Aug 13 07:08:43.046379 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Aug 13 07:08:43.063742 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Aug 13 07:08:43.063909 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Aug 13 07:08:43.076459 systemd[1]: ignition-files.service: Deactivated successfully. Aug 13 07:08:43.076613 systemd[1]: Stopped ignition-files.service - Ignition (files). Aug 13 07:08:43.086981 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Aug 13 07:08:43.087122 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Aug 13 07:08:43.117402 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Aug 13 07:08:43.138516 ignition[1123]: INFO : Ignition 2.20.0 Aug 13 07:08:43.138516 ignition[1123]: INFO : Stage: umount Aug 13 07:08:43.138516 ignition[1123]: INFO : no configs at "/usr/lib/ignition/base.d" Aug 13 07:08:43.138516 ignition[1123]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Aug 13 07:08:43.138516 ignition[1123]: INFO : umount: umount passed Aug 13 07:08:43.138516 ignition[1123]: INFO : Ignition finished successfully Aug 13 07:08:43.132240 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Aug 13 07:08:43.137352 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:08:43.159096 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Aug 13 07:08:43.172968 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Aug 13 07:08:43.173175 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:08:43.185669 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Aug 13 07:08:43.185823 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Aug 13 07:08:43.204821 systemd[1]: ignition-mount.service: Deactivated successfully. Aug 13 07:08:43.206299 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Aug 13 07:08:43.214354 systemd[1]: sysroot-boot.mount: Deactivated successfully. Aug 13 07:08:43.216972 systemd[1]: initrd-cleanup.service: Deactivated successfully. Aug 13 07:08:43.217265 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Aug 13 07:08:43.231695 systemd[1]: ignition-disks.service: Deactivated successfully. Aug 13 07:08:43.231751 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Aug 13 07:08:43.241017 systemd[1]: ignition-kargs.service: Deactivated successfully. Aug 13 07:08:43.241061 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Aug 13 07:08:43.251569 systemd[1]: ignition-fetch.service: Deactivated successfully. Aug 13 07:08:43.251611 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Aug 13 07:08:43.261143 systemd[1]: Stopped target network.target - Network. Aug 13 07:08:43.270771 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Aug 13 07:08:43.270821 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Aug 13 07:08:43.281265 systemd[1]: Stopped target paths.target - Path Units. Aug 13 07:08:43.290217 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Aug 13 07:08:43.295299 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:08:43.308346 systemd[1]: Stopped target slices.target - Slice Units. Aug 13 07:08:43.318571 systemd[1]: Stopped target sockets.target - Socket Units. Aug 13 07:08:43.334004 systemd[1]: iscsid.socket: Deactivated successfully. Aug 13 07:08:43.334053 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Aug 13 07:08:43.344501 systemd[1]: iscsiuio.socket: Deactivated successfully. Aug 13 07:08:43.344529 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Aug 13 07:08:43.354725 systemd[1]: ignition-setup.service: Deactivated successfully. Aug 13 07:08:43.354776 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Aug 13 07:08:43.364981 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Aug 13 07:08:43.365023 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Aug 13 07:08:43.376425 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Aug 13 07:08:43.387920 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Aug 13 07:08:43.398890 systemd[1]: systemd-resolved.service: Deactivated successfully. Aug 13 07:08:43.398975 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Aug 13 07:08:43.419143 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Aug 13 07:08:43.419418 systemd[1]: systemd-networkd.service: Deactivated successfully. Aug 13 07:08:43.419521 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Aug 13 07:08:43.429204 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Aug 13 07:08:43.429424 systemd[1]: sysroot-boot.service: Deactivated successfully. Aug 13 07:08:43.429511 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Aug 13 07:08:43.438982 systemd[1]: systemd-networkd.socket: Deactivated successfully. Aug 13 07:08:43.439040 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:08:43.448082 systemd[1]: initrd-setup-root.service: Deactivated successfully. Aug 13 07:08:43.448142 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Aug 13 07:08:43.478454 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Aug 13 07:08:43.487543 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Aug 13 07:08:43.487644 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Aug 13 07:08:43.498999 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:08:43.499053 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:08:43.515222 systemd[1]: systemd-modules-load.service: Deactivated successfully. Aug 13 07:08:43.682492 kernel: hv_netvsc 000d3afe-f8e0-000d-3afe-f8e0000d3afe eth0: Data path switched from VF: enP10003s1 Aug 13 07:08:43.515265 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Aug 13 07:08:43.521073 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Aug 13 07:08:43.521114 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:08:43.537391 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:08:43.547307 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Aug 13 07:08:43.547381 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Aug 13 07:08:43.558736 systemd[1]: systemd-udevd.service: Deactivated successfully. Aug 13 07:08:43.559761 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:08:43.570178 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Aug 13 07:08:43.570242 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Aug 13 07:08:43.581260 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Aug 13 07:08:43.581404 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:08:43.591084 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Aug 13 07:08:43.591128 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Aug 13 07:08:43.605552 systemd[1]: dracut-cmdline.service: Deactivated successfully. Aug 13 07:08:43.605596 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Aug 13 07:08:43.622159 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Aug 13 07:08:43.622203 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Aug 13 07:08:43.662731 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Aug 13 07:08:43.677396 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Aug 13 07:08:43.677465 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:08:43.693993 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Aug 13 07:08:43.694045 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:43.705780 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Aug 13 07:08:43.705842 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Aug 13 07:08:43.706185 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Aug 13 07:08:43.706337 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Aug 13 07:08:43.717515 systemd[1]: network-cleanup.service: Deactivated successfully. Aug 13 07:08:43.717604 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Aug 13 07:08:43.729099 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Aug 13 07:08:43.758475 systemd[1]: Starting initrd-switch-root.service - Switch Root... Aug 13 07:08:43.790540 systemd[1]: Switching root. Aug 13 07:08:43.902745 systemd-journald[218]: Received SIGTERM from PID 1 (systemd). Aug 13 07:08:43.902783 systemd-journald[218]: Journal stopped Aug 13 07:08:48.950791 kernel: mlx5_core 2713:00:02.0: poll_health:835:(pid 0): device's health compromised - reached miss count Aug 13 07:08:48.950816 kernel: SELinux: policy capability network_peer_controls=1 Aug 13 07:08:48.950829 kernel: SELinux: policy capability open_perms=1 Aug 13 07:08:48.950837 kernel: SELinux: policy capability extended_socket_class=1 Aug 13 07:08:48.950845 kernel: SELinux: policy capability always_check_network=0 Aug 13 07:08:48.950852 kernel: SELinux: policy capability cgroup_seclabel=1 Aug 13 07:08:48.950861 kernel: SELinux: policy capability nnp_nosuid_transition=1 Aug 13 07:08:48.950869 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Aug 13 07:08:48.950876 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Aug 13 07:08:48.950884 kernel: audit: type=1403 audit(1755068925.056:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Aug 13 07:08:48.950895 systemd[1]: Successfully loaded SELinux policy in 160.468ms. Aug 13 07:08:48.950904 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 10.300ms. Aug 13 07:08:48.950914 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Aug 13 07:08:48.950923 systemd[1]: Detected virtualization microsoft. Aug 13 07:08:48.950934 systemd[1]: Detected architecture arm64. Aug 13 07:08:48.950942 systemd[1]: Detected first boot. Aug 13 07:08:48.950951 systemd[1]: Hostname set to . Aug 13 07:08:48.950959 systemd[1]: Initializing machine ID from random generator. Aug 13 07:08:48.950968 zram_generator::config[1166]: No configuration found. Aug 13 07:08:48.950977 kernel: NET: Registered PF_VSOCK protocol family Aug 13 07:08:48.950987 systemd[1]: Populated /etc with preset unit settings. Aug 13 07:08:48.950997 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Aug 13 07:08:48.951006 systemd[1]: initrd-switch-root.service: Deactivated successfully. Aug 13 07:08:48.951015 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Aug 13 07:08:48.951023 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Aug 13 07:08:48.951032 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Aug 13 07:08:48.951041 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Aug 13 07:08:48.951050 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Aug 13 07:08:48.951060 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Aug 13 07:08:48.951069 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Aug 13 07:08:48.951078 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Aug 13 07:08:48.951087 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Aug 13 07:08:48.951096 systemd[1]: Created slice user.slice - User and Session Slice. Aug 13 07:08:48.951105 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Aug 13 07:08:48.951113 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Aug 13 07:08:48.951122 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Aug 13 07:08:48.951132 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Aug 13 07:08:48.951142 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Aug 13 07:08:48.951151 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Aug 13 07:08:48.951159 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Aug 13 07:08:48.951171 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Aug 13 07:08:48.951181 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Aug 13 07:08:48.951189 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Aug 13 07:08:48.951198 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Aug 13 07:08:48.951209 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Aug 13 07:08:48.951218 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Aug 13 07:08:48.951227 systemd[1]: Reached target remote-fs.target - Remote File Systems. Aug 13 07:08:48.951236 systemd[1]: Reached target slices.target - Slice Units. Aug 13 07:08:48.951244 systemd[1]: Reached target swap.target - Swaps. Aug 13 07:08:48.951253 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Aug 13 07:08:48.951262 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Aug 13 07:08:48.951271 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Aug 13 07:08:48.951301 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Aug 13 07:08:48.951311 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Aug 13 07:08:48.951321 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Aug 13 07:08:48.951330 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Aug 13 07:08:48.951339 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Aug 13 07:08:48.951349 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Aug 13 07:08:48.951359 systemd[1]: Mounting media.mount - External Media Directory... Aug 13 07:08:48.951368 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Aug 13 07:08:48.951377 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Aug 13 07:08:48.951386 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Aug 13 07:08:48.951396 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Aug 13 07:08:48.951406 systemd[1]: Reached target machines.target - Containers. Aug 13 07:08:48.951415 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Aug 13 07:08:48.951425 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:08:48.951435 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Aug 13 07:08:48.951444 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Aug 13 07:08:48.951453 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:08:48.951465 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:08:48.951474 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:08:48.951483 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Aug 13 07:08:48.951492 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:08:48.951503 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Aug 13 07:08:48.951512 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Aug 13 07:08:48.951521 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Aug 13 07:08:48.951530 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Aug 13 07:08:48.951539 systemd[1]: Stopped systemd-fsck-usr.service. Aug 13 07:08:48.951548 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Aug 13 07:08:48.951558 kernel: fuse: init (API version 7.39) Aug 13 07:08:48.951566 systemd[1]: Starting systemd-journald.service - Journal Service... Aug 13 07:08:48.951576 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Aug 13 07:08:48.951586 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Aug 13 07:08:48.951596 kernel: loop: module loaded Aug 13 07:08:48.951604 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Aug 13 07:08:48.951613 kernel: ACPI: bus type drm_connector registered Aug 13 07:08:48.951621 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Aug 13 07:08:48.951649 systemd-journald[1270]: Collecting audit messages is disabled. Aug 13 07:08:48.951670 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Aug 13 07:08:48.951680 systemd-journald[1270]: Journal started Aug 13 07:08:48.951699 systemd-journald[1270]: Runtime Journal (/run/log/journal/284e92d8f2a848d694744eee9f7e8f13) is 8M, max 78.5M, 70.5M free. Aug 13 07:08:48.043488 systemd[1]: Queued start job for default target multi-user.target. Aug 13 07:08:48.054014 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Aug 13 07:08:48.054445 systemd[1]: systemd-journald.service: Deactivated successfully. Aug 13 07:08:48.055486 systemd[1]: systemd-journald.service: Consumed 3.062s CPU time. Aug 13 07:08:48.971762 systemd[1]: verity-setup.service: Deactivated successfully. Aug 13 07:08:48.971819 systemd[1]: Stopped verity-setup.service. Aug 13 07:08:48.989012 systemd[1]: Started systemd-journald.service - Journal Service. Aug 13 07:08:48.989822 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Aug 13 07:08:48.995249 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Aug 13 07:08:49.000937 systemd[1]: Mounted media.mount - External Media Directory. Aug 13 07:08:49.005877 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Aug 13 07:08:49.011890 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Aug 13 07:08:49.017748 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Aug 13 07:08:49.022796 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Aug 13 07:08:49.029409 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Aug 13 07:08:49.036312 systemd[1]: modprobe@configfs.service: Deactivated successfully. Aug 13 07:08:49.036477 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Aug 13 07:08:49.042834 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:08:49.042986 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:08:49.049053 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:08:49.049209 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:08:49.054854 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:08:49.055005 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:08:49.063818 systemd[1]: modprobe@fuse.service: Deactivated successfully. Aug 13 07:08:49.063969 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Aug 13 07:08:49.069732 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:08:49.069886 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:08:49.076147 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Aug 13 07:08:49.082792 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Aug 13 07:08:49.089684 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Aug 13 07:08:49.096380 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Aug 13 07:08:49.103233 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Aug 13 07:08:49.121667 systemd[1]: Reached target network-pre.target - Preparation for Network. Aug 13 07:08:49.132361 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Aug 13 07:08:49.139259 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Aug 13 07:08:49.145181 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Aug 13 07:08:49.145221 systemd[1]: Reached target local-fs.target - Local File Systems. Aug 13 07:08:49.151523 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Aug 13 07:08:49.158881 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Aug 13 07:08:49.166005 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Aug 13 07:08:49.171665 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:08:49.172973 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Aug 13 07:08:49.179423 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Aug 13 07:08:49.185684 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:08:49.188450 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Aug 13 07:08:49.194530 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:08:49.197453 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:08:49.204519 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Aug 13 07:08:49.215450 systemd[1]: Starting systemd-sysusers.service - Create System Users... Aug 13 07:08:49.225543 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Aug 13 07:08:49.239026 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Aug 13 07:08:49.247787 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Aug 13 07:08:49.254260 systemd-journald[1270]: Time spent on flushing to /var/log/journal/284e92d8f2a848d694744eee9f7e8f13 is 21.447ms for 920 entries. Aug 13 07:08:49.254260 systemd-journald[1270]: System Journal (/var/log/journal/284e92d8f2a848d694744eee9f7e8f13) is 8M, max 2.6G, 2.6G free. Aug 13 07:08:49.305545 systemd-journald[1270]: Received client request to flush runtime journal. Aug 13 07:08:49.305579 kernel: loop0: detected capacity change from 0 to 123192 Aug 13 07:08:49.260095 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Aug 13 07:08:49.269116 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Aug 13 07:08:49.280173 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Aug 13 07:08:49.295578 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Aug 13 07:08:49.302527 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:08:49.309353 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Aug 13 07:08:49.319753 udevadm[1309]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Aug 13 07:08:49.838961 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Aug 13 07:08:49.839624 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Aug 13 07:08:49.972340 systemd[1]: Finished systemd-sysusers.service - Create System Users. Aug 13 07:08:49.987999 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Aug 13 07:08:50.090540 systemd-tmpfiles[1323]: ACLs are not supported, ignoring. Aug 13 07:08:50.090937 systemd-tmpfiles[1323]: ACLs are not supported, ignoring. Aug 13 07:08:50.092309 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Aug 13 07:08:50.095495 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Aug 13 07:08:50.448303 kernel: loop1: detected capacity change from 0 to 113512 Aug 13 07:08:50.775298 kernel: loop2: detected capacity change from 0 to 28720 Aug 13 07:08:52.015061 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Aug 13 07:08:52.025170 kernel: loop3: detected capacity change from 0 to 203944 Aug 13 07:08:52.036423 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Aug 13 07:08:52.057127 systemd-udevd[1332]: Using default interface naming scheme 'v255'. Aug 13 07:08:52.059299 kernel: loop4: detected capacity change from 0 to 123192 Aug 13 07:08:52.068296 kernel: loop5: detected capacity change from 0 to 113512 Aug 13 07:08:52.076301 kernel: loop6: detected capacity change from 0 to 28720 Aug 13 07:08:52.084294 kernel: loop7: detected capacity change from 0 to 203944 Aug 13 07:08:52.088401 (sd-merge)[1333]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Aug 13 07:08:52.088808 (sd-merge)[1333]: Merged extensions into '/usr'. Aug 13 07:08:52.091626 systemd[1]: Reload requested from client PID 1306 ('systemd-sysext') (unit systemd-sysext.service)... Aug 13 07:08:52.091638 systemd[1]: Reloading... Aug 13 07:08:52.151311 zram_generator::config[1358]: No configuration found. Aug 13 07:08:52.284571 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:08:52.353750 systemd[1]: Reloading finished in 261 ms. Aug 13 07:08:52.371172 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Aug 13 07:08:52.384354 systemd[1]: Starting ensure-sysext.service... Aug 13 07:08:52.389444 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Aug 13 07:08:52.432531 systemd-tmpfiles[1417]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Aug 13 07:08:52.433079 systemd-tmpfiles[1417]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Aug 13 07:08:52.433769 systemd-tmpfiles[1417]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Aug 13 07:08:52.433985 systemd-tmpfiles[1417]: ACLs are not supported, ignoring. Aug 13 07:08:52.434027 systemd-tmpfiles[1417]: ACLs are not supported, ignoring. Aug 13 07:08:52.450310 systemd-tmpfiles[1417]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:08:52.450322 systemd-tmpfiles[1417]: Skipping /boot Aug 13 07:08:52.458560 systemd-tmpfiles[1417]: Detected autofs mount point /boot during canonicalization of boot. Aug 13 07:08:52.458574 systemd-tmpfiles[1417]: Skipping /boot Aug 13 07:08:52.465249 systemd[1]: Reload requested from client PID 1416 ('systemctl') (unit ensure-sysext.service)... Aug 13 07:08:52.465263 systemd[1]: Reloading... Aug 13 07:08:52.577527 zram_generator::config[1471]: No configuration found. Aug 13 07:08:52.685120 kernel: mousedev: PS/2 mouse device common for all mice Aug 13 07:08:52.708857 kernel: hv_vmbus: registering driver hv_balloon Aug 13 07:08:52.708942 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Aug 13 07:08:52.720302 kernel: hv_balloon: Memory hot add disabled on ARM64 Aug 13 07:08:52.720374 kernel: hv_vmbus: registering driver hyperv_fb Aug 13 07:08:52.723290 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Aug 13 07:08:52.731291 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Aug 13 07:08:52.737833 kernel: Console: switching to colour dummy device 80x25 Aug 13 07:08:52.740306 kernel: Console: switching to colour frame buffer device 128x48 Aug 13 07:08:52.746809 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:08:52.836610 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Aug 13 07:08:52.836753 systemd[1]: Reloading finished in 371 ms. Aug 13 07:08:52.849266 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Aug 13 07:08:52.865575 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Aug 13 07:08:52.896480 systemd[1]: Finished ensure-sysext.service. Aug 13 07:08:52.915494 systemd[1]: Starting audit-rules.service - Load Audit Rules... Aug 13 07:08:52.923449 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Aug 13 07:08:52.932862 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Aug 13 07:08:52.934464 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Aug 13 07:08:52.945417 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 42 scanned by (udev-worker) (1427) Aug 13 07:08:52.956329 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Aug 13 07:08:52.974578 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Aug 13 07:08:52.989472 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Aug 13 07:08:52.999526 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Aug 13 07:08:52.999578 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Aug 13 07:08:53.003629 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Aug 13 07:08:53.031554 systemd[1]: Starting systemd-networkd.service - Network Configuration... Aug 13 07:08:53.045482 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Aug 13 07:08:53.052510 systemd[1]: Reached target time-set.target - System Time Set. Aug 13 07:08:53.059644 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Aug 13 07:08:53.067364 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Aug 13 07:08:53.073797 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Aug 13 07:08:53.075320 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Aug 13 07:08:53.081708 systemd[1]: modprobe@drm.service: Deactivated successfully. Aug 13 07:08:53.081890 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Aug 13 07:08:53.088908 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Aug 13 07:08:53.089074 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Aug 13 07:08:53.096820 systemd[1]: modprobe@loop.service: Deactivated successfully. Aug 13 07:08:53.096975 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Aug 13 07:08:53.126003 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Aug 13 07:08:53.144174 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Aug 13 07:08:53.156456 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Aug 13 07:08:53.164331 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Aug 13 07:08:53.164388 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Aug 13 07:08:53.167441 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Aug 13 07:08:53.176740 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Aug 13 07:08:53.198441 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Aug 13 07:08:53.214705 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Aug 13 07:08:53.220896 systemd[1]: Started systemd-userdbd.service - User Database Manager. Aug 13 07:08:53.556780 augenrules[1653]: No rules Aug 13 07:08:53.558239 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 07:08:53.558486 systemd[1]: Finished audit-rules.service - Load Audit Rules. Aug 13 07:08:53.568047 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Aug 13 07:08:53.694957 lvm[1642]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:08:53.722314 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Aug 13 07:08:53.729484 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Aug 13 07:08:53.740633 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Aug 13 07:08:53.743967 lvm[1661]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Aug 13 07:08:53.766439 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Aug 13 07:08:53.800807 systemd-resolved[1619]: Positive Trust Anchors: Aug 13 07:08:53.800830 systemd-resolved[1619]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Aug 13 07:08:53.800862 systemd-resolved[1619]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Aug 13 07:08:53.944107 systemd-resolved[1619]: Using system hostname 'ci-4230.2.2-a-ad5fc6e7e6'. Aug 13 07:08:53.945498 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Aug 13 07:08:53.951111 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Aug 13 07:08:53.989776 systemd-networkd[1609]: lo: Link UP Aug 13 07:08:53.989788 systemd-networkd[1609]: lo: Gained carrier Aug 13 07:08:53.991848 systemd-networkd[1609]: Enumeration completed Aug 13 07:08:53.991981 systemd[1]: Started systemd-networkd.service - Network Configuration. Aug 13 07:08:53.992159 systemd-networkd[1609]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:53.992162 systemd-networkd[1609]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:08:53.997806 systemd[1]: Reached target network.target - Network. Aug 13 07:08:54.007431 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Aug 13 07:08:54.014908 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Aug 13 07:08:54.065296 kernel: mlx5_core 2713:00:02.0 enP10003s1: Link up Aug 13 07:08:54.108310 kernel: hv_netvsc 000d3afe-f8e0-000d-3afe-f8e0000d3afe eth0: Data path switched to VF: enP10003s1 Aug 13 07:08:54.109479 systemd-networkd[1609]: enP10003s1: Link UP Aug 13 07:08:54.109570 systemd-networkd[1609]: eth0: Link UP Aug 13 07:08:54.109573 systemd-networkd[1609]: eth0: Gained carrier Aug 13 07:08:54.109587 systemd-networkd[1609]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:08:54.114867 systemd-networkd[1609]: enP10003s1: Gained carrier Aug 13 07:08:54.124318 systemd-networkd[1609]: eth0: DHCPv4 address 10.200.20.42/24, gateway 10.200.20.1 acquired from 168.63.129.16 Aug 13 07:08:54.166022 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Aug 13 07:08:54.215722 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Aug 13 07:08:54.710935 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Aug 13 07:08:54.717934 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Aug 13 07:08:55.421451 systemd-networkd[1609]: eth0: Gained IPv6LL Aug 13 07:08:55.424299 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Aug 13 07:08:55.431175 systemd[1]: Reached target network-online.target - Network is Online. Aug 13 07:08:58.419299 ldconfig[1301]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Aug 13 07:08:58.432976 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Aug 13 07:08:58.443489 systemd[1]: Starting systemd-update-done.service - Update is Completed... Aug 13 07:08:58.456131 systemd[1]: Finished systemd-update-done.service - Update is Completed. Aug 13 07:08:58.462207 systemd[1]: Reached target sysinit.target - System Initialization. Aug 13 07:08:58.467623 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Aug 13 07:08:58.474463 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Aug 13 07:08:58.480902 systemd[1]: Started logrotate.timer - Daily rotation of log files. Aug 13 07:08:58.486307 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Aug 13 07:08:58.493132 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Aug 13 07:08:58.499378 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Aug 13 07:08:58.499408 systemd[1]: Reached target paths.target - Path Units. Aug 13 07:08:58.503876 systemd[1]: Reached target timers.target - Timer Units. Aug 13 07:08:58.509702 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Aug 13 07:08:58.516845 systemd[1]: Starting docker.socket - Docker Socket for the API... Aug 13 07:08:58.523656 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Aug 13 07:08:58.530159 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Aug 13 07:08:58.536447 systemd[1]: Reached target ssh-access.target - SSH Access Available. Aug 13 07:08:58.548921 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Aug 13 07:08:58.555643 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Aug 13 07:08:58.562042 systemd[1]: Listening on docker.socket - Docker Socket for the API. Aug 13 07:08:58.567412 systemd[1]: Reached target sockets.target - Socket Units. Aug 13 07:08:58.572444 systemd[1]: Reached target basic.target - Basic System. Aug 13 07:08:58.577206 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:08:58.577230 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Aug 13 07:08:58.588368 systemd[1]: Starting chronyd.service - NTP client/server... Aug 13 07:08:58.596455 systemd[1]: Starting containerd.service - containerd container runtime... Aug 13 07:08:58.606579 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Aug 13 07:08:58.614457 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Aug 13 07:08:58.623404 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Aug 13 07:08:58.629704 (chronyd)[1678]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Aug 13 07:08:58.632459 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Aug 13 07:08:58.633377 jq[1685]: false Aug 13 07:08:58.638723 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Aug 13 07:08:58.638768 systemd[1]: hv_fcopy_daemon.service - Hyper-V FCOPY daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_fcopy). Aug 13 07:08:58.639817 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Aug 13 07:08:58.647422 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Aug 13 07:08:58.648444 KVP[1687]: KVP starting; pid is:1687 Aug 13 07:08:58.649948 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:08:58.655290 chronyd[1690]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Aug 13 07:08:58.658799 KVP[1687]: KVP LIC Version: 3.1 Aug 13 07:08:58.661313 kernel: hv_utils: KVP IC version 4.0 Aug 13 07:08:58.662446 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Aug 13 07:08:58.670469 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Aug 13 07:08:58.676372 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Aug 13 07:08:58.683925 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Aug 13 07:08:58.695443 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Aug 13 07:08:58.704803 chronyd[1690]: Timezone right/UTC failed leap second check, ignoring Aug 13 07:08:58.704951 chronyd[1690]: Loaded seccomp filter (level 2) Aug 13 07:08:58.707930 systemd[1]: Starting systemd-logind.service - User Login Management... Aug 13 07:08:58.714898 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Aug 13 07:08:58.715545 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Aug 13 07:08:58.716638 systemd[1]: Starting update-engine.service - Update Engine... Aug 13 07:08:58.723842 extend-filesystems[1686]: Found loop4 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found loop5 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found loop6 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found loop7 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda1 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda2 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda3 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found usr Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda4 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda6 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda7 Aug 13 07:08:58.735121 extend-filesystems[1686]: Found sda9 Aug 13 07:08:58.735121 extend-filesystems[1686]: Checking size of /dev/sda9 Aug 13 07:08:58.732446 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Aug 13 07:08:58.791546 dbus-daemon[1681]: [system] SELinux support is enabled Aug 13 07:08:58.851865 extend-filesystems[1686]: Old size kept for /dev/sda9 Aug 13 07:08:58.851865 extend-filesystems[1686]: Found sr0 Aug 13 07:08:58.867087 update_engine[1708]: I20250813 07:08:58.790233 1708 main.cc:92] Flatcar Update Engine starting Aug 13 07:08:58.867087 update_engine[1708]: I20250813 07:08:58.804102 1708 update_check_scheduler.cc:74] Next update check in 11m1s Aug 13 07:08:58.751499 systemd[1]: Started chronyd.service - NTP client/server. Aug 13 07:08:58.867487 jq[1709]: true Aug 13 07:08:58.777004 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Aug 13 07:08:58.777185 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Aug 13 07:08:58.778203 systemd[1]: motdgen.service: Deactivated successfully. Aug 13 07:08:58.778420 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Aug 13 07:08:58.792731 systemd[1]: Started dbus.service - D-Bus System Message Bus. Aug 13 07:08:58.818307 systemd[1]: extend-filesystems.service: Deactivated successfully. Aug 13 07:08:58.819328 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Aug 13 07:08:58.831207 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Aug 13 07:08:58.860503 systemd-logind[1702]: New seat seat0. Aug 13 07:08:58.861949 systemd-logind[1702]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard) Aug 13 07:08:58.890621 systemd[1]: Started systemd-logind.service - User Login Management. Aug 13 07:08:58.898898 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Aug 13 07:08:58.899095 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Aug 13 07:08:58.926287 jq[1740]: true Aug 13 07:08:58.926147 dbus-daemon[1681]: [system] Successfully activated service 'org.freedesktop.systemd1' Aug 13 07:08:58.932073 (ntainerd)[1741]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Aug 13 07:08:58.937404 systemd[1]: Started update-engine.service - Update Engine. Aug 13 07:08:58.949970 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Aug 13 07:08:58.950118 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Aug 13 07:08:58.961247 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Aug 13 07:08:58.961528 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Aug 13 07:08:58.982310 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 42 scanned by (udev-worker) (1736) Aug 13 07:08:58.984636 systemd[1]: Started locksmithd.service - Cluster reboot manager. Aug 13 07:08:59.101026 coreos-metadata[1680]: Aug 13 07:08:59.100 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Aug 13 07:08:59.105547 coreos-metadata[1680]: Aug 13 07:08:59.105 INFO Fetch successful Aug 13 07:08:59.105547 coreos-metadata[1680]: Aug 13 07:08:59.105 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Aug 13 07:08:59.110548 coreos-metadata[1680]: Aug 13 07:08:59.110 INFO Fetch successful Aug 13 07:08:59.110986 coreos-metadata[1680]: Aug 13 07:08:59.110 INFO Fetching http://168.63.129.16/machine/ed42ca63-c986-44b5-96e8-f9ff183fb09f/6334fc56%2D8bb0%2D4366%2D8682%2Dba8e94a58823.%5Fci%2D4230.2.2%2Da%2Dad5fc6e7e6?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Aug 13 07:08:59.112958 coreos-metadata[1680]: Aug 13 07:08:59.112 INFO Fetch successful Aug 13 07:08:59.113085 coreos-metadata[1680]: Aug 13 07:08:59.113 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Aug 13 07:08:59.123910 coreos-metadata[1680]: Aug 13 07:08:59.123 INFO Fetch successful Aug 13 07:08:59.155367 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Aug 13 07:08:59.165532 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Aug 13 07:08:59.166588 tar[1730]: linux-arm64/helm Aug 13 07:08:59.812620 locksmithd[1762]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Aug 13 07:08:59.998697 bash[1785]: Updated "/home/core/.ssh/authorized_keys" Aug 13 07:08:59.990012 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Aug 13 07:08:59.997834 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Aug 13 07:09:00.042852 tar[1730]: linux-arm64/LICENSE Aug 13 07:09:00.042852 tar[1730]: linux-arm64/README.md Aug 13 07:09:00.056499 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Aug 13 07:09:00.279118 sshd_keygen[1715]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Aug 13 07:09:00.305539 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Aug 13 07:09:00.318507 systemd[1]: Starting issuegen.service - Generate /run/issue... Aug 13 07:09:00.325134 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Aug 13 07:09:00.333173 systemd[1]: issuegen.service: Deactivated successfully. Aug 13 07:09:00.334118 systemd[1]: Finished issuegen.service - Generate /run/issue. Aug 13 07:09:00.349906 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Aug 13 07:09:00.364481 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Aug 13 07:09:00.372261 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Aug 13 07:09:00.395210 systemd[1]: Started getty@tty1.service - Getty on tty1. Aug 13 07:09:00.401225 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Aug 13 07:09:00.407487 systemd[1]: Reached target getty.target - Login Prompts. Aug 13 07:09:00.427575 (kubelet)[1863]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:00.427778 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:00.510474 containerd[1741]: time="2025-08-13T07:09:00.510226140Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Aug 13 07:09:00.535238 containerd[1741]: time="2025-08-13T07:09:00.535145940Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.536513 containerd[1741]: time="2025-08-13T07:09:00.536479780Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.100-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:09:00.537316 containerd[1741]: time="2025-08-13T07:09:00.537116780Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Aug 13 07:09:00.537316 containerd[1741]: time="2025-08-13T07:09:00.537145740Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Aug 13 07:09:00.537316 containerd[1741]: time="2025-08-13T07:09:00.537309380Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Aug 13 07:09:00.537417 containerd[1741]: time="2025-08-13T07:09:00.537328140Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.537417 containerd[1741]: time="2025-08-13T07:09:00.537389100Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:09:00.537417 containerd[1741]: time="2025-08-13T07:09:00.537401340Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538343 containerd[1741]: time="2025-08-13T07:09:00.538315900Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538378 containerd[1741]: time="2025-08-13T07:09:00.538343380Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538378 containerd[1741]: time="2025-08-13T07:09:00.538358060Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538378 containerd[1741]: time="2025-08-13T07:09:00.538367780Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538461 containerd[1741]: time="2025-08-13T07:09:00.538448180Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538656 containerd[1741]: time="2025-08-13T07:09:00.538633900Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538780 containerd[1741]: time="2025-08-13T07:09:00.538760740Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Aug 13 07:09:00.538805 containerd[1741]: time="2025-08-13T07:09:00.538780580Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Aug 13 07:09:00.538871 containerd[1741]: time="2025-08-13T07:09:00.538854900Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Aug 13 07:09:00.538915 containerd[1741]: time="2025-08-13T07:09:00.538899220Z" level=info msg="metadata content store policy set" policy=shared Aug 13 07:09:00.780444 kubelet[1863]: E0813 07:09:00.780360 1863 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:00.782450 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:00.782577 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:00.783014 systemd[1]: kubelet.service: Consumed 697ms CPU time, 256.4M memory peak. Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088324820Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088397500Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088417620Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088433820Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088448780Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088613260Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088844380Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088932620Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088946420Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088959780Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088973380Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088985220Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.088996340Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090331 containerd[1741]: time="2025-08-13T07:09:01.089010020Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089025540Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089038460Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089050500Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089061260Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089081220Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089094180Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089105980Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089118140Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089129340Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089141940Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089152980Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089165420Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089177140Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090642 containerd[1741]: time="2025-08-13T07:09:01.089190660Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089201140Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089212420Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089224220Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089241220Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089262420Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089305660Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089320340Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089366940Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089386180Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089396460Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089408900Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089417300Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089428860Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Aug 13 07:09:01.090868 containerd[1741]: time="2025-08-13T07:09:01.089445100Z" level=info msg="NRI interface is disabled by configuration." Aug 13 07:09:01.091092 containerd[1741]: time="2025-08-13T07:09:01.089458460Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.089749140Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.089801460Z" level=info msg="Connect containerd service" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.089834300Z" level=info msg="using legacy CRI server" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.089840700Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.089954020Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090601340Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090764180Z" level=info msg="Start subscribing containerd event" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090801700Z" level=info msg="Start recovering state" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090856180Z" level=info msg="Start event monitor" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090865780Z" level=info msg="Start snapshots syncer" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090873660Z" level=info msg="Start cni network conf syncer for default" Aug 13 07:09:01.091111 containerd[1741]: time="2025-08-13T07:09:01.090880780Z" level=info msg="Start streaming server" Aug 13 07:09:01.091395 containerd[1741]: time="2025-08-13T07:09:01.091238620Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Aug 13 07:09:01.091395 containerd[1741]: time="2025-08-13T07:09:01.091327300Z" level=info msg=serving... address=/run/containerd/containerd.sock Aug 13 07:09:01.091461 systemd[1]: Started containerd.service - containerd container runtime. Aug 13 07:09:01.094216 containerd[1741]: time="2025-08-13T07:09:01.094094860Z" level=info msg="containerd successfully booted in 0.585174s" Aug 13 07:09:01.097465 systemd[1]: Reached target multi-user.target - Multi-User System. Aug 13 07:09:01.106345 systemd[1]: Startup finished in 631ms (kernel) + 12.043s (initrd) + 16.208s (userspace) = 28.884s. Aug 13 07:09:01.570825 login[1861]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:01.572698 login[1862]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:01.581434 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Aug 13 07:09:01.586496 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Aug 13 07:09:01.588762 systemd-logind[1702]: New session 2 of user core. Aug 13 07:09:01.594153 systemd-logind[1702]: New session 1 of user core. Aug 13 07:09:01.599327 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Aug 13 07:09:01.606544 systemd[1]: Starting user@500.service - User Manager for UID 500... Aug 13 07:09:01.609145 (systemd)[1880]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Aug 13 07:09:01.611474 systemd-logind[1702]: New session c1 of user core. Aug 13 07:09:01.864566 systemd[1880]: Queued start job for default target default.target. Aug 13 07:09:01.874125 systemd[1880]: Created slice app.slice - User Application Slice. Aug 13 07:09:01.874294 systemd[1880]: Reached target paths.target - Paths. Aug 13 07:09:01.874417 systemd[1880]: Reached target timers.target - Timers. Aug 13 07:09:01.875659 systemd[1880]: Starting dbus.socket - D-Bus User Message Bus Socket... Aug 13 07:09:01.884144 systemd[1880]: Listening on dbus.socket - D-Bus User Message Bus Socket. Aug 13 07:09:01.884195 systemd[1880]: Reached target sockets.target - Sockets. Aug 13 07:09:01.884229 systemd[1880]: Reached target basic.target - Basic System. Aug 13 07:09:01.884260 systemd[1880]: Reached target default.target - Main User Target. Aug 13 07:09:01.884311 systemd[1880]: Startup finished in 267ms. Aug 13 07:09:01.884492 systemd[1]: Started user@500.service - User Manager for UID 500. Aug 13 07:09:01.886400 systemd[1]: Started session-1.scope - Session 1 of User core. Aug 13 07:09:01.887124 systemd[1]: Started session-2.scope - Session 2 of User core. Aug 13 07:09:02.988105 waagent[1855]: 2025-08-13T07:09:02.988017Z INFO Daemon Daemon Azure Linux Agent Version: 2.9.1.1 Aug 13 07:09:02.993218 waagent[1855]: 2025-08-13T07:09:02.993163Z INFO Daemon Daemon OS: flatcar 4230.2.2 Aug 13 07:09:02.997528 waagent[1855]: 2025-08-13T07:09:02.997484Z INFO Daemon Daemon Python: 3.11.11 Aug 13 07:09:03.001917 waagent[1855]: 2025-08-13T07:09:03.001725Z INFO Daemon Daemon Run daemon Aug 13 07:09:03.005991 waagent[1855]: 2025-08-13T07:09:03.005947Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4230.2.2' Aug 13 07:09:03.014631 waagent[1855]: 2025-08-13T07:09:03.014583Z INFO Daemon Daemon Using waagent for provisioning Aug 13 07:09:03.019682 waagent[1855]: 2025-08-13T07:09:03.019642Z INFO Daemon Daemon Activate resource disk Aug 13 07:09:03.024242 waagent[1855]: 2025-08-13T07:09:03.024203Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Aug 13 07:09:03.036345 waagent[1855]: 2025-08-13T07:09:03.036300Z INFO Daemon Daemon Found device: None Aug 13 07:09:03.040319 waagent[1855]: 2025-08-13T07:09:03.040267Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Aug 13 07:09:03.048348 waagent[1855]: 2025-08-13T07:09:03.048309Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Aug 13 07:09:03.058734 waagent[1855]: 2025-08-13T07:09:03.058692Z INFO Daemon Daemon Clean protocol and wireserver endpoint Aug 13 07:09:03.064735 waagent[1855]: 2025-08-13T07:09:03.064696Z INFO Daemon Daemon Running default provisioning handler Aug 13 07:09:03.075784 waagent[1855]: 2025-08-13T07:09:03.075725Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Aug 13 07:09:03.089007 waagent[1855]: 2025-08-13T07:09:03.088957Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Aug 13 07:09:03.098867 waagent[1855]: 2025-08-13T07:09:03.098524Z INFO Daemon Daemon cloud-init is enabled: False Aug 13 07:09:03.106208 waagent[1855]: 2025-08-13T07:09:03.106146Z INFO Daemon Daemon Copying ovf-env.xml Aug 13 07:09:03.169340 waagent[1855]: 2025-08-13T07:09:03.169226Z INFO Daemon Daemon Successfully mounted dvd Aug 13 07:09:03.198526 waagent[1855]: 2025-08-13T07:09:03.198440Z INFO Daemon Daemon Detect protocol endpoint Aug 13 07:09:03.199388 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Aug 13 07:09:03.203328 waagent[1855]: 2025-08-13T07:09:03.203262Z INFO Daemon Daemon Clean protocol and wireserver endpoint Aug 13 07:09:03.208837 waagent[1855]: 2025-08-13T07:09:03.208786Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Aug 13 07:09:03.214989 waagent[1855]: 2025-08-13T07:09:03.214942Z INFO Daemon Daemon Test for route to 168.63.129.16 Aug 13 07:09:03.220184 waagent[1855]: 2025-08-13T07:09:03.220137Z INFO Daemon Daemon Route to 168.63.129.16 exists Aug 13 07:09:03.224951 waagent[1855]: 2025-08-13T07:09:03.224907Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Aug 13 07:09:03.256796 waagent[1855]: 2025-08-13T07:09:03.256708Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Aug 13 07:09:03.263431 waagent[1855]: 2025-08-13T07:09:03.263401Z INFO Daemon Daemon Wire protocol version:2012-11-30 Aug 13 07:09:03.268828 waagent[1855]: 2025-08-13T07:09:03.268780Z INFO Daemon Daemon Server preferred version:2015-04-05 Aug 13 07:09:03.552315 waagent[1855]: 2025-08-13T07:09:03.551492Z INFO Daemon Daemon Initializing goal state during protocol detection Aug 13 07:09:03.558352 waagent[1855]: 2025-08-13T07:09:03.558301Z INFO Daemon Daemon Forcing an update of the goal state. Aug 13 07:09:03.566930 waagent[1855]: 2025-08-13T07:09:03.566879Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Aug 13 07:09:03.585792 waagent[1855]: 2025-08-13T07:09:03.585748Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.175 Aug 13 07:09:03.591166 waagent[1855]: 2025-08-13T07:09:03.591117Z INFO Daemon Aug 13 07:09:03.593822 waagent[1855]: 2025-08-13T07:09:03.593777Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: 99c4b8df-2144-42b3-9ca8-6f5c474fa7e6 eTag: 1798490211998072079 source: Fabric] Aug 13 07:09:03.604715 waagent[1855]: 2025-08-13T07:09:03.604668Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Aug 13 07:09:03.611136 waagent[1855]: 2025-08-13T07:09:03.611088Z INFO Daemon Aug 13 07:09:03.613840 waagent[1855]: 2025-08-13T07:09:03.613795Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Aug 13 07:09:03.623793 waagent[1855]: 2025-08-13T07:09:03.623753Z INFO Daemon Daemon Downloading artifacts profile blob Aug 13 07:09:03.764493 waagent[1855]: 2025-08-13T07:09:03.764418Z INFO Daemon Downloaded certificate {'thumbprint': '9D5CD3A456091BFC2CC4D35626F9D5772C302DE5', 'hasPrivateKey': False} Aug 13 07:09:03.773894 waagent[1855]: 2025-08-13T07:09:03.773845Z INFO Daemon Downloaded certificate {'thumbprint': '1DA5BF181F45A40662141BA2CE4E21F780FAFED1', 'hasPrivateKey': True} Aug 13 07:09:03.783038 waagent[1855]: 2025-08-13T07:09:03.782988Z INFO Daemon Fetch goal state completed Aug 13 07:09:03.822735 waagent[1855]: 2025-08-13T07:09:03.822582Z INFO Daemon Daemon Starting provisioning Aug 13 07:09:03.827822 waagent[1855]: 2025-08-13T07:09:03.827770Z INFO Daemon Daemon Handle ovf-env.xml. Aug 13 07:09:03.832541 waagent[1855]: 2025-08-13T07:09:03.832497Z INFO Daemon Daemon Set hostname [ci-4230.2.2-a-ad5fc6e7e6] Aug 13 07:09:03.845300 waagent[1855]: 2025-08-13T07:09:03.840339Z INFO Daemon Daemon Publish hostname [ci-4230.2.2-a-ad5fc6e7e6] Aug 13 07:09:03.846692 waagent[1855]: 2025-08-13T07:09:03.846642Z INFO Daemon Daemon Examine /proc/net/route for primary interface Aug 13 07:09:03.852754 waagent[1855]: 2025-08-13T07:09:03.852708Z INFO Daemon Daemon Primary interface is [eth0] Aug 13 07:09:03.864309 systemd-networkd[1609]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Aug 13 07:09:03.864316 systemd-networkd[1609]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Aug 13 07:09:03.864358 systemd-networkd[1609]: eth0: DHCP lease lost Aug 13 07:09:03.865411 waagent[1855]: 2025-08-13T07:09:03.865354Z INFO Daemon Daemon Create user account if not exists Aug 13 07:09:03.870617 waagent[1855]: 2025-08-13T07:09:03.870556Z INFO Daemon Daemon User core already exists, skip useradd Aug 13 07:09:03.876182 waagent[1855]: 2025-08-13T07:09:03.876131Z INFO Daemon Daemon Configure sudoer Aug 13 07:09:03.880564 waagent[1855]: 2025-08-13T07:09:03.880507Z INFO Daemon Daemon Configure sshd Aug 13 07:09:03.885230 waagent[1855]: 2025-08-13T07:09:03.885172Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Aug 13 07:09:03.901475 waagent[1855]: 2025-08-13T07:09:03.897251Z INFO Daemon Daemon Deploy ssh public key. Aug 13 07:09:03.911343 systemd-networkd[1609]: eth0: DHCPv4 address 10.200.20.42/24, gateway 10.200.20.1 acquired from 168.63.129.16 Aug 13 07:09:05.012300 waagent[1855]: 2025-08-13T07:09:05.011169Z INFO Daemon Daemon Provisioning complete Aug 13 07:09:05.028649 waagent[1855]: 2025-08-13T07:09:05.028604Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Aug 13 07:09:05.034399 waagent[1855]: 2025-08-13T07:09:05.034353Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Aug 13 07:09:05.044034 waagent[1855]: 2025-08-13T07:09:05.043987Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.9.1.1 is the most current agent Aug 13 07:09:05.169098 waagent[1937]: 2025-08-13T07:09:05.169028Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.9.1.1) Aug 13 07:09:05.169899 waagent[1937]: 2025-08-13T07:09:05.169530Z INFO ExtHandler ExtHandler OS: flatcar 4230.2.2 Aug 13 07:09:05.169899 waagent[1937]: 2025-08-13T07:09:05.169602Z INFO ExtHandler ExtHandler Python: 3.11.11 Aug 13 07:09:05.176318 waagent[1937]: 2025-08-13T07:09:05.175935Z INFO ExtHandler ExtHandler Distro: flatcar-4230.2.2; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.11; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Aug 13 07:09:05.176318 waagent[1937]: 2025-08-13T07:09:05.176103Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Aug 13 07:09:05.176318 waagent[1937]: 2025-08-13T07:09:05.176158Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Aug 13 07:09:05.183748 waagent[1937]: 2025-08-13T07:09:05.183696Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Aug 13 07:09:05.188634 waagent[1937]: 2025-08-13T07:09:05.188594Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.175 Aug 13 07:09:05.189069 waagent[1937]: 2025-08-13T07:09:05.189026Z INFO ExtHandler Aug 13 07:09:05.189134 waagent[1937]: 2025-08-13T07:09:05.189106Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 93d0beb8-d35d-4e57-b9e6-77ad9b613f51 eTag: 1798490211998072079 source: Fabric] Aug 13 07:09:05.189436 waagent[1937]: 2025-08-13T07:09:05.189395Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Aug 13 07:09:05.189973 waagent[1937]: 2025-08-13T07:09:05.189930Z INFO ExtHandler Aug 13 07:09:05.190036 waagent[1937]: 2025-08-13T07:09:05.190006Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Aug 13 07:09:05.194003 waagent[1937]: 2025-08-13T07:09:05.193969Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Aug 13 07:09:05.265422 waagent[1937]: 2025-08-13T07:09:05.265261Z INFO ExtHandler Downloaded certificate {'thumbprint': '9D5CD3A456091BFC2CC4D35626F9D5772C302DE5', 'hasPrivateKey': False} Aug 13 07:09:05.265798 waagent[1937]: 2025-08-13T07:09:05.265754Z INFO ExtHandler Downloaded certificate {'thumbprint': '1DA5BF181F45A40662141BA2CE4E21F780FAFED1', 'hasPrivateKey': True} Aug 13 07:09:05.266204 waagent[1937]: 2025-08-13T07:09:05.266163Z INFO ExtHandler Fetch goal state completed Aug 13 07:09:05.282405 waagent[1937]: 2025-08-13T07:09:05.282345Z INFO ExtHandler ExtHandler WALinuxAgent-2.9.1.1 running as process 1937 Aug 13 07:09:05.282555 waagent[1937]: 2025-08-13T07:09:05.282519Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Aug 13 07:09:05.284194 waagent[1937]: 2025-08-13T07:09:05.284148Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '4230.2.2', '', 'Flatcar Container Linux by Kinvolk'] Aug 13 07:09:05.284591 waagent[1937]: 2025-08-13T07:09:05.284554Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Aug 13 07:09:05.360750 waagent[1937]: 2025-08-13T07:09:05.360705Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Aug 13 07:09:05.360939 waagent[1937]: 2025-08-13T07:09:05.360901Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Aug 13 07:09:05.366199 waagent[1937]: 2025-08-13T07:09:05.366160Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Aug 13 07:09:05.372069 systemd[1]: Reload requested from client PID 1952 ('systemctl') (unit waagent.service)... Aug 13 07:09:05.372082 systemd[1]: Reloading... Aug 13 07:09:05.448299 zram_generator::config[1991]: No configuration found. Aug 13 07:09:05.554677 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:09:05.657133 systemd[1]: Reloading finished in 284 ms. Aug 13 07:09:05.670300 waagent[1937]: 2025-08-13T07:09:05.668333Z INFO ExtHandler ExtHandler Executing systemctl daemon-reload for setting up waagent-network-setup.service Aug 13 07:09:05.675837 systemd[1]: Reload requested from client PID 2047 ('systemctl') (unit waagent.service)... Aug 13 07:09:05.675849 systemd[1]: Reloading... Aug 13 07:09:05.758307 zram_generator::config[2084]: No configuration found. Aug 13 07:09:05.858974 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:09:05.955847 systemd[1]: Reloading finished in 279 ms. Aug 13 07:09:05.968349 waagent[1937]: 2025-08-13T07:09:05.967517Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Aug 13 07:09:05.968349 waagent[1937]: 2025-08-13T07:09:05.967695Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Aug 13 07:09:06.261944 waagent[1937]: 2025-08-13T07:09:06.261869Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Aug 13 07:09:06.262932 waagent[1937]: 2025-08-13T07:09:06.262856Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Aug 13 07:09:06.263913 waagent[1937]: 2025-08-13T07:09:06.263857Z INFO ExtHandler ExtHandler Starting env monitor service. Aug 13 07:09:06.264040 waagent[1937]: 2025-08-13T07:09:06.263983Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Aug 13 07:09:06.264127 waagent[1937]: 2025-08-13T07:09:06.264091Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Aug 13 07:09:06.264373 waagent[1937]: 2025-08-13T07:09:06.264325Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Aug 13 07:09:06.264899 waagent[1937]: 2025-08-13T07:09:06.264845Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Aug 13 07:09:06.265166 waagent[1937]: 2025-08-13T07:09:06.265053Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Aug 13 07:09:06.265166 waagent[1937]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Aug 13 07:09:06.265166 waagent[1937]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Aug 13 07:09:06.265166 waagent[1937]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Aug 13 07:09:06.265166 waagent[1937]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Aug 13 07:09:06.265166 waagent[1937]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Aug 13 07:09:06.265166 waagent[1937]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Aug 13 07:09:06.265369 waagent[1937]: 2025-08-13T07:09:06.265235Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Aug 13 07:09:06.265369 waagent[1937]: 2025-08-13T07:09:06.265319Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Aug 13 07:09:06.265628 waagent[1937]: 2025-08-13T07:09:06.265443Z INFO EnvHandler ExtHandler Configure routes Aug 13 07:09:06.265712 waagent[1937]: 2025-08-13T07:09:06.265673Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Aug 13 07:09:06.265865 waagent[1937]: 2025-08-13T07:09:06.265805Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Aug 13 07:09:06.266055 waagent[1937]: 2025-08-13T07:09:06.266010Z INFO EnvHandler ExtHandler Gateway:None Aug 13 07:09:06.266595 waagent[1937]: 2025-08-13T07:09:06.266525Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Aug 13 07:09:06.266731 waagent[1937]: 2025-08-13T07:09:06.266675Z INFO EnvHandler ExtHandler Routes:None Aug 13 07:09:06.267139 waagent[1937]: 2025-08-13T07:09:06.267030Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Aug 13 07:09:06.267749 waagent[1937]: 2025-08-13T07:09:06.267272Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Aug 13 07:09:06.273456 waagent[1937]: 2025-08-13T07:09:06.273405Z INFO ExtHandler ExtHandler Aug 13 07:09:06.273732 waagent[1937]: 2025-08-13T07:09:06.273691Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: e51fe2b2-c3ed-4ea6-b33e-0657e8205c80 correlation 592077da-5ac9-43ee-ade4-a226d14acba3 created: 2025-08-13T07:07:53.706749Z] Aug 13 07:09:06.274664 waagent[1937]: 2025-08-13T07:09:06.274618Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Aug 13 07:09:06.275209 waagent[1937]: 2025-08-13T07:09:06.275169Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 1 ms] Aug 13 07:09:06.311238 waagent[1937]: 2025-08-13T07:09:06.311170Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.9.1.1 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 164C07AB-FCA3-4548-B43C-0BD025A81446;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0] Aug 13 07:09:06.319112 waagent[1937]: 2025-08-13T07:09:06.319038Z INFO MonitorHandler ExtHandler Network interfaces: Aug 13 07:09:06.319112 waagent[1937]: Executing ['ip', '-a', '-o', 'link']: Aug 13 07:09:06.319112 waagent[1937]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Aug 13 07:09:06.319112 waagent[1937]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:fe:f8:e0 brd ff:ff:ff:ff:ff:ff Aug 13 07:09:06.319112 waagent[1937]: 3: enP10003s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:fe:f8:e0 brd ff:ff:ff:ff:ff:ff\ altname enP10003p0s2 Aug 13 07:09:06.319112 waagent[1937]: Executing ['ip', '-4', '-a', '-o', 'address']: Aug 13 07:09:06.319112 waagent[1937]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Aug 13 07:09:06.319112 waagent[1937]: 2: eth0 inet 10.200.20.42/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Aug 13 07:09:06.319112 waagent[1937]: Executing ['ip', '-6', '-a', '-o', 'address']: Aug 13 07:09:06.319112 waagent[1937]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Aug 13 07:09:06.319112 waagent[1937]: 2: eth0 inet6 fe80::20d:3aff:fefe:f8e0/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Aug 13 07:09:06.342688 waagent[1937]: 2025-08-13T07:09:06.342615Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Aug 13 07:09:06.342688 waagent[1937]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 07:09:06.342688 waagent[1937]: pkts bytes target prot opt in out source destination Aug 13 07:09:06.342688 waagent[1937]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Aug 13 07:09:06.342688 waagent[1937]: pkts bytes target prot opt in out source destination Aug 13 07:09:06.342688 waagent[1937]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 07:09:06.342688 waagent[1937]: pkts bytes target prot opt in out source destination Aug 13 07:09:06.342688 waagent[1937]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Aug 13 07:09:06.342688 waagent[1937]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Aug 13 07:09:06.342688 waagent[1937]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Aug 13 07:09:06.345567 waagent[1937]: 2025-08-13T07:09:06.345504Z INFO EnvHandler ExtHandler Current Firewall rules: Aug 13 07:09:06.345567 waagent[1937]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 07:09:06.345567 waagent[1937]: pkts bytes target prot opt in out source destination Aug 13 07:09:06.345567 waagent[1937]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Aug 13 07:09:06.345567 waagent[1937]: pkts bytes target prot opt in out source destination Aug 13 07:09:06.345567 waagent[1937]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Aug 13 07:09:06.345567 waagent[1937]: pkts bytes target prot opt in out source destination Aug 13 07:09:06.345567 waagent[1937]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Aug 13 07:09:06.345567 waagent[1937]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Aug 13 07:09:06.345567 waagent[1937]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Aug 13 07:09:06.345817 waagent[1937]: 2025-08-13T07:09:06.345777Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Aug 13 07:09:11.011905 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Aug 13 07:09:11.021423 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:11.122006 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:11.125698 (kubelet)[2179]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:11.261781 kubelet[2179]: E0813 07:09:11.261699 2179 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:11.264264 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:11.264425 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:11.264897 systemd[1]: kubelet.service: Consumed 120ms CPU time, 107.8M memory peak. Aug 13 07:09:21.512084 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Aug 13 07:09:21.520426 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:21.609269 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:21.612722 (kubelet)[2194]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:21.756061 kubelet[2194]: E0813 07:09:21.755995 2194 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:21.757757 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:21.757879 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:21.758400 systemd[1]: kubelet.service: Consumed 119ms CPU time, 105.9M memory peak. Aug 13 07:09:22.493170 chronyd[1690]: Selected source PHC0 Aug 13 07:09:31.762058 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Aug 13 07:09:31.770434 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:31.858367 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:31.861518 (kubelet)[2210]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:31.991497 kubelet[2210]: E0813 07:09:31.991447 2210 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:31.993875 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:31.994121 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:31.994560 systemd[1]: kubelet.service: Consumed 120ms CPU time, 107.4M memory peak. Aug 13 07:09:40.703950 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Aug 13 07:09:40.705392 systemd[1]: Started sshd@0-10.200.20.42:22-10.200.16.10:57648.service - OpenSSH per-connection server daemon (10.200.16.10:57648). Aug 13 07:09:40.851479 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Aug 13 07:09:41.294061 sshd[2218]: Accepted publickey for core from 10.200.16.10 port 57648 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:41.295237 sshd-session[2218]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:41.298939 systemd-logind[1702]: New session 3 of user core. Aug 13 07:09:41.308543 systemd[1]: Started session-3.scope - Session 3 of User core. Aug 13 07:09:41.728311 systemd[1]: Started sshd@1-10.200.20.42:22-10.200.16.10:57656.service - OpenSSH per-connection server daemon (10.200.16.10:57656). Aug 13 07:09:42.011877 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Aug 13 07:09:42.020451 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:42.110215 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:42.113062 (kubelet)[2233]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:42.226934 sshd[2223]: Accepted publickey for core from 10.200.16.10 port 57656 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:42.228122 sshd-session[2223]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:42.235339 systemd-logind[1702]: New session 4 of user core. Aug 13 07:09:42.243526 systemd[1]: Started session-4.scope - Session 4 of User core. Aug 13 07:09:42.252187 kubelet[2233]: E0813 07:09:42.252142 2233 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:42.254437 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:42.254574 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:42.254858 systemd[1]: kubelet.service: Consumed 123ms CPU time, 106M memory peak. Aug 13 07:09:42.579774 sshd[2239]: Connection closed by 10.200.16.10 port 57656 Aug 13 07:09:42.580368 sshd-session[2223]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:42.583740 systemd[1]: sshd@1-10.200.20.42:22-10.200.16.10:57656.service: Deactivated successfully. Aug 13 07:09:42.585241 systemd[1]: session-4.scope: Deactivated successfully. Aug 13 07:09:42.585885 systemd-logind[1702]: Session 4 logged out. Waiting for processes to exit. Aug 13 07:09:42.587072 systemd-logind[1702]: Removed session 4. Aug 13 07:09:42.666189 systemd[1]: Started sshd@2-10.200.20.42:22-10.200.16.10:57662.service - OpenSSH per-connection server daemon (10.200.16.10:57662). Aug 13 07:09:43.146827 sshd[2246]: Accepted publickey for core from 10.200.16.10 port 57662 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:43.148328 sshd-session[2246]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:43.152171 systemd-logind[1702]: New session 5 of user core. Aug 13 07:09:43.158401 systemd[1]: Started session-5.scope - Session 5 of User core. Aug 13 07:09:43.487512 sshd[2248]: Connection closed by 10.200.16.10 port 57662 Aug 13 07:09:43.487367 sshd-session[2246]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:43.490659 systemd[1]: sshd@2-10.200.20.42:22-10.200.16.10:57662.service: Deactivated successfully. Aug 13 07:09:43.492087 systemd[1]: session-5.scope: Deactivated successfully. Aug 13 07:09:43.493735 systemd-logind[1702]: Session 5 logged out. Waiting for processes to exit. Aug 13 07:09:43.494627 systemd-logind[1702]: Removed session 5. Aug 13 07:09:43.571464 systemd[1]: Started sshd@3-10.200.20.42:22-10.200.16.10:57676.service - OpenSSH per-connection server daemon (10.200.16.10:57676). Aug 13 07:09:44.022893 sshd[2254]: Accepted publickey for core from 10.200.16.10 port 57676 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:44.025562 sshd-session[2254]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:44.031020 systemd-logind[1702]: New session 6 of user core. Aug 13 07:09:44.037431 systemd[1]: Started session-6.scope - Session 6 of User core. Aug 13 07:09:44.223349 update_engine[1708]: I20250813 07:09:44.223289 1708 update_attempter.cc:509] Updating boot flags... Aug 13 07:09:44.294366 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 42 scanned by (udev-worker) (2272) Aug 13 07:09:44.365033 sshd[2256]: Connection closed by 10.200.16.10 port 57676 Aug 13 07:09:44.365528 sshd-session[2254]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:44.370508 systemd[1]: session-6.scope: Deactivated successfully. Aug 13 07:09:44.371186 systemd[1]: sshd@3-10.200.20.42:22-10.200.16.10:57676.service: Deactivated successfully. Aug 13 07:09:44.372826 systemd-logind[1702]: Session 6 logged out. Waiting for processes to exit. Aug 13 07:09:44.373993 systemd-logind[1702]: Removed session 6. Aug 13 07:09:44.450197 systemd[1]: Started sshd@4-10.200.20.42:22-10.200.16.10:57684.service - OpenSSH per-connection server daemon (10.200.16.10:57684). Aug 13 07:09:44.903262 sshd[2325]: Accepted publickey for core from 10.200.16.10 port 57684 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:44.904514 sshd-session[2325]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:44.909860 systemd-logind[1702]: New session 7 of user core. Aug 13 07:09:44.912404 systemd[1]: Started session-7.scope - Session 7 of User core. Aug 13 07:09:45.304180 sudo[2328]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Aug 13 07:09:45.304475 sudo[2328]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:09:45.318989 sudo[2328]: pam_unix(sudo:session): session closed for user root Aug 13 07:09:45.396229 sshd[2327]: Connection closed by 10.200.16.10 port 57684 Aug 13 07:09:45.397192 sshd-session[2325]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:45.400751 systemd-logind[1702]: Session 7 logged out. Waiting for processes to exit. Aug 13 07:09:45.401584 systemd[1]: sshd@4-10.200.20.42:22-10.200.16.10:57684.service: Deactivated successfully. Aug 13 07:09:45.403504 systemd[1]: session-7.scope: Deactivated successfully. Aug 13 07:09:45.406062 systemd-logind[1702]: Removed session 7. Aug 13 07:09:45.491482 systemd[1]: Started sshd@5-10.200.20.42:22-10.200.16.10:57690.service - OpenSSH per-connection server daemon (10.200.16.10:57690). Aug 13 07:09:45.987359 sshd[2334]: Accepted publickey for core from 10.200.16.10 port 57690 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:45.988891 sshd-session[2334]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:45.992697 systemd-logind[1702]: New session 8 of user core. Aug 13 07:09:46.003402 systemd[1]: Started session-8.scope - Session 8 of User core. Aug 13 07:09:46.265240 sudo[2338]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Aug 13 07:09:46.265518 sudo[2338]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:09:46.268343 sudo[2338]: pam_unix(sudo:session): session closed for user root Aug 13 07:09:46.272555 sudo[2337]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Aug 13 07:09:46.272789 sudo[2337]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:09:46.289544 systemd[1]: Starting audit-rules.service - Load Audit Rules... Aug 13 07:09:46.309990 augenrules[2360]: No rules Aug 13 07:09:46.311564 systemd[1]: audit-rules.service: Deactivated successfully. Aug 13 07:09:46.312352 systemd[1]: Finished audit-rules.service - Load Audit Rules. Aug 13 07:09:46.313923 sudo[2337]: pam_unix(sudo:session): session closed for user root Aug 13 07:09:46.385270 sshd[2336]: Connection closed by 10.200.16.10 port 57690 Aug 13 07:09:46.385185 sshd-session[2334]: pam_unix(sshd:session): session closed for user core Aug 13 07:09:46.387881 systemd[1]: sshd@5-10.200.20.42:22-10.200.16.10:57690.service: Deactivated successfully. Aug 13 07:09:46.389573 systemd[1]: session-8.scope: Deactivated successfully. Aug 13 07:09:46.391450 systemd-logind[1702]: Session 8 logged out. Waiting for processes to exit. Aug 13 07:09:46.392469 systemd-logind[1702]: Removed session 8. Aug 13 07:09:46.478505 systemd[1]: Started sshd@6-10.200.20.42:22-10.200.16.10:57698.service - OpenSSH per-connection server daemon (10.200.16.10:57698). Aug 13 07:09:46.929999 sshd[2369]: Accepted publickey for core from 10.200.16.10 port 57698 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:09:46.931121 sshd-session[2369]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:09:46.934836 systemd-logind[1702]: New session 9 of user core. Aug 13 07:09:46.941430 systemd[1]: Started session-9.scope - Session 9 of User core. Aug 13 07:09:47.185146 sudo[2372]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Aug 13 07:09:47.185425 sudo[2372]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Aug 13 07:09:48.126595 (dockerd)[2389]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Aug 13 07:09:48.126944 systemd[1]: Starting docker.service - Docker Application Container Engine... Aug 13 07:09:48.854305 dockerd[2389]: time="2025-08-13T07:09:48.854068204Z" level=info msg="Starting up" Aug 13 07:09:49.102189 systemd[1]: var-lib-docker-metacopy\x2dcheck2771551470-merged.mount: Deactivated successfully. Aug 13 07:09:49.123838 dockerd[2389]: time="2025-08-13T07:09:49.123755964Z" level=info msg="Loading containers: start." Aug 13 07:09:49.292299 kernel: Initializing XFRM netlink socket Aug 13 07:09:49.394172 systemd-networkd[1609]: docker0: Link UP Aug 13 07:09:49.450244 dockerd[2389]: time="2025-08-13T07:09:49.450197425Z" level=info msg="Loading containers: done." Aug 13 07:09:49.461588 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3664545607-merged.mount: Deactivated successfully. Aug 13 07:09:49.494330 dockerd[2389]: time="2025-08-13T07:09:49.493925020Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Aug 13 07:09:49.494330 dockerd[2389]: time="2025-08-13T07:09:49.494032700Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Aug 13 07:09:49.494330 dockerd[2389]: time="2025-08-13T07:09:49.494148340Z" level=info msg="Daemon has completed initialization" Aug 13 07:09:49.565600 dockerd[2389]: time="2025-08-13T07:09:49.565522785Z" level=info msg="API listen on /run/docker.sock" Aug 13 07:09:49.565683 systemd[1]: Started docker.service - Docker Application Container Engine. Aug 13 07:09:50.617923 containerd[1741]: time="2025-08-13T07:09:50.617879693Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\"" Aug 13 07:09:51.565309 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2904394381.mount: Deactivated successfully. Aug 13 07:09:52.261884 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Aug 13 07:09:52.268448 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:09:52.363309 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:09:52.366190 (kubelet)[2612]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:09:52.494399 kubelet[2612]: E0813 07:09:52.494309 2612 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:09:52.496390 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:09:52.496535 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:09:52.496988 systemd[1]: kubelet.service: Consumed 121ms CPU time, 109.7M memory peak. Aug 13 07:09:53.463330 containerd[1741]: time="2025-08-13T07:09:53.462546020Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:53.466313 containerd[1741]: time="2025-08-13T07:09:53.466254736Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.31.11: active requests=0, bytes read=25651813" Aug 13 07:09:53.470099 containerd[1741]: time="2025-08-13T07:09:53.470059612Z" level=info msg="ImageCreate event name:\"sha256:00a68b619a4bfa14c989a2181a7aa0726a5cb1272a7f65394e6a594ad6eade27\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:53.474693 containerd[1741]: time="2025-08-13T07:09:53.474642487Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a3d1c4440817725a1b503a7ccce94f3dce2b208ebf257b405dc2d97817df3dde\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:53.475775 containerd[1741]: time="2025-08-13T07:09:53.475618326Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.31.11\" with image id \"sha256:00a68b619a4bfa14c989a2181a7aa0726a5cb1272a7f65394e6a594ad6eade27\", repo tag \"registry.k8s.io/kube-apiserver:v1.31.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a3d1c4440817725a1b503a7ccce94f3dce2b208ebf257b405dc2d97817df3dde\", size \"25648613\" in 2.857694753s" Aug 13 07:09:53.475775 containerd[1741]: time="2025-08-13T07:09:53.475651966Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.11\" returns image reference \"sha256:00a68b619a4bfa14c989a2181a7aa0726a5cb1272a7f65394e6a594ad6eade27\"" Aug 13 07:09:53.477496 containerd[1741]: time="2025-08-13T07:09:53.477374205Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\"" Aug 13 07:09:54.837316 containerd[1741]: time="2025-08-13T07:09:54.836780665Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:54.840994 containerd[1741]: time="2025-08-13T07:09:54.840784103Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.31.11: active requests=0, bytes read=22460283" Aug 13 07:09:54.845352 containerd[1741]: time="2025-08-13T07:09:54.845322862Z" level=info msg="ImageCreate event name:\"sha256:5c5dc52b837451e0fe6108fdfb9cfa431191ce227ce71d103dec8a8c655c4e71\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:54.850893 containerd[1741]: time="2025-08-13T07:09:54.850865701Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:0f19de157f3d251f5ddeb6e9d026895bc55cb02592874b326fa345c57e5e2848\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:54.852010 containerd[1741]: time="2025-08-13T07:09:54.851889500Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.31.11\" with image id \"sha256:5c5dc52b837451e0fe6108fdfb9cfa431191ce227ce71d103dec8a8c655c4e71\", repo tag \"registry.k8s.io/kube-controller-manager:v1.31.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:0f19de157f3d251f5ddeb6e9d026895bc55cb02592874b326fa345c57e5e2848\", size \"23996073\" in 1.374458096s" Aug 13 07:09:54.852010 containerd[1741]: time="2025-08-13T07:09:54.851920140Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.11\" returns image reference \"sha256:5c5dc52b837451e0fe6108fdfb9cfa431191ce227ce71d103dec8a8c655c4e71\"" Aug 13 07:09:54.852503 containerd[1741]: time="2025-08-13T07:09:54.852472820Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\"" Aug 13 07:09:56.055010 containerd[1741]: time="2025-08-13T07:09:56.054952816Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:56.057692 containerd[1741]: time="2025-08-13T07:09:56.057611056Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.31.11: active requests=0, bytes read=17125089" Aug 13 07:09:56.065034 containerd[1741]: time="2025-08-13T07:09:56.064992294Z" level=info msg="ImageCreate event name:\"sha256:89be0efdc4ab1793b9b1b05e836e33dc50f5b2911b57609b315b58608b2d3746\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:56.073800 containerd[1741]: time="2025-08-13T07:09:56.073761331Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:1a9b59b3bfa6c1f1911f6f865a795620c461d079e413061bb71981cadd67f39d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:56.075077 containerd[1741]: time="2025-08-13T07:09:56.074758371Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.31.11\" with image id \"sha256:89be0efdc4ab1793b9b1b05e836e33dc50f5b2911b57609b315b58608b2d3746\", repo tag \"registry.k8s.io/kube-scheduler:v1.31.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:1a9b59b3bfa6c1f1911f6f865a795620c461d079e413061bb71981cadd67f39d\", size \"18660897\" in 1.222248591s" Aug 13 07:09:56.075077 containerd[1741]: time="2025-08-13T07:09:56.074790571Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.11\" returns image reference \"sha256:89be0efdc4ab1793b9b1b05e836e33dc50f5b2911b57609b315b58608b2d3746\"" Aug 13 07:09:56.075251 containerd[1741]: time="2025-08-13T07:09:56.075218691Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\"" Aug 13 07:09:57.298781 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3631708866.mount: Deactivated successfully. Aug 13 07:09:57.635104 containerd[1741]: time="2025-08-13T07:09:57.634985271Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.31.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:57.638599 containerd[1741]: time="2025-08-13T07:09:57.638454950Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.31.11: active requests=0, bytes read=26915993" Aug 13 07:09:57.642688 containerd[1741]: time="2025-08-13T07:09:57.642639389Z" level=info msg="ImageCreate event name:\"sha256:7d1e7db6660181423f98acbe3a495b3fe5cec9b85cdef245540cc2cb3b180ab0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:57.647442 containerd[1741]: time="2025-08-13T07:09:57.647268188Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:a31da847792c5e7e92e91b78da1ad21d693e4b2b48d0e9f4610c8764dc2a5d79\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:57.648014 containerd[1741]: time="2025-08-13T07:09:57.647985507Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.31.11\" with image id \"sha256:7d1e7db6660181423f98acbe3a495b3fe5cec9b85cdef245540cc2cb3b180ab0\", repo tag \"registry.k8s.io/kube-proxy:v1.31.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:a31da847792c5e7e92e91b78da1ad21d693e4b2b48d0e9f4610c8764dc2a5d79\", size \"26915012\" in 1.572732656s" Aug 13 07:09:57.648085 containerd[1741]: time="2025-08-13T07:09:57.648017907Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.11\" returns image reference \"sha256:7d1e7db6660181423f98acbe3a495b3fe5cec9b85cdef245540cc2cb3b180ab0\"" Aug 13 07:09:57.648734 containerd[1741]: time="2025-08-13T07:09:57.648709707Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Aug 13 07:09:58.411468 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount302402872.mount: Deactivated successfully. Aug 13 07:09:59.781062 containerd[1741]: time="2025-08-13T07:09:59.781008173Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:59.788201 containerd[1741]: time="2025-08-13T07:09:59.788130531Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951622" Aug 13 07:09:59.793366 containerd[1741]: time="2025-08-13T07:09:59.793317970Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:59.798798 containerd[1741]: time="2025-08-13T07:09:59.798743968Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:09:59.800189 containerd[1741]: time="2025-08-13T07:09:59.799888528Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 2.151146501s" Aug 13 07:09:59.800189 containerd[1741]: time="2025-08-13T07:09:59.799921528Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Aug 13 07:09:59.800678 containerd[1741]: time="2025-08-13T07:09:59.800657568Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Aug 13 07:10:00.497472 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3234021819.mount: Deactivated successfully. Aug 13 07:10:00.528312 containerd[1741]: time="2025-08-13T07:10:00.528256341Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:00.531304 containerd[1741]: time="2025-08-13T07:10:00.531258978Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Aug 13 07:10:00.537696 containerd[1741]: time="2025-08-13T07:10:00.537649172Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:00.553602 containerd[1741]: time="2025-08-13T07:10:00.553514236Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:00.555170 containerd[1741]: time="2025-08-13T07:10:00.554976995Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 754.228147ms" Aug 13 07:10:00.555170 containerd[1741]: time="2025-08-13T07:10:00.555029635Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Aug 13 07:10:00.556063 containerd[1741]: time="2025-08-13T07:10:00.555766634Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Aug 13 07:10:01.869203 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount729057155.mount: Deactivated successfully. Aug 13 07:10:02.511986 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Aug 13 07:10:02.518557 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:10:02.618595 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:02.621457 (kubelet)[2743]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Aug 13 07:10:02.741383 kubelet[2743]: E0813 07:10:02.741232 2743 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Aug 13 07:10:02.744078 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Aug 13 07:10:02.744227 systemd[1]: kubelet.service: Failed with result 'exit-code'. Aug 13 07:10:02.744668 systemd[1]: kubelet.service: Consumed 119ms CPU time, 106.9M memory peak. Aug 13 07:10:04.549152 containerd[1741]: time="2025-08-13T07:10:04.549093077Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.15-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:04.555460 containerd[1741]: time="2025-08-13T07:10:04.555400912Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.15-0: active requests=0, bytes read=66406465" Aug 13 07:10:04.559350 containerd[1741]: time="2025-08-13T07:10:04.559271708Z" level=info msg="ImageCreate event name:\"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:04.589444 containerd[1741]: time="2025-08-13T07:10:04.589362320Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:04.591790 containerd[1741]: time="2025-08-13T07:10:04.591210718Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.15-0\" with image id \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\", repo tag \"registry.k8s.io/etcd:3.5.15-0\", repo digest \"registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a\", size \"66535646\" in 4.035414164s" Aug 13 07:10:04.591790 containerd[1741]: time="2025-08-13T07:10:04.591240318Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:27e3830e1402783674d8b594038967deea9d51f0d91b34c93c8f39d2f68af7da\"" Aug 13 07:10:11.587720 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:11.587858 systemd[1]: kubelet.service: Consumed 119ms CPU time, 106.9M memory peak. Aug 13 07:10:11.592476 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:10:11.618053 systemd[1]: Reload requested from client PID 2812 ('systemctl') (unit session-9.scope)... Aug 13 07:10:11.618065 systemd[1]: Reloading... Aug 13 07:10:11.729339 zram_generator::config[2865]: No configuration found. Aug 13 07:10:11.819632 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:10:11.920089 systemd[1]: Reloading finished in 301 ms. Aug 13 07:10:11.970679 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:11.976029 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:10:11.980752 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 07:10:11.980952 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:11.980993 systemd[1]: kubelet.service: Consumed 84ms CPU time, 94.9M memory peak. Aug 13 07:10:11.982491 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:10:12.120038 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:12.129527 (kubelet)[2928]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:10:12.160522 kubelet[2928]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:10:12.160522 kubelet[2928]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 07:10:12.160522 kubelet[2928]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:10:12.160848 kubelet[2928]: I0813 07:10:12.160571 2928 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:10:13.403914 kubelet[2928]: I0813 07:10:13.403550 2928 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 07:10:13.403914 kubelet[2928]: I0813 07:10:13.403580 2928 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:10:13.404300 kubelet[2928]: I0813 07:10:13.404131 2928 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 07:10:13.422020 kubelet[2928]: E0813 07:10:13.421975 2928 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.20.42:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:13.423194 kubelet[2928]: I0813 07:10:13.423179 2928 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:10:13.428511 kubelet[2928]: E0813 07:10:13.428433 2928 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:10:13.428511 kubelet[2928]: I0813 07:10:13.428462 2928 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:10:13.432263 kubelet[2928]: I0813 07:10:13.432238 2928 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:10:13.432910 kubelet[2928]: I0813 07:10:13.432889 2928 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 07:10:13.433045 kubelet[2928]: I0813 07:10:13.433017 2928 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:10:13.433194 kubelet[2928]: I0813 07:10:13.433044 2928 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230.2.2-a-ad5fc6e7e6","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 07:10:13.433288 kubelet[2928]: I0813 07:10:13.433200 2928 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:10:13.433288 kubelet[2928]: I0813 07:10:13.433209 2928 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 07:10:13.433347 kubelet[2928]: I0813 07:10:13.433330 2928 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:10:13.435224 kubelet[2928]: I0813 07:10:13.435042 2928 kubelet.go:408] "Attempting to sync node with API server" Aug 13 07:10:13.435224 kubelet[2928]: I0813 07:10:13.435067 2928 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:10:13.435224 kubelet[2928]: I0813 07:10:13.435088 2928 kubelet.go:314] "Adding apiserver pod source" Aug 13 07:10:13.435224 kubelet[2928]: I0813 07:10:13.435101 2928 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:10:13.440042 kubelet[2928]: W0813 07:10:13.439525 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.42:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-a-ad5fc6e7e6&limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:13.440042 kubelet[2928]: E0813 07:10:13.439591 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.20.42:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-a-ad5fc6e7e6&limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:13.440042 kubelet[2928]: W0813 07:10:13.439894 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.42:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:13.440042 kubelet[2928]: E0813 07:10:13.439926 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.20.42:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:13.440292 kubelet[2928]: I0813 07:10:13.440256 2928 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Aug 13 07:10:13.441132 kubelet[2928]: I0813 07:10:13.440732 2928 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 07:10:13.441132 kubelet[2928]: W0813 07:10:13.440780 2928 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Aug 13 07:10:13.442008 kubelet[2928]: I0813 07:10:13.441981 2928 server.go:1274] "Started kubelet" Aug 13 07:10:13.443498 kubelet[2928]: I0813 07:10:13.443471 2928 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:10:13.444333 kubelet[2928]: I0813 07:10:13.444318 2928 server.go:449] "Adding debug handlers to kubelet server" Aug 13 07:10:13.444568 kubelet[2928]: I0813 07:10:13.444505 2928 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:10:13.445140 kubelet[2928]: I0813 07:10:13.445109 2928 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:10:13.446778 kubelet[2928]: I0813 07:10:13.446746 2928 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:10:13.448368 kubelet[2928]: E0813 07:10:13.446712 2928 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.42:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.42:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4230.2.2-a-ad5fc6e7e6.185b41f80e80e473 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230.2.2-a-ad5fc6e7e6,UID:ci-4230.2.2-a-ad5fc6e7e6,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230.2.2-a-ad5fc6e7e6,},FirstTimestamp:2025-08-13 07:10:13.441963123 +0000 UTC m=+1.309750807,LastTimestamp:2025-08-13 07:10:13.441963123 +0000 UTC m=+1.309750807,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230.2.2-a-ad5fc6e7e6,}" Aug 13 07:10:13.449726 kubelet[2928]: I0813 07:10:13.449703 2928 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:10:13.451931 kubelet[2928]: I0813 07:10:13.451915 2928 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 07:10:13.453336 kubelet[2928]: E0813 07:10:13.453316 2928 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230.2.2-a-ad5fc6e7e6\" not found" Aug 13 07:10:13.454271 kubelet[2928]: I0813 07:10:13.454256 2928 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 07:10:13.454928 kubelet[2928]: E0813 07:10:13.454639 2928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.42:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-a-ad5fc6e7e6?timeout=10s\": dial tcp 10.200.20.42:6443: connect: connection refused" interval="200ms" Aug 13 07:10:13.455631 kubelet[2928]: I0813 07:10:13.454722 2928 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:10:13.455730 kubelet[2928]: I0813 07:10:13.455151 2928 factory.go:221] Registration of the systemd container factory successfully Aug 13 07:10:13.455862 kubelet[2928]: I0813 07:10:13.455845 2928 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:10:13.457944 kubelet[2928]: W0813 07:10:13.457481 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.42:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:13.457944 kubelet[2928]: E0813 07:10:13.457524 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.20.42:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:13.458774 kubelet[2928]: I0813 07:10:13.458757 2928 factory.go:221] Registration of the containerd container factory successfully Aug 13 07:10:13.462147 kubelet[2928]: E0813 07:10:13.462006 2928 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:10:13.472844 kubelet[2928]: I0813 07:10:13.472727 2928 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 07:10:13.473650 kubelet[2928]: I0813 07:10:13.473633 2928 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 07:10:13.473959 kubelet[2928]: I0813 07:10:13.473721 2928 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 07:10:13.473959 kubelet[2928]: I0813 07:10:13.473742 2928 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 07:10:13.473959 kubelet[2928]: E0813 07:10:13.473776 2928 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:10:13.478813 kubelet[2928]: W0813 07:10:13.478785 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.42:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:13.478974 kubelet[2928]: E0813 07:10:13.478955 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.20.42:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:13.479674 kubelet[2928]: I0813 07:10:13.479652 2928 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 07:10:13.479674 kubelet[2928]: I0813 07:10:13.479669 2928 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 07:10:13.479768 kubelet[2928]: I0813 07:10:13.479687 2928 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:10:13.486199 kubelet[2928]: I0813 07:10:13.486175 2928 policy_none.go:49] "None policy: Start" Aug 13 07:10:13.486898 kubelet[2928]: I0813 07:10:13.486881 2928 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 07:10:13.487233 kubelet[2928]: I0813 07:10:13.487004 2928 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:10:13.495871 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Aug 13 07:10:13.508679 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Aug 13 07:10:13.511634 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Aug 13 07:10:13.526062 kubelet[2928]: I0813 07:10:13.526038 2928 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 07:10:13.527101 kubelet[2928]: I0813 07:10:13.526667 2928 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:10:13.527101 kubelet[2928]: I0813 07:10:13.526689 2928 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:10:13.527101 kubelet[2928]: I0813 07:10:13.526945 2928 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:10:13.529128 kubelet[2928]: E0813 07:10:13.529113 2928 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4230.2.2-a-ad5fc6e7e6\" not found" Aug 13 07:10:13.584083 systemd[1]: Created slice kubepods-burstable-poda1e86d3a89911a8cd45e2eee93c504f8.slice - libcontainer container kubepods-burstable-poda1e86d3a89911a8cd45e2eee93c504f8.slice. Aug 13 07:10:13.606853 systemd[1]: Created slice kubepods-burstable-pod19973cd5fe66cbaf3cc4ec7be1d3c8a2.slice - libcontainer container kubepods-burstable-pod19973cd5fe66cbaf3cc4ec7be1d3c8a2.slice. Aug 13 07:10:13.621247 systemd[1]: Created slice kubepods-burstable-podcfbc4e22bdbd14f9d0eca7156b11f436.slice - libcontainer container kubepods-burstable-podcfbc4e22bdbd14f9d0eca7156b11f436.slice. Aug 13 07:10:13.628411 kubelet[2928]: I0813 07:10:13.628375 2928 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.628783 kubelet[2928]: E0813 07:10:13.628740 2928 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.42:6443/api/v1/nodes\": dial tcp 10.200.20.42:6443: connect: connection refused" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.656564 kubelet[2928]: E0813 07:10:13.656456 2928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.42:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-a-ad5fc6e7e6?timeout=10s\": dial tcp 10.200.20.42:6443: connect: connection refused" interval="400ms" Aug 13 07:10:13.756344 kubelet[2928]: I0813 07:10:13.756231 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-flexvolume-dir\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756344 kubelet[2928]: I0813 07:10:13.756291 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-kubeconfig\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756344 kubelet[2928]: I0813 07:10:13.756310 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a1e86d3a89911a8cd45e2eee93c504f8-ca-certs\") pod \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"a1e86d3a89911a8cd45e2eee93c504f8\") " pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756344 kubelet[2928]: I0813 07:10:13.756333 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a1e86d3a89911a8cd45e2eee93c504f8-k8s-certs\") pod \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"a1e86d3a89911a8cd45e2eee93c504f8\") " pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756547 kubelet[2928]: I0813 07:10:13.756362 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a1e86d3a89911a8cd45e2eee93c504f8-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"a1e86d3a89911a8cd45e2eee93c504f8\") " pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756547 kubelet[2928]: I0813 07:10:13.756378 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-ca-certs\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756547 kubelet[2928]: I0813 07:10:13.756395 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-k8s-certs\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756547 kubelet[2928]: I0813 07:10:13.756412 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.756547 kubelet[2928]: I0813 07:10:13.756429 2928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/cfbc4e22bdbd14f9d0eca7156b11f436-kubeconfig\") pod \"kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"cfbc4e22bdbd14f9d0eca7156b11f436\") " pod="kube-system/kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.830867 kubelet[2928]: I0813 07:10:13.830838 2928 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.831182 kubelet[2928]: E0813 07:10:13.831149 2928 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.42:6443/api/v1/nodes\": dial tcp 10.200.20.42:6443: connect: connection refused" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:13.905691 containerd[1741]: time="2025-08-13T07:10:13.905649380Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6,Uid:a1e86d3a89911a8cd45e2eee93c504f8,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:13.920383 containerd[1741]: time="2025-08-13T07:10:13.920156407Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6,Uid:19973cd5fe66cbaf3cc4ec7be1d3c8a2,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:13.923899 containerd[1741]: time="2025-08-13T07:10:13.923828084Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6,Uid:cfbc4e22bdbd14f9d0eca7156b11f436,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:14.057322 kubelet[2928]: E0813 07:10:14.057231 2928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.42:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-a-ad5fc6e7e6?timeout=10s\": dial tcp 10.200.20.42:6443: connect: connection refused" interval="800ms" Aug 13 07:10:14.233471 kubelet[2928]: I0813 07:10:14.233383 2928 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:14.233738 kubelet[2928]: E0813 07:10:14.233676 2928 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.42:6443/api/v1/nodes\": dial tcp 10.200.20.42:6443: connect: connection refused" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:14.300826 kubelet[2928]: W0813 07:10:14.300734 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.42:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:14.300959 kubelet[2928]: E0813 07:10:14.300842 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.200.20.42:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:14.350774 kubelet[2928]: W0813 07:10:14.350696 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.42:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-a-ad5fc6e7e6&limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:14.350881 kubelet[2928]: E0813 07:10:14.350789 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.200.20.42:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4230.2.2-a-ad5fc6e7e6&limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:14.645823 kubelet[2928]: W0813 07:10:14.645779 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.42:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:14.646149 kubelet[2928]: E0813 07:10:14.645828 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.200.20.42:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:14.675346 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4128231045.mount: Deactivated successfully. Aug 13 07:10:14.715516 containerd[1741]: time="2025-08-13T07:10:14.715463642Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:10:14.733987 containerd[1741]: time="2025-08-13T07:10:14.733929465Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Aug 13 07:10:14.739126 containerd[1741]: time="2025-08-13T07:10:14.739086741Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:10:14.747375 containerd[1741]: time="2025-08-13T07:10:14.746394054Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:10:14.756751 containerd[1741]: time="2025-08-13T07:10:14.756503445Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:10:14.762301 containerd[1741]: time="2025-08-13T07:10:14.762262320Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:10:14.765772 containerd[1741]: time="2025-08-13T07:10:14.765734556Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Aug 13 07:10:14.766653 containerd[1741]: time="2025-08-13T07:10:14.766624236Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 860.894616ms" Aug 13 07:10:14.769449 containerd[1741]: time="2025-08-13T07:10:14.769408753Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Aug 13 07:10:14.775453 containerd[1741]: time="2025-08-13T07:10:14.775316628Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 855.092621ms" Aug 13 07:10:14.818219 containerd[1741]: time="2025-08-13T07:10:14.818167349Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 894.280945ms" Aug 13 07:10:14.858373 kubelet[2928]: E0813 07:10:14.858327 2928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.42:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4230.2.2-a-ad5fc6e7e6?timeout=10s\": dial tcp 10.200.20.42:6443: connect: connection refused" interval="1.6s" Aug 13 07:10:15.036028 kubelet[2928]: I0813 07:10:15.035986 2928 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:15.036326 kubelet[2928]: E0813 07:10:15.036299 2928 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.200.20.42:6443/api/v1/nodes\": dial tcp 10.200.20.42:6443: connect: connection refused" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:15.080575 kubelet[2928]: W0813 07:10:15.080486 2928 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.42:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.42:6443: connect: connection refused Aug 13 07:10:15.080575 kubelet[2928]: E0813 07:10:15.080531 2928 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.200.20.42:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:15.344284 containerd[1741]: time="2025-08-13T07:10:15.344048029Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:15.344284 containerd[1741]: time="2025-08-13T07:10:15.344132549Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:15.344284 containerd[1741]: time="2025-08-13T07:10:15.344158509Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:15.346246 containerd[1741]: time="2025-08-13T07:10:15.345367228Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:15.346246 containerd[1741]: time="2025-08-13T07:10:15.345872868Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:15.346246 containerd[1741]: time="2025-08-13T07:10:15.345884948Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:15.346246 containerd[1741]: time="2025-08-13T07:10:15.345951667Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:15.347623 containerd[1741]: time="2025-08-13T07:10:15.346576187Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:15.349230 containerd[1741]: time="2025-08-13T07:10:15.349004225Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:15.349230 containerd[1741]: time="2025-08-13T07:10:15.349058065Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:15.349230 containerd[1741]: time="2025-08-13T07:10:15.349073665Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:15.349230 containerd[1741]: time="2025-08-13T07:10:15.349138385Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:15.368599 systemd[1]: Started cri-containerd-5bc4cdcb38ab7bad478ccbb5274e0398f9e82e4dcf0aed9106f93dc263673081.scope - libcontainer container 5bc4cdcb38ab7bad478ccbb5274e0398f9e82e4dcf0aed9106f93dc263673081. Aug 13 07:10:15.373520 systemd[1]: Started cri-containerd-a59b2d96e8aa08bd66db50b1d13a2ec42a4d081940cd41183b811653717f6496.scope - libcontainer container a59b2d96e8aa08bd66db50b1d13a2ec42a4d081940cd41183b811653717f6496. Aug 13 07:10:15.375232 systemd[1]: Started cri-containerd-e2043e815bac74a47d843b086c5d0d8c83aa613ca47319783a88adaed5d4035c.scope - libcontainer container e2043e815bac74a47d843b086c5d0d8c83aa613ca47319783a88adaed5d4035c. Aug 13 07:10:15.417052 containerd[1741]: time="2025-08-13T07:10:15.416645363Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6,Uid:a1e86d3a89911a8cd45e2eee93c504f8,Namespace:kube-system,Attempt:0,} returns sandbox id \"5bc4cdcb38ab7bad478ccbb5274e0398f9e82e4dcf0aed9106f93dc263673081\"" Aug 13 07:10:15.424052 containerd[1741]: time="2025-08-13T07:10:15.423906516Z" level=info msg="CreateContainer within sandbox \"5bc4cdcb38ab7bad478ccbb5274e0398f9e82e4dcf0aed9106f93dc263673081\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Aug 13 07:10:15.430395 kubelet[2928]: E0813 07:10:15.430356 2928 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.200.20.42:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.42:6443: connect: connection refused" logger="UnhandledError" Aug 13 07:10:15.432120 containerd[1741]: time="2025-08-13T07:10:15.431826269Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6,Uid:19973cd5fe66cbaf3cc4ec7be1d3c8a2,Namespace:kube-system,Attempt:0,} returns sandbox id \"a59b2d96e8aa08bd66db50b1d13a2ec42a4d081940cd41183b811653717f6496\"" Aug 13 07:10:15.435532 containerd[1741]: time="2025-08-13T07:10:15.435436226Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6,Uid:cfbc4e22bdbd14f9d0eca7156b11f436,Namespace:kube-system,Attempt:0,} returns sandbox id \"e2043e815bac74a47d843b086c5d0d8c83aa613ca47319783a88adaed5d4035c\"" Aug 13 07:10:15.436560 containerd[1741]: time="2025-08-13T07:10:15.436536105Z" level=info msg="CreateContainer within sandbox \"a59b2d96e8aa08bd66db50b1d13a2ec42a4d081940cd41183b811653717f6496\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Aug 13 07:10:15.438819 containerd[1741]: time="2025-08-13T07:10:15.438782463Z" level=info msg="CreateContainer within sandbox \"e2043e815bac74a47d843b086c5d0d8c83aa613ca47319783a88adaed5d4035c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Aug 13 07:10:15.562044 containerd[1741]: time="2025-08-13T07:10:15.562001591Z" level=info msg="CreateContainer within sandbox \"5bc4cdcb38ab7bad478ccbb5274e0398f9e82e4dcf0aed9106f93dc263673081\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b7ced0d563b2c5a573e1d9ab2de17be3e9f611c7e94c50c250b445ee721f662d\"" Aug 13 07:10:15.563432 containerd[1741]: time="2025-08-13T07:10:15.562849630Z" level=info msg="StartContainer for \"b7ced0d563b2c5a573e1d9ab2de17be3e9f611c7e94c50c250b445ee721f662d\"" Aug 13 07:10:15.583424 systemd[1]: Started cri-containerd-b7ced0d563b2c5a573e1d9ab2de17be3e9f611c7e94c50c250b445ee721f662d.scope - libcontainer container b7ced0d563b2c5a573e1d9ab2de17be3e9f611c7e94c50c250b445ee721f662d. Aug 13 07:10:15.591013 containerd[1741]: time="2025-08-13T07:10:15.590783844Z" level=info msg="CreateContainer within sandbox \"a59b2d96e8aa08bd66db50b1d13a2ec42a4d081940cd41183b811653717f6496\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"d56d18975ccffe9aebb757dab5e6939a29a5d3c0f68bed6b12874885f1de5c08\"" Aug 13 07:10:15.591414 containerd[1741]: time="2025-08-13T07:10:15.591385204Z" level=info msg="StartContainer for \"d56d18975ccffe9aebb757dab5e6939a29a5d3c0f68bed6b12874885f1de5c08\"" Aug 13 07:10:15.597816 containerd[1741]: time="2025-08-13T07:10:15.597634518Z" level=info msg="CreateContainer within sandbox \"e2043e815bac74a47d843b086c5d0d8c83aa613ca47319783a88adaed5d4035c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"f54be8455768e9880d708dcfa080c559aaeb661e38a644fd13f109cbc8b59c0a\"" Aug 13 07:10:15.598936 containerd[1741]: time="2025-08-13T07:10:15.598914157Z" level=info msg="StartContainer for \"f54be8455768e9880d708dcfa080c559aaeb661e38a644fd13f109cbc8b59c0a\"" Aug 13 07:10:15.621461 systemd[1]: Started cri-containerd-d56d18975ccffe9aebb757dab5e6939a29a5d3c0f68bed6b12874885f1de5c08.scope - libcontainer container d56d18975ccffe9aebb757dab5e6939a29a5d3c0f68bed6b12874885f1de5c08. Aug 13 07:10:15.649214 containerd[1741]: time="2025-08-13T07:10:15.648961511Z" level=info msg="StartContainer for \"b7ced0d563b2c5a573e1d9ab2de17be3e9f611c7e94c50c250b445ee721f662d\" returns successfully" Aug 13 07:10:15.655126 systemd[1]: Started cri-containerd-f54be8455768e9880d708dcfa080c559aaeb661e38a644fd13f109cbc8b59c0a.scope - libcontainer container f54be8455768e9880d708dcfa080c559aaeb661e38a644fd13f109cbc8b59c0a. Aug 13 07:10:15.688009 containerd[1741]: time="2025-08-13T07:10:15.687963916Z" level=info msg="StartContainer for \"d56d18975ccffe9aebb757dab5e6939a29a5d3c0f68bed6b12874885f1de5c08\" returns successfully" Aug 13 07:10:15.739497 containerd[1741]: time="2025-08-13T07:10:15.739451269Z" level=info msg="StartContainer for \"f54be8455768e9880d708dcfa080c559aaeb661e38a644fd13f109cbc8b59c0a\" returns successfully" Aug 13 07:10:16.638246 kubelet[2928]: I0813 07:10:16.638207 2928 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:17.663394 kubelet[2928]: E0813 07:10:17.663357 2928 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4230.2.2-a-ad5fc6e7e6\" not found" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:17.759727 kubelet[2928]: I0813 07:10:17.759687 2928 kubelet_node_status.go:75] "Successfully registered node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:17.759727 kubelet[2928]: E0813 07:10:17.759725 2928 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-4230.2.2-a-ad5fc6e7e6\": node \"ci-4230.2.2-a-ad5fc6e7e6\" not found" Aug 13 07:10:17.821370 kubelet[2928]: E0813 07:10:17.821260 2928 event.go:359] "Server rejected event (will not retry!)" err="namespaces \"default\" not found" event="&Event{ObjectMeta:{ci-4230.2.2-a-ad5fc6e7e6.185b41f80e80e473 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4230.2.2-a-ad5fc6e7e6,UID:ci-4230.2.2-a-ad5fc6e7e6,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4230.2.2-a-ad5fc6e7e6,},FirstTimestamp:2025-08-13 07:10:13.441963123 +0000 UTC m=+1.309750807,LastTimestamp:2025-08-13 07:10:13.441963123 +0000 UTC m=+1.309750807,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4230.2.2-a-ad5fc6e7e6,}" Aug 13 07:10:18.441695 kubelet[2928]: I0813 07:10:18.441658 2928 apiserver.go:52] "Watching apiserver" Aug 13 07:10:18.455119 kubelet[2928]: I0813 07:10:18.455058 2928 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 07:10:18.514805 kubelet[2928]: W0813 07:10:18.514751 2928 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:10:18.515912 kubelet[2928]: W0813 07:10:18.515750 2928 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:10:20.160743 systemd[1]: Reload requested from client PID 3202 ('systemctl') (unit session-9.scope)... Aug 13 07:10:20.160763 systemd[1]: Reloading... Aug 13 07:10:20.314347 zram_generator::config[3261]: No configuration found. Aug 13 07:10:20.422551 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Aug 13 07:10:20.537850 systemd[1]: Reloading finished in 376 ms. Aug 13 07:10:20.561126 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:10:20.581594 systemd[1]: kubelet.service: Deactivated successfully. Aug 13 07:10:20.581816 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:20.581858 systemd[1]: kubelet.service: Consumed 1.660s CPU time, 129M memory peak. Aug 13 07:10:20.589515 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Aug 13 07:10:20.868273 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Aug 13 07:10:20.878071 (kubelet)[3313]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Aug 13 07:10:20.916810 kubelet[3313]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:10:20.917398 kubelet[3313]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Aug 13 07:10:20.917398 kubelet[3313]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Aug 13 07:10:20.918318 kubelet[3313]: I0813 07:10:20.917748 3313 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Aug 13 07:10:20.924305 kubelet[3313]: I0813 07:10:20.924240 3313 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Aug 13 07:10:20.924305 kubelet[3313]: I0813 07:10:20.924268 3313 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Aug 13 07:10:20.924529 kubelet[3313]: I0813 07:10:20.924508 3313 server.go:934] "Client rotation is on, will bootstrap in background" Aug 13 07:10:20.925899 kubelet[3313]: I0813 07:10:20.925868 3313 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Aug 13 07:10:20.932241 kubelet[3313]: I0813 07:10:20.931641 3313 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Aug 13 07:10:20.936203 kubelet[3313]: E0813 07:10:20.935949 3313 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Aug 13 07:10:20.936203 kubelet[3313]: I0813 07:10:20.936077 3313 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Aug 13 07:10:20.940536 kubelet[3313]: I0813 07:10:20.939760 3313 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Aug 13 07:10:20.940536 kubelet[3313]: I0813 07:10:20.939896 3313 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Aug 13 07:10:20.940536 kubelet[3313]: I0813 07:10:20.939989 3313 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Aug 13 07:10:20.940987 kubelet[3313]: I0813 07:10:20.940008 3313 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4230.2.2-a-ad5fc6e7e6","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Aug 13 07:10:20.942170 kubelet[3313]: I0813 07:10:20.942148 3313 topology_manager.go:138] "Creating topology manager with none policy" Aug 13 07:10:20.942273 kubelet[3313]: I0813 07:10:20.942263 3313 container_manager_linux.go:300] "Creating device plugin manager" Aug 13 07:10:20.942418 kubelet[3313]: I0813 07:10:20.942407 3313 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:10:20.942582 kubelet[3313]: I0813 07:10:20.942571 3313 kubelet.go:408] "Attempting to sync node with API server" Aug 13 07:10:20.942648 kubelet[3313]: I0813 07:10:20.942639 3313 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Aug 13 07:10:20.946557 kubelet[3313]: I0813 07:10:20.945071 3313 kubelet.go:314] "Adding apiserver pod source" Aug 13 07:10:20.946557 kubelet[3313]: I0813 07:10:20.945094 3313 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Aug 13 07:10:20.947822 kubelet[3313]: I0813 07:10:20.947716 3313 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Aug 13 07:10:20.951291 kubelet[3313]: I0813 07:10:20.948178 3313 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Aug 13 07:10:20.951291 kubelet[3313]: I0813 07:10:20.948732 3313 server.go:1274] "Started kubelet" Aug 13 07:10:20.956299 kubelet[3313]: I0813 07:10:20.955694 3313 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Aug 13 07:10:20.956444 kubelet[3313]: I0813 07:10:20.955252 3313 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Aug 13 07:10:20.957953 kubelet[3313]: I0813 07:10:20.956709 3313 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Aug 13 07:10:20.958166 kubelet[3313]: I0813 07:10:20.958135 3313 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Aug 13 07:10:20.959004 kubelet[3313]: I0813 07:10:20.958985 3313 server.go:449] "Adding debug handlers to kubelet server" Aug 13 07:10:20.968348 kubelet[3313]: I0813 07:10:20.968318 3313 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Aug 13 07:10:20.970032 kubelet[3313]: I0813 07:10:20.970000 3313 volume_manager.go:289] "Starting Kubelet Volume Manager" Aug 13 07:10:20.970438 kubelet[3313]: E0813 07:10:20.970413 3313 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-4230.2.2-a-ad5fc6e7e6\" not found" Aug 13 07:10:20.972245 kubelet[3313]: I0813 07:10:20.972227 3313 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Aug 13 07:10:20.972627 kubelet[3313]: I0813 07:10:20.972615 3313 reconciler.go:26] "Reconciler: start to sync state" Aug 13 07:10:20.974241 kubelet[3313]: I0813 07:10:20.974206 3313 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Aug 13 07:10:20.975207 kubelet[3313]: I0813 07:10:20.975188 3313 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Aug 13 07:10:20.975317 kubelet[3313]: I0813 07:10:20.975306 3313 status_manager.go:217] "Starting to sync pod status with apiserver" Aug 13 07:10:20.975463 kubelet[3313]: I0813 07:10:20.975452 3313 kubelet.go:2321] "Starting kubelet main sync loop" Aug 13 07:10:20.975565 kubelet[3313]: E0813 07:10:20.975550 3313 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Aug 13 07:10:20.987506 kubelet[3313]: I0813 07:10:20.987478 3313 factory.go:221] Registration of the systemd container factory successfully Aug 13 07:10:20.987772 kubelet[3313]: I0813 07:10:20.987752 3313 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Aug 13 07:10:20.990391 kubelet[3313]: E0813 07:10:20.990366 3313 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Aug 13 07:10:20.997823 kubelet[3313]: I0813 07:10:20.997799 3313 factory.go:221] Registration of the containerd container factory successfully Aug 13 07:10:21.039021 kubelet[3313]: I0813 07:10:21.038991 3313 cpu_manager.go:214] "Starting CPU manager" policy="none" Aug 13 07:10:21.039021 kubelet[3313]: I0813 07:10:21.039010 3313 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Aug 13 07:10:21.039021 kubelet[3313]: I0813 07:10:21.039032 3313 state_mem.go:36] "Initialized new in-memory state store" Aug 13 07:10:21.039198 kubelet[3313]: I0813 07:10:21.039175 3313 state_mem.go:88] "Updated default CPUSet" cpuSet="" Aug 13 07:10:21.039222 kubelet[3313]: I0813 07:10:21.039191 3313 state_mem.go:96] "Updated CPUSet assignments" assignments={} Aug 13 07:10:21.039243 kubelet[3313]: I0813 07:10:21.039230 3313 policy_none.go:49] "None policy: Start" Aug 13 07:10:21.040301 kubelet[3313]: I0813 07:10:21.040033 3313 memory_manager.go:170] "Starting memorymanager" policy="None" Aug 13 07:10:21.040301 kubelet[3313]: I0813 07:10:21.040069 3313 state_mem.go:35] "Initializing new in-memory state store" Aug 13 07:10:21.040301 kubelet[3313]: I0813 07:10:21.040213 3313 state_mem.go:75] "Updated machine memory state" Aug 13 07:10:21.045162 kubelet[3313]: I0813 07:10:21.045140 3313 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Aug 13 07:10:21.045729 kubelet[3313]: I0813 07:10:21.045315 3313 eviction_manager.go:189] "Eviction manager: starting control loop" Aug 13 07:10:21.045729 kubelet[3313]: I0813 07:10:21.045333 3313 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Aug 13 07:10:21.046030 kubelet[3313]: I0813 07:10:21.045884 3313 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Aug 13 07:10:21.093357 kubelet[3313]: W0813 07:10:21.092314 3313 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:10:21.099056 kubelet[3313]: W0813 07:10:21.099019 3313 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:10:21.099428 kubelet[3313]: W0813 07:10:21.099337 3313 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Aug 13 07:10:21.099428 kubelet[3313]: E0813 07:10:21.099392 3313 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6\" already exists" pod="kube-system/kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.099428 kubelet[3313]: E0813 07:10:21.099392 3313 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" already exists" pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.148876 kubelet[3313]: I0813 07:10:21.148740 3313 kubelet_node_status.go:72] "Attempting to register node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.169474 kubelet[3313]: I0813 07:10:21.169384 3313 kubelet_node_status.go:111] "Node was previously registered" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.169474 kubelet[3313]: I0813 07:10:21.169462 3313 kubelet_node_status.go:75] "Successfully registered node" node="ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.173862 kubelet[3313]: I0813 07:10:21.173781 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/cfbc4e22bdbd14f9d0eca7156b11f436-kubeconfig\") pod \"kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"cfbc4e22bdbd14f9d0eca7156b11f436\") " pod="kube-system/kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.173942 kubelet[3313]: I0813 07:10:21.173917 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a1e86d3a89911a8cd45e2eee93c504f8-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"a1e86d3a89911a8cd45e2eee93c504f8\") " pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.173986 kubelet[3313]: I0813 07:10:21.173945 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-ca-certs\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.174093 kubelet[3313]: I0813 07:10:21.174064 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-flexvolume-dir\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.174124 kubelet[3313]: I0813 07:10:21.174095 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-k8s-certs\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.174124 kubelet[3313]: I0813 07:10:21.174113 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-kubeconfig\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.174246 kubelet[3313]: I0813 07:10:21.174224 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/19973cd5fe66cbaf3cc4ec7be1d3c8a2-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"19973cd5fe66cbaf3cc4ec7be1d3c8a2\") " pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.174317 kubelet[3313]: I0813 07:10:21.174254 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a1e86d3a89911a8cd45e2eee93c504f8-ca-certs\") pod \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"a1e86d3a89911a8cd45e2eee93c504f8\") " pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.174387 kubelet[3313]: I0813 07:10:21.174270 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a1e86d3a89911a8cd45e2eee93c504f8-k8s-certs\") pod \"kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6\" (UID: \"a1e86d3a89911a8cd45e2eee93c504f8\") " pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" Aug 13 07:10:21.189907 sudo[3345]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Aug 13 07:10:21.190200 sudo[3345]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Aug 13 07:10:21.634320 sudo[3345]: pam_unix(sudo:session): session closed for user root Aug 13 07:10:21.946912 kubelet[3313]: I0813 07:10:21.946323 3313 apiserver.go:52] "Watching apiserver" Aug 13 07:10:21.973211 kubelet[3313]: I0813 07:10:21.973143 3313 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Aug 13 07:10:22.064296 kubelet[3313]: I0813 07:10:22.063983 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4230.2.2-a-ad5fc6e7e6" podStartSLOduration=4.063965659 podStartE2EDuration="4.063965659s" podCreationTimestamp="2025-08-13 07:10:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:22.06371862 +0000 UTC m=+1.182032362" watchObservedRunningTime="2025-08-13 07:10:22.063965659 +0000 UTC m=+1.182279401" Aug 13 07:10:22.064296 kubelet[3313]: I0813 07:10:22.064095 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4230.2.2-a-ad5fc6e7e6" podStartSLOduration=1.064091379 podStartE2EDuration="1.064091379s" podCreationTimestamp="2025-08-13 07:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:22.047767035 +0000 UTC m=+1.166080817" watchObservedRunningTime="2025-08-13 07:10:22.064091379 +0000 UTC m=+1.182405121" Aug 13 07:10:22.090741 kubelet[3313]: I0813 07:10:22.090337 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4230.2.2-a-ad5fc6e7e6" podStartSLOduration=4.090321194 podStartE2EDuration="4.090321194s" podCreationTimestamp="2025-08-13 07:10:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:22.077437447 +0000 UTC m=+1.195751189" watchObservedRunningTime="2025-08-13 07:10:22.090321194 +0000 UTC m=+1.208634936" Aug 13 07:10:23.369042 sudo[2372]: pam_unix(sudo:session): session closed for user root Aug 13 07:10:23.448314 sshd[2371]: Connection closed by 10.200.16.10 port 57698 Aug 13 07:10:23.448771 sshd-session[2369]: pam_unix(sshd:session): session closed for user core Aug 13 07:10:23.451825 systemd[1]: sshd@6-10.200.20.42:22-10.200.16.10:57698.service: Deactivated successfully. Aug 13 07:10:23.454497 systemd[1]: session-9.scope: Deactivated successfully. Aug 13 07:10:23.454773 systemd[1]: session-9.scope: Consumed 8.593s CPU time, 257.3M memory peak. Aug 13 07:10:23.456168 systemd-logind[1702]: Session 9 logged out. Waiting for processes to exit. Aug 13 07:10:23.457939 systemd-logind[1702]: Removed session 9. Aug 13 07:10:24.893237 kubelet[3313]: I0813 07:10:24.893182 3313 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Aug 13 07:10:24.894449 containerd[1741]: time="2025-08-13T07:10:24.894406339Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Aug 13 07:10:24.894970 kubelet[3313]: I0813 07:10:24.894716 3313 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Aug 13 07:10:25.941236 systemd[1]: Created slice kubepods-besteffort-pod93af169b_e52e_4240_8a6a_dd7028943863.slice - libcontainer container kubepods-besteffort-pod93af169b_e52e_4240_8a6a_dd7028943863.slice. Aug 13 07:10:25.953311 systemd[1]: Created slice kubepods-burstable-pod615a2f11_5eb4_45b2_96cd_bf70284e640c.slice - libcontainer container kubepods-burstable-pod615a2f11_5eb4_45b2_96cd_bf70284e640c.slice. Aug 13 07:10:25.998324 systemd[1]: Created slice kubepods-besteffort-pod62b95808_a811_4451_90ed_261de488a09c.slice - libcontainer container kubepods-besteffort-pod62b95808_a811_4451_90ed_261de488a09c.slice. Aug 13 07:10:26.002161 kubelet[3313]: I0813 07:10:26.002107 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cni-path\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003373 kubelet[3313]: I0813 07:10:26.003338 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zds8g\" (UniqueName: \"kubernetes.io/projected/93af169b-e52e-4240-8a6a-dd7028943863-kube-api-access-zds8g\") pod \"kube-proxy-xv8kw\" (UID: \"93af169b-e52e-4240-8a6a-dd7028943863\") " pod="kube-system/kube-proxy-xv8kw" Aug 13 07:10:26.003450 kubelet[3313]: I0813 07:10:26.003379 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/615a2f11-5eb4-45b2-96cd-bf70284e640c-clustermesh-secrets\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003450 kubelet[3313]: I0813 07:10:26.003408 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/62b95808-a811-4451-90ed-261de488a09c-cilium-config-path\") pod \"cilium-operator-5d85765b45-vdl2l\" (UID: \"62b95808-a811-4451-90ed-261de488a09c\") " pod="kube-system/cilium-operator-5d85765b45-vdl2l" Aug 13 07:10:26.003450 kubelet[3313]: I0813 07:10:26.003427 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-cgroup\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003450 kubelet[3313]: I0813 07:10:26.003442 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-config-path\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003544 kubelet[3313]: I0813 07:10:26.003459 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-kernel\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003544 kubelet[3313]: I0813 07:10:26.003475 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-run\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003544 kubelet[3313]: I0813 07:10:26.003494 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pksj6\" (UniqueName: \"kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-kube-api-access-pksj6\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003544 kubelet[3313]: I0813 07:10:26.003511 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g6vd\" (UniqueName: \"kubernetes.io/projected/62b95808-a811-4451-90ed-261de488a09c-kube-api-access-5g6vd\") pod \"cilium-operator-5d85765b45-vdl2l\" (UID: \"62b95808-a811-4451-90ed-261de488a09c\") " pod="kube-system/cilium-operator-5d85765b45-vdl2l" Aug 13 07:10:26.003544 kubelet[3313]: I0813 07:10:26.003528 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/93af169b-e52e-4240-8a6a-dd7028943863-kube-proxy\") pod \"kube-proxy-xv8kw\" (UID: \"93af169b-e52e-4240-8a6a-dd7028943863\") " pod="kube-system/kube-proxy-xv8kw" Aug 13 07:10:26.003647 kubelet[3313]: I0813 07:10:26.003543 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/93af169b-e52e-4240-8a6a-dd7028943863-xtables-lock\") pod \"kube-proxy-xv8kw\" (UID: \"93af169b-e52e-4240-8a6a-dd7028943863\") " pod="kube-system/kube-proxy-xv8kw" Aug 13 07:10:26.003647 kubelet[3313]: I0813 07:10:26.003560 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-etc-cni-netd\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003647 kubelet[3313]: I0813 07:10:26.003586 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-lib-modules\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003647 kubelet[3313]: I0813 07:10:26.003600 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-xtables-lock\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003647 kubelet[3313]: I0813 07:10:26.003613 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-hubble-tls\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003647 kubelet[3313]: I0813 07:10:26.003629 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-hostproc\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003766 kubelet[3313]: I0813 07:10:26.003649 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-bpf-maps\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.003766 kubelet[3313]: I0813 07:10:26.003664 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/93af169b-e52e-4240-8a6a-dd7028943863-lib-modules\") pod \"kube-proxy-xv8kw\" (UID: \"93af169b-e52e-4240-8a6a-dd7028943863\") " pod="kube-system/kube-proxy-xv8kw" Aug 13 07:10:26.003766 kubelet[3313]: I0813 07:10:26.003680 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-net\") pod \"cilium-7847w\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " pod="kube-system/cilium-7847w" Aug 13 07:10:26.250078 containerd[1741]: time="2025-08-13T07:10:26.249898295Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xv8kw,Uid:93af169b-e52e-4240-8a6a-dd7028943863,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:26.258773 containerd[1741]: time="2025-08-13T07:10:26.258490647Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-7847w,Uid:615a2f11-5eb4-45b2-96cd-bf70284e640c,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:26.299150 containerd[1741]: time="2025-08-13T07:10:26.299062248Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:26.299468 containerd[1741]: time="2025-08-13T07:10:26.299321848Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:26.299468 containerd[1741]: time="2025-08-13T07:10:26.299367248Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:26.299630 containerd[1741]: time="2025-08-13T07:10:26.299590248Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:26.301954 containerd[1741]: time="2025-08-13T07:10:26.301900366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-vdl2l,Uid:62b95808-a811-4451-90ed-261de488a09c,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:26.316459 systemd[1]: Started cri-containerd-452a3544fac40da1a49b8cfe526eb0526351f7bef44d41bfea17add00845926e.scope - libcontainer container 452a3544fac40da1a49b8cfe526eb0526351f7bef44d41bfea17add00845926e. Aug 13 07:10:26.332973 containerd[1741]: time="2025-08-13T07:10:26.332858376Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:26.332973 containerd[1741]: time="2025-08-13T07:10:26.332922776Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:26.333503 containerd[1741]: time="2025-08-13T07:10:26.332934176Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:26.333503 containerd[1741]: time="2025-08-13T07:10:26.333006856Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:26.343143 containerd[1741]: time="2025-08-13T07:10:26.342868967Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xv8kw,Uid:93af169b-e52e-4240-8a6a-dd7028943863,Namespace:kube-system,Attempt:0,} returns sandbox id \"452a3544fac40da1a49b8cfe526eb0526351f7bef44d41bfea17add00845926e\"" Aug 13 07:10:26.352751 containerd[1741]: time="2025-08-13T07:10:26.352618798Z" level=info msg="CreateContainer within sandbox \"452a3544fac40da1a49b8cfe526eb0526351f7bef44d41bfea17add00845926e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Aug 13 07:10:26.361511 systemd[1]: Started cri-containerd-310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7.scope - libcontainer container 310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7. Aug 13 07:10:26.372397 containerd[1741]: time="2025-08-13T07:10:26.372255419Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:26.372664 containerd[1741]: time="2025-08-13T07:10:26.372431579Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:26.372820 containerd[1741]: time="2025-08-13T07:10:26.372736778Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:26.373035 containerd[1741]: time="2025-08-13T07:10:26.372942938Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:26.392725 systemd[1]: Started cri-containerd-9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3.scope - libcontainer container 9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3. Aug 13 07:10:26.394043 containerd[1741]: time="2025-08-13T07:10:26.393988678Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-7847w,Uid:615a2f11-5eb4-45b2-96cd-bf70284e640c,Namespace:kube-system,Attempt:0,} returns sandbox id \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\"" Aug 13 07:10:26.398013 containerd[1741]: time="2025-08-13T07:10:26.397775635Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Aug 13 07:10:26.411472 containerd[1741]: time="2025-08-13T07:10:26.411360942Z" level=info msg="CreateContainer within sandbox \"452a3544fac40da1a49b8cfe526eb0526351f7bef44d41bfea17add00845926e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"f8e0759abcdd31e4c99362f36c7f466401d31ff548c12467324042a6cac63c36\"" Aug 13 07:10:26.413480 containerd[1741]: time="2025-08-13T07:10:26.413374020Z" level=info msg="StartContainer for \"f8e0759abcdd31e4c99362f36c7f466401d31ff548c12467324042a6cac63c36\"" Aug 13 07:10:26.438968 systemd[1]: Started cri-containerd-f8e0759abcdd31e4c99362f36c7f466401d31ff548c12467324042a6cac63c36.scope - libcontainer container f8e0759abcdd31e4c99362f36c7f466401d31ff548c12467324042a6cac63c36. Aug 13 07:10:26.440950 containerd[1741]: time="2025-08-13T07:10:26.440912154Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-vdl2l,Uid:62b95808-a811-4451-90ed-261de488a09c,Namespace:kube-system,Attempt:0,} returns sandbox id \"9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3\"" Aug 13 07:10:26.476479 containerd[1741]: time="2025-08-13T07:10:26.476426280Z" level=info msg="StartContainer for \"f8e0759abcdd31e4c99362f36c7f466401d31ff548c12467324042a6cac63c36\" returns successfully" Aug 13 07:10:30.744985 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1072651784.mount: Deactivated successfully. Aug 13 07:10:30.810818 kubelet[3313]: I0813 07:10:30.810758 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-xv8kw" podStartSLOduration=5.810740185 podStartE2EDuration="5.810740185s" podCreationTimestamp="2025-08-13 07:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:27.050104307 +0000 UTC m=+6.168418049" watchObservedRunningTime="2025-08-13 07:10:30.810740185 +0000 UTC m=+9.929053927" Aug 13 07:10:32.338314 containerd[1741]: time="2025-08-13T07:10:32.337451281Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:32.341966 containerd[1741]: time="2025-08-13T07:10:32.341833637Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Aug 13 07:10:32.580406 containerd[1741]: time="2025-08-13T07:10:32.579670157Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:32.584903 containerd[1741]: time="2025-08-13T07:10:32.584864551Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 6.186849756s" Aug 13 07:10:32.584903 containerd[1741]: time="2025-08-13T07:10:32.584902831Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Aug 13 07:10:32.586553 containerd[1741]: time="2025-08-13T07:10:32.585998470Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Aug 13 07:10:32.588418 containerd[1741]: time="2025-08-13T07:10:32.588331668Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 07:10:32.628095 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4088896825.mount: Deactivated successfully. Aug 13 07:10:32.670042 containerd[1741]: time="2025-08-13T07:10:32.669992585Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0\"" Aug 13 07:10:32.671310 containerd[1741]: time="2025-08-13T07:10:32.670584625Z" level=info msg="StartContainer for \"b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0\"" Aug 13 07:10:32.695405 systemd[1]: Started cri-containerd-b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0.scope - libcontainer container b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0. Aug 13 07:10:32.720497 containerd[1741]: time="2025-08-13T07:10:32.720462454Z" level=info msg="StartContainer for \"b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0\" returns successfully" Aug 13 07:10:32.727441 systemd[1]: cri-containerd-b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0.scope: Deactivated successfully. Aug 13 07:10:33.624123 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0-rootfs.mount: Deactivated successfully. Aug 13 07:10:34.146298 containerd[1741]: time="2025-08-13T07:10:34.146198493Z" level=info msg="shim disconnected" id=b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0 namespace=k8s.io Aug 13 07:10:34.146298 containerd[1741]: time="2025-08-13T07:10:34.146301413Z" level=warning msg="cleaning up after shim disconnected" id=b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0 namespace=k8s.io Aug 13 07:10:34.146665 containerd[1741]: time="2025-08-13T07:10:34.146312733Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:35.054798 containerd[1741]: time="2025-08-13T07:10:35.054628748Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Aug 13 07:10:35.202477 containerd[1741]: time="2025-08-13T07:10:35.202434322Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13\"" Aug 13 07:10:35.203227 containerd[1741]: time="2025-08-13T07:10:35.203147562Z" level=info msg="StartContainer for \"4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13\"" Aug 13 07:10:35.231423 systemd[1]: Started cri-containerd-4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13.scope - libcontainer container 4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13. Aug 13 07:10:35.258354 containerd[1741]: time="2025-08-13T07:10:35.258082708Z" level=info msg="StartContainer for \"4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13\" returns successfully" Aug 13 07:10:35.263818 systemd[1]: systemd-sysctl.service: Deactivated successfully. Aug 13 07:10:35.264015 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:10:35.264171 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:10:35.268234 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Aug 13 07:10:35.269654 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Aug 13 07:10:35.272713 systemd[1]: cri-containerd-4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13.scope: Deactivated successfully. Aug 13 07:10:35.292090 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Aug 13 07:10:35.478858 containerd[1741]: time="2025-08-13T07:10:35.478024132Z" level=info msg="shim disconnected" id=4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13 namespace=k8s.io Aug 13 07:10:35.478858 containerd[1741]: time="2025-08-13T07:10:35.478079092Z" level=warning msg="cleaning up after shim disconnected" id=4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13 namespace=k8s.io Aug 13 07:10:35.478858 containerd[1741]: time="2025-08-13T07:10:35.478088292Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:36.058704 containerd[1741]: time="2025-08-13T07:10:36.058464562Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Aug 13 07:10:36.163826 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13-rootfs.mount: Deactivated successfully. Aug 13 07:10:36.886205 containerd[1741]: time="2025-08-13T07:10:36.886156069Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2\"" Aug 13 07:10:36.888166 containerd[1741]: time="2025-08-13T07:10:36.886822868Z" level=info msg="StartContainer for \"e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2\"" Aug 13 07:10:36.913457 systemd[1]: Started cri-containerd-e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2.scope - libcontainer container e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2. Aug 13 07:10:36.940519 systemd[1]: cri-containerd-e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2.scope: Deactivated successfully. Aug 13 07:10:36.975380 containerd[1741]: time="2025-08-13T07:10:36.975260661Z" level=info msg="StartContainer for \"e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2\" returns successfully" Aug 13 07:10:36.993762 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2-rootfs.mount: Deactivated successfully. Aug 13 07:10:38.034681 containerd[1741]: time="2025-08-13T07:10:38.034587021Z" level=info msg="shim disconnected" id=e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2 namespace=k8s.io Aug 13 07:10:38.034681 containerd[1741]: time="2025-08-13T07:10:38.034722380Z" level=warning msg="cleaning up after shim disconnected" id=e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2 namespace=k8s.io Aug 13 07:10:38.034681 containerd[1741]: time="2025-08-13T07:10:38.034733220Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:38.087764 containerd[1741]: time="2025-08-13T07:10:38.087717528Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Aug 13 07:10:38.434432 containerd[1741]: time="2025-08-13T07:10:38.434227228Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb\"" Aug 13 07:10:38.435273 containerd[1741]: time="2025-08-13T07:10:38.435190747Z" level=info msg="StartContainer for \"165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb\"" Aug 13 07:10:38.461415 systemd[1]: Started cri-containerd-165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb.scope - libcontainer container 165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb. Aug 13 07:10:38.480428 systemd[1]: cri-containerd-165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb.scope: Deactivated successfully. Aug 13 07:10:38.485649 containerd[1741]: time="2025-08-13T07:10:38.485609418Z" level=info msg="StartContainer for \"165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb\" returns successfully" Aug 13 07:10:38.500107 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb-rootfs.mount: Deactivated successfully. Aug 13 07:10:40.754653 containerd[1741]: time="2025-08-13T07:10:40.754576909Z" level=info msg="shim disconnected" id=165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb namespace=k8s.io Aug 13 07:10:40.754653 containerd[1741]: time="2025-08-13T07:10:40.754622389Z" level=warning msg="cleaning up after shim disconnected" id=165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb namespace=k8s.io Aug 13 07:10:40.754653 containerd[1741]: time="2025-08-13T07:10:40.754629989Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:10:40.823319 containerd[1741]: time="2025-08-13T07:10:40.822562922Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:40.825321 containerd[1741]: time="2025-08-13T07:10:40.825122800Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Aug 13 07:10:40.829362 containerd[1741]: time="2025-08-13T07:10:40.829319036Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Aug 13 07:10:40.830845 containerd[1741]: time="2025-08-13T07:10:40.830718994Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 8.244685644s" Aug 13 07:10:40.830845 containerd[1741]: time="2025-08-13T07:10:40.830753874Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Aug 13 07:10:40.833208 containerd[1741]: time="2025-08-13T07:10:40.833152912Z" level=info msg="CreateContainer within sandbox \"9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Aug 13 07:10:40.894828 containerd[1741]: time="2025-08-13T07:10:40.894777091Z" level=info msg="CreateContainer within sandbox \"9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\"" Aug 13 07:10:40.895827 containerd[1741]: time="2025-08-13T07:10:40.895333131Z" level=info msg="StartContainer for \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\"" Aug 13 07:10:40.924422 systemd[1]: Started cri-containerd-0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407.scope - libcontainer container 0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407. Aug 13 07:10:40.954608 containerd[1741]: time="2025-08-13T07:10:40.953825313Z" level=info msg="StartContainer for \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\" returns successfully" Aug 13 07:10:41.113828 containerd[1741]: time="2025-08-13T07:10:41.113782236Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Aug 13 07:10:41.179979 containerd[1741]: time="2025-08-13T07:10:41.179938131Z" level=info msg="CreateContainer within sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\"" Aug 13 07:10:41.181698 containerd[1741]: time="2025-08-13T07:10:41.180688290Z" level=info msg="StartContainer for \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\"" Aug 13 07:10:41.186007 kubelet[3313]: I0813 07:10:41.185954 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-vdl2l" podStartSLOduration=1.7971113239999998 podStartE2EDuration="16.185934765s" podCreationTimestamp="2025-08-13 07:10:25 +0000 UTC" firstStartedPulling="2025-08-13 07:10:26.442727752 +0000 UTC m=+5.561041454" lastFinishedPulling="2025-08-13 07:10:40.831551153 +0000 UTC m=+19.949864895" observedRunningTime="2025-08-13 07:10:41.142703488 +0000 UTC m=+20.261017230" watchObservedRunningTime="2025-08-13 07:10:41.185934765 +0000 UTC m=+20.304248507" Aug 13 07:10:41.226450 systemd[1]: Started cri-containerd-9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810.scope - libcontainer container 9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810. Aug 13 07:10:41.272148 containerd[1741]: time="2025-08-13T07:10:41.272104001Z" level=info msg="StartContainer for \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\" returns successfully" Aug 13 07:10:41.475396 kubelet[3313]: I0813 07:10:41.473944 3313 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Aug 13 07:10:41.694492 systemd[1]: Created slice kubepods-burstable-pod1f6377a0_4a51_4387_b613_e1a50416c431.slice - libcontainer container kubepods-burstable-pod1f6377a0_4a51_4387_b613_e1a50416c431.slice. Aug 13 07:10:41.697495 kubelet[3313]: I0813 07:10:41.697418 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2c2c\" (UniqueName: \"kubernetes.io/projected/1f6377a0-4a51-4387-b613-e1a50416c431-kube-api-access-p2c2c\") pod \"coredns-7c65d6cfc9-hxswf\" (UID: \"1f6377a0-4a51-4387-b613-e1a50416c431\") " pod="kube-system/coredns-7c65d6cfc9-hxswf" Aug 13 07:10:41.697495 kubelet[3313]: I0813 07:10:41.697450 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1f6377a0-4a51-4387-b613-e1a50416c431-config-volume\") pod \"coredns-7c65d6cfc9-hxswf\" (UID: \"1f6377a0-4a51-4387-b613-e1a50416c431\") " pod="kube-system/coredns-7c65d6cfc9-hxswf" Aug 13 07:10:41.705093 systemd[1]: Created slice kubepods-burstable-pod783c4528_8824_449c_a5b0_8a02319b0579.slice - libcontainer container kubepods-burstable-pod783c4528_8824_449c_a5b0_8a02319b0579.slice. Aug 13 07:10:41.798150 kubelet[3313]: I0813 07:10:41.798107 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/783c4528-8824-449c-a5b0-8a02319b0579-config-volume\") pod \"coredns-7c65d6cfc9-xm7n8\" (UID: \"783c4528-8824-449c-a5b0-8a02319b0579\") " pod="kube-system/coredns-7c65d6cfc9-xm7n8" Aug 13 07:10:41.798150 kubelet[3313]: I0813 07:10:41.798149 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwvpl\" (UniqueName: \"kubernetes.io/projected/783c4528-8824-449c-a5b0-8a02319b0579-kube-api-access-rwvpl\") pod \"coredns-7c65d6cfc9-xm7n8\" (UID: \"783c4528-8824-449c-a5b0-8a02319b0579\") " pod="kube-system/coredns-7c65d6cfc9-xm7n8" Aug 13 07:10:42.001852 containerd[1741]: time="2025-08-13T07:10:42.001788084Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-hxswf,Uid:1f6377a0-4a51-4387-b613-e1a50416c431,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:42.010634 containerd[1741]: time="2025-08-13T07:10:42.010156196Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-xm7n8,Uid:783c4528-8824-449c-a5b0-8a02319b0579,Namespace:kube-system,Attempt:0,}" Aug 13 07:10:42.137823 kubelet[3313]: I0813 07:10:42.136491 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-7847w" podStartSLOduration=10.946236919 podStartE2EDuration="17.136474152s" podCreationTimestamp="2025-08-13 07:10:25 +0000 UTC" firstStartedPulling="2025-08-13 07:10:26.395493037 +0000 UTC m=+5.513806739" lastFinishedPulling="2025-08-13 07:10:32.58573023 +0000 UTC m=+11.704043972" observedRunningTime="2025-08-13 07:10:42.136353632 +0000 UTC m=+21.254667454" watchObservedRunningTime="2025-08-13 07:10:42.136474152 +0000 UTC m=+21.254787894" Aug 13 07:10:44.421001 systemd-networkd[1609]: cilium_host: Link UP Aug 13 07:10:44.421108 systemd-networkd[1609]: cilium_net: Link UP Aug 13 07:10:44.421111 systemd-networkd[1609]: cilium_net: Gained carrier Aug 13 07:10:44.421234 systemd-networkd[1609]: cilium_host: Gained carrier Aug 13 07:10:44.569260 systemd-networkd[1609]: cilium_vxlan: Link UP Aug 13 07:10:44.569268 systemd-networkd[1609]: cilium_vxlan: Gained carrier Aug 13 07:10:44.804374 kernel: NET: Registered PF_ALG protocol family Aug 13 07:10:44.989441 systemd-networkd[1609]: cilium_host: Gained IPv6LL Aug 13 07:10:45.181424 systemd-networkd[1609]: cilium_net: Gained IPv6LL Aug 13 07:10:45.474360 systemd-networkd[1609]: lxc_health: Link UP Aug 13 07:10:45.489725 systemd-networkd[1609]: lxc_health: Gained carrier Aug 13 07:10:45.605381 systemd-networkd[1609]: lxc9522490680c9: Link UP Aug 13 07:10:45.614309 kernel: eth0: renamed from tmp259fd Aug 13 07:10:45.623101 systemd-networkd[1609]: lxc9522490680c9: Gained carrier Aug 13 07:10:45.643136 systemd-networkd[1609]: lxce662669747d3: Link UP Aug 13 07:10:45.651588 kernel: eth0: renamed from tmp3be13 Aug 13 07:10:45.659033 systemd-networkd[1609]: lxce662669747d3: Gained carrier Aug 13 07:10:46.141414 systemd-networkd[1609]: cilium_vxlan: Gained IPv6LL Aug 13 07:10:46.845505 systemd-networkd[1609]: lxc9522490680c9: Gained IPv6LL Aug 13 07:10:47.102516 systemd-networkd[1609]: lxc_health: Gained IPv6LL Aug 13 07:10:47.678429 systemd-networkd[1609]: lxce662669747d3: Gained IPv6LL Aug 13 07:10:49.071111 containerd[1741]: time="2025-08-13T07:10:49.070979892Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:49.071111 containerd[1741]: time="2025-08-13T07:10:49.071072452Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:49.073706 containerd[1741]: time="2025-08-13T07:10:49.071084332Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:49.077614 containerd[1741]: time="2025-08-13T07:10:49.073686250Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:49.093155 containerd[1741]: time="2025-08-13T07:10:49.092199789Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:10:49.093155 containerd[1741]: time="2025-08-13T07:10:49.092251989Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:10:49.093155 containerd[1741]: time="2025-08-13T07:10:49.092263669Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:49.093155 containerd[1741]: time="2025-08-13T07:10:49.092462749Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:10:49.108753 systemd[1]: Started cri-containerd-259fd9a415d21d8dcd601c2536afa0c45e88620ddb5713860bc5b994b30468f4.scope - libcontainer container 259fd9a415d21d8dcd601c2536afa0c45e88620ddb5713860bc5b994b30468f4. Aug 13 07:10:49.122057 systemd[1]: Started cri-containerd-3be136800b6958b2e935bc2088a213879e5e1a05b0308c0084e16f12fe4e9cb7.scope - libcontainer container 3be136800b6958b2e935bc2088a213879e5e1a05b0308c0084e16f12fe4e9cb7. Aug 13 07:10:49.162679 containerd[1741]: time="2025-08-13T07:10:49.162613433Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-hxswf,Uid:1f6377a0-4a51-4387-b613-e1a50416c431,Namespace:kube-system,Attempt:0,} returns sandbox id \"259fd9a415d21d8dcd601c2536afa0c45e88620ddb5713860bc5b994b30468f4\"" Aug 13 07:10:49.169181 containerd[1741]: time="2025-08-13T07:10:49.168271427Z" level=info msg="CreateContainer within sandbox \"259fd9a415d21d8dcd601c2536afa0c45e88620ddb5713860bc5b994b30468f4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:10:49.173531 containerd[1741]: time="2025-08-13T07:10:49.173454302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-xm7n8,Uid:783c4528-8824-449c-a5b0-8a02319b0579,Namespace:kube-system,Attempt:0,} returns sandbox id \"3be136800b6958b2e935bc2088a213879e5e1a05b0308c0084e16f12fe4e9cb7\"" Aug 13 07:10:49.178235 containerd[1741]: time="2025-08-13T07:10:49.178203656Z" level=info msg="CreateContainer within sandbox \"3be136800b6958b2e935bc2088a213879e5e1a05b0308c0084e16f12fe4e9cb7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Aug 13 07:10:49.229857 containerd[1741]: time="2025-08-13T07:10:49.229744641Z" level=info msg="CreateContainer within sandbox \"259fd9a415d21d8dcd601c2536afa0c45e88620ddb5713860bc5b994b30468f4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ae4bebeb81d0dff4c79e563c16175a4c19651b7674c64d5be5e3ea56840d19b7\"" Aug 13 07:10:49.231233 containerd[1741]: time="2025-08-13T07:10:49.230254120Z" level=info msg="StartContainer for \"ae4bebeb81d0dff4c79e563c16175a4c19651b7674c64d5be5e3ea56840d19b7\"" Aug 13 07:10:49.247347 containerd[1741]: time="2025-08-13T07:10:49.246982142Z" level=info msg="CreateContainer within sandbox \"3be136800b6958b2e935bc2088a213879e5e1a05b0308c0084e16f12fe4e9cb7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"27d8dcb4b58cbb9b75f4deabe275c42e9e20ee291a0ab5817b8d4c1be6a79c5c\"" Aug 13 07:10:49.249734 containerd[1741]: time="2025-08-13T07:10:49.248554740Z" level=info msg="StartContainer for \"27d8dcb4b58cbb9b75f4deabe275c42e9e20ee291a0ab5817b8d4c1be6a79c5c\"" Aug 13 07:10:49.253419 systemd[1]: Started cri-containerd-ae4bebeb81d0dff4c79e563c16175a4c19651b7674c64d5be5e3ea56840d19b7.scope - libcontainer container ae4bebeb81d0dff4c79e563c16175a4c19651b7674c64d5be5e3ea56840d19b7. Aug 13 07:10:49.275456 systemd[1]: Started cri-containerd-27d8dcb4b58cbb9b75f4deabe275c42e9e20ee291a0ab5817b8d4c1be6a79c5c.scope - libcontainer container 27d8dcb4b58cbb9b75f4deabe275c42e9e20ee291a0ab5817b8d4c1be6a79c5c. Aug 13 07:10:49.290846 containerd[1741]: time="2025-08-13T07:10:49.290794895Z" level=info msg="StartContainer for \"ae4bebeb81d0dff4c79e563c16175a4c19651b7674c64d5be5e3ea56840d19b7\" returns successfully" Aug 13 07:10:49.314495 containerd[1741]: time="2025-08-13T07:10:49.314451229Z" level=info msg="StartContainer for \"27d8dcb4b58cbb9b75f4deabe275c42e9e20ee291a0ab5817b8d4c1be6a79c5c\" returns successfully" Aug 13 07:10:50.170433 kubelet[3313]: I0813 07:10:50.170320 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-xm7n8" podStartSLOduration=25.170268663 podStartE2EDuration="25.170268663s" podCreationTimestamp="2025-08-13 07:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:50.151433403 +0000 UTC m=+29.269747145" watchObservedRunningTime="2025-08-13 07:10:50.170268663 +0000 UTC m=+29.288582405" Aug 13 07:10:50.198374 kubelet[3313]: I0813 07:10:50.198194 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-hxswf" podStartSLOduration=25.198176792 podStartE2EDuration="25.198176792s" podCreationTimestamp="2025-08-13 07:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:10:50.172006061 +0000 UTC m=+29.290319803" watchObservedRunningTime="2025-08-13 07:10:50.198176792 +0000 UTC m=+29.316490534" Aug 13 07:10:51.248021 kubelet[3313]: I0813 07:10:51.247903 3313 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Aug 13 07:11:56.611391 systemd[1]: Started sshd@7-10.200.20.42:22-10.200.16.10:49050.service - OpenSSH per-connection server daemon (10.200.16.10:49050). Aug 13 07:11:57.110463 sshd[4691]: Accepted publickey for core from 10.200.16.10 port 49050 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:11:57.111780 sshd-session[4691]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:11:57.115673 systemd-logind[1702]: New session 10 of user core. Aug 13 07:11:57.132414 systemd[1]: Started session-10.scope - Session 10 of User core. Aug 13 07:11:57.532432 sshd[4693]: Connection closed by 10.200.16.10 port 49050 Aug 13 07:11:57.532965 sshd-session[4691]: pam_unix(sshd:session): session closed for user core Aug 13 07:11:57.536393 systemd-logind[1702]: Session 10 logged out. Waiting for processes to exit. Aug 13 07:11:57.536407 systemd[1]: sshd@7-10.200.20.42:22-10.200.16.10:49050.service: Deactivated successfully. Aug 13 07:11:57.538754 systemd[1]: session-10.scope: Deactivated successfully. Aug 13 07:11:57.539724 systemd-logind[1702]: Removed session 10. Aug 13 07:12:02.624203 systemd[1]: Started sshd@8-10.200.20.42:22-10.200.16.10:57294.service - OpenSSH per-connection server daemon (10.200.16.10:57294). Aug 13 07:12:03.105288 sshd[4707]: Accepted publickey for core from 10.200.16.10 port 57294 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:03.106497 sshd-session[4707]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:03.111722 systemd-logind[1702]: New session 11 of user core. Aug 13 07:12:03.115518 systemd[1]: Started session-11.scope - Session 11 of User core. Aug 13 07:12:03.508171 sshd[4709]: Connection closed by 10.200.16.10 port 57294 Aug 13 07:12:03.508861 sshd-session[4707]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:03.512178 systemd[1]: sshd@8-10.200.20.42:22-10.200.16.10:57294.service: Deactivated successfully. Aug 13 07:12:03.513713 systemd[1]: session-11.scope: Deactivated successfully. Aug 13 07:12:03.514544 systemd-logind[1702]: Session 11 logged out. Waiting for processes to exit. Aug 13 07:12:03.515519 systemd-logind[1702]: Removed session 11. Aug 13 07:12:08.596980 systemd[1]: Started sshd@9-10.200.20.42:22-10.200.16.10:57296.service - OpenSSH per-connection server daemon (10.200.16.10:57296). Aug 13 07:12:09.095715 sshd[4722]: Accepted publickey for core from 10.200.16.10 port 57296 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:09.096949 sshd-session[4722]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:09.101670 systemd-logind[1702]: New session 12 of user core. Aug 13 07:12:09.107567 systemd[1]: Started session-12.scope - Session 12 of User core. Aug 13 07:12:09.500667 sshd[4724]: Connection closed by 10.200.16.10 port 57296 Aug 13 07:12:09.501213 sshd-session[4722]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:09.508627 systemd[1]: sshd@9-10.200.20.42:22-10.200.16.10:57296.service: Deactivated successfully. Aug 13 07:12:09.510476 systemd[1]: session-12.scope: Deactivated successfully. Aug 13 07:12:09.511448 systemd-logind[1702]: Session 12 logged out. Waiting for processes to exit. Aug 13 07:12:09.512403 systemd-logind[1702]: Removed session 12. Aug 13 07:12:14.593531 systemd[1]: Started sshd@10-10.200.20.42:22-10.200.16.10:41834.service - OpenSSH per-connection server daemon (10.200.16.10:41834). Aug 13 07:12:15.083674 sshd[4737]: Accepted publickey for core from 10.200.16.10 port 41834 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:15.084959 sshd-session[4737]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:15.089172 systemd-logind[1702]: New session 13 of user core. Aug 13 07:12:15.095418 systemd[1]: Started session-13.scope - Session 13 of User core. Aug 13 07:12:15.514791 sshd[4739]: Connection closed by 10.200.16.10 port 41834 Aug 13 07:12:15.514705 sshd-session[4737]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:15.517863 systemd[1]: sshd@10-10.200.20.42:22-10.200.16.10:41834.service: Deactivated successfully. Aug 13 07:12:15.520503 systemd[1]: session-13.scope: Deactivated successfully. Aug 13 07:12:15.521436 systemd-logind[1702]: Session 13 logged out. Waiting for processes to exit. Aug 13 07:12:15.522503 systemd-logind[1702]: Removed session 13. Aug 13 07:12:15.605489 systemd[1]: Started sshd@11-10.200.20.42:22-10.200.16.10:41840.service - OpenSSH per-connection server daemon (10.200.16.10:41840). Aug 13 07:12:16.058837 sshd[4752]: Accepted publickey for core from 10.200.16.10 port 41840 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:16.060041 sshd-session[4752]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:16.065251 systemd-logind[1702]: New session 14 of user core. Aug 13 07:12:16.072446 systemd[1]: Started session-14.scope - Session 14 of User core. Aug 13 07:12:16.490561 sshd[4754]: Connection closed by 10.200.16.10 port 41840 Aug 13 07:12:16.491164 sshd-session[4752]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:16.494500 systemd[1]: sshd@11-10.200.20.42:22-10.200.16.10:41840.service: Deactivated successfully. Aug 13 07:12:16.496875 systemd[1]: session-14.scope: Deactivated successfully. Aug 13 07:12:16.497618 systemd-logind[1702]: Session 14 logged out. Waiting for processes to exit. Aug 13 07:12:16.498734 systemd-logind[1702]: Removed session 14. Aug 13 07:12:16.583525 systemd[1]: Started sshd@12-10.200.20.42:22-10.200.16.10:41842.service - OpenSSH per-connection server daemon (10.200.16.10:41842). Aug 13 07:12:17.077173 sshd[4764]: Accepted publickey for core from 10.200.16.10 port 41842 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:17.078449 sshd-session[4764]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:17.083205 systemd-logind[1702]: New session 15 of user core. Aug 13 07:12:17.085436 systemd[1]: Started session-15.scope - Session 15 of User core. Aug 13 07:12:17.503729 sshd[4766]: Connection closed by 10.200.16.10 port 41842 Aug 13 07:12:17.504185 sshd-session[4764]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:17.507880 systemd-logind[1702]: Session 15 logged out. Waiting for processes to exit. Aug 13 07:12:17.508728 systemd[1]: sshd@12-10.200.20.42:22-10.200.16.10:41842.service: Deactivated successfully. Aug 13 07:12:17.511116 systemd[1]: session-15.scope: Deactivated successfully. Aug 13 07:12:17.512765 systemd-logind[1702]: Removed session 15. Aug 13 07:12:22.596521 systemd[1]: Started sshd@13-10.200.20.42:22-10.200.16.10:59130.service - OpenSSH per-connection server daemon (10.200.16.10:59130). Aug 13 07:12:23.087305 sshd[4781]: Accepted publickey for core from 10.200.16.10 port 59130 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:23.088576 sshd-session[4781]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:23.093361 systemd-logind[1702]: New session 16 of user core. Aug 13 07:12:23.102432 systemd[1]: Started session-16.scope - Session 16 of User core. Aug 13 07:12:23.492876 sshd[4783]: Connection closed by 10.200.16.10 port 59130 Aug 13 07:12:23.493541 sshd-session[4781]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:23.496765 systemd[1]: sshd@13-10.200.20.42:22-10.200.16.10:59130.service: Deactivated successfully. Aug 13 07:12:23.498851 systemd[1]: session-16.scope: Deactivated successfully. Aug 13 07:12:23.499895 systemd-logind[1702]: Session 16 logged out. Waiting for processes to exit. Aug 13 07:12:23.501098 systemd-logind[1702]: Removed session 16. Aug 13 07:12:28.579508 systemd[1]: Started sshd@14-10.200.20.42:22-10.200.16.10:59134.service - OpenSSH per-connection server daemon (10.200.16.10:59134). Aug 13 07:12:29.030608 sshd[4797]: Accepted publickey for core from 10.200.16.10 port 59134 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:29.031796 sshd-session[4797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:29.036335 systemd-logind[1702]: New session 17 of user core. Aug 13 07:12:29.040609 systemd[1]: Started session-17.scope - Session 17 of User core. Aug 13 07:12:29.430954 sshd[4799]: Connection closed by 10.200.16.10 port 59134 Aug 13 07:12:29.430344 sshd-session[4797]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:29.433266 systemd-logind[1702]: Session 17 logged out. Waiting for processes to exit. Aug 13 07:12:29.434806 systemd[1]: sshd@14-10.200.20.42:22-10.200.16.10:59134.service: Deactivated successfully. Aug 13 07:12:29.437082 systemd[1]: session-17.scope: Deactivated successfully. Aug 13 07:12:29.438441 systemd-logind[1702]: Removed session 17. Aug 13 07:12:29.519501 systemd[1]: Started sshd@15-10.200.20.42:22-10.200.16.10:59144.service - OpenSSH per-connection server daemon (10.200.16.10:59144). Aug 13 07:12:30.012544 sshd[4810]: Accepted publickey for core from 10.200.16.10 port 59144 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:30.013749 sshd-session[4810]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:30.018507 systemd-logind[1702]: New session 18 of user core. Aug 13 07:12:30.027398 systemd[1]: Started session-18.scope - Session 18 of User core. Aug 13 07:12:30.458243 sshd[4812]: Connection closed by 10.200.16.10 port 59144 Aug 13 07:12:30.459538 sshd-session[4810]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:30.462678 systemd-logind[1702]: Session 18 logged out. Waiting for processes to exit. Aug 13 07:12:30.463478 systemd[1]: sshd@15-10.200.20.42:22-10.200.16.10:59144.service: Deactivated successfully. Aug 13 07:12:30.466025 systemd[1]: session-18.scope: Deactivated successfully. Aug 13 07:12:30.467125 systemd-logind[1702]: Removed session 18. Aug 13 07:12:30.553502 systemd[1]: Started sshd@16-10.200.20.42:22-10.200.16.10:53572.service - OpenSSH per-connection server daemon (10.200.16.10:53572). Aug 13 07:12:31.045342 sshd[4822]: Accepted publickey for core from 10.200.16.10 port 53572 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:31.046542 sshd-session[4822]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:31.050488 systemd-logind[1702]: New session 19 of user core. Aug 13 07:12:31.059413 systemd[1]: Started session-19.scope - Session 19 of User core. Aug 13 07:12:32.568830 sshd[4824]: Connection closed by 10.200.16.10 port 53572 Aug 13 07:12:32.569260 sshd-session[4822]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:32.573842 systemd[1]: sshd@16-10.200.20.42:22-10.200.16.10:53572.service: Deactivated successfully. Aug 13 07:12:32.577266 systemd[1]: session-19.scope: Deactivated successfully. Aug 13 07:12:32.578506 systemd-logind[1702]: Session 19 logged out. Waiting for processes to exit. Aug 13 07:12:32.579395 systemd-logind[1702]: Removed session 19. Aug 13 07:12:32.661570 systemd[1]: Started sshd@17-10.200.20.42:22-10.200.16.10:53576.service - OpenSSH per-connection server daemon (10.200.16.10:53576). Aug 13 07:12:33.154306 sshd[4841]: Accepted publickey for core from 10.200.16.10 port 53576 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:33.155575 sshd-session[4841]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:33.159458 systemd-logind[1702]: New session 20 of user core. Aug 13 07:12:33.165423 systemd[1]: Started session-20.scope - Session 20 of User core. Aug 13 07:12:33.694098 sshd[4843]: Connection closed by 10.200.16.10 port 53576 Aug 13 07:12:33.694779 sshd-session[4841]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:33.698559 systemd[1]: sshd@17-10.200.20.42:22-10.200.16.10:53576.service: Deactivated successfully. Aug 13 07:12:33.701496 systemd[1]: session-20.scope: Deactivated successfully. Aug 13 07:12:33.702265 systemd-logind[1702]: Session 20 logged out. Waiting for processes to exit. Aug 13 07:12:33.703114 systemd-logind[1702]: Removed session 20. Aug 13 07:12:33.788502 systemd[1]: Started sshd@18-10.200.20.42:22-10.200.16.10:53588.service - OpenSSH per-connection server daemon (10.200.16.10:53588). Aug 13 07:12:34.279026 sshd[4852]: Accepted publickey for core from 10.200.16.10 port 53588 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:34.280231 sshd-session[4852]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:34.285333 systemd-logind[1702]: New session 21 of user core. Aug 13 07:12:34.293456 systemd[1]: Started session-21.scope - Session 21 of User core. Aug 13 07:12:34.687764 sshd[4854]: Connection closed by 10.200.16.10 port 53588 Aug 13 07:12:34.688546 sshd-session[4852]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:34.691572 systemd[1]: sshd@18-10.200.20.42:22-10.200.16.10:53588.service: Deactivated successfully. Aug 13 07:12:34.693796 systemd[1]: session-21.scope: Deactivated successfully. Aug 13 07:12:34.694781 systemd-logind[1702]: Session 21 logged out. Waiting for processes to exit. Aug 13 07:12:34.695747 systemd-logind[1702]: Removed session 21. Aug 13 07:12:39.781510 systemd[1]: Started sshd@19-10.200.20.42:22-10.200.16.10:53590.service - OpenSSH per-connection server daemon (10.200.16.10:53590). Aug 13 07:12:40.277793 sshd[4868]: Accepted publickey for core from 10.200.16.10 port 53590 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:40.279562 sshd-session[4868]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:40.284890 systemd-logind[1702]: New session 22 of user core. Aug 13 07:12:40.289437 systemd[1]: Started session-22.scope - Session 22 of User core. Aug 13 07:12:40.686568 sshd[4870]: Connection closed by 10.200.16.10 port 53590 Aug 13 07:12:40.687074 sshd-session[4868]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:40.690664 systemd[1]: sshd@19-10.200.20.42:22-10.200.16.10:53590.service: Deactivated successfully. Aug 13 07:12:40.692239 systemd[1]: session-22.scope: Deactivated successfully. Aug 13 07:12:40.693625 systemd-logind[1702]: Session 22 logged out. Waiting for processes to exit. Aug 13 07:12:40.694908 systemd-logind[1702]: Removed session 22. Aug 13 07:12:45.775356 systemd[1]: Started sshd@20-10.200.20.42:22-10.200.16.10:43792.service - OpenSSH per-connection server daemon (10.200.16.10:43792). Aug 13 07:12:46.270148 sshd[4882]: Accepted publickey for core from 10.200.16.10 port 43792 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:46.271395 sshd-session[4882]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:46.275212 systemd-logind[1702]: New session 23 of user core. Aug 13 07:12:46.279415 systemd[1]: Started session-23.scope - Session 23 of User core. Aug 13 07:12:46.690822 sshd[4884]: Connection closed by 10.200.16.10 port 43792 Aug 13 07:12:46.691622 sshd-session[4882]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:46.694732 systemd[1]: sshd@20-10.200.20.42:22-10.200.16.10:43792.service: Deactivated successfully. Aug 13 07:12:46.696954 systemd[1]: session-23.scope: Deactivated successfully. Aug 13 07:12:46.698218 systemd-logind[1702]: Session 23 logged out. Waiting for processes to exit. Aug 13 07:12:46.699673 systemd-logind[1702]: Removed session 23. Aug 13 07:12:51.784590 systemd[1]: Started sshd@21-10.200.20.42:22-10.200.16.10:45154.service - OpenSSH per-connection server daemon (10.200.16.10:45154). Aug 13 07:12:52.279561 sshd[4897]: Accepted publickey for core from 10.200.16.10 port 45154 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:52.280772 sshd-session[4897]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:52.285532 systemd-logind[1702]: New session 24 of user core. Aug 13 07:12:52.289419 systemd[1]: Started session-24.scope - Session 24 of User core. Aug 13 07:12:52.690987 sshd[4899]: Connection closed by 10.200.16.10 port 45154 Aug 13 07:12:52.691460 sshd-session[4897]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:52.695244 systemd[1]: sshd@21-10.200.20.42:22-10.200.16.10:45154.service: Deactivated successfully. Aug 13 07:12:52.697150 systemd[1]: session-24.scope: Deactivated successfully. Aug 13 07:12:52.698042 systemd-logind[1702]: Session 24 logged out. Waiting for processes to exit. Aug 13 07:12:52.699500 systemd-logind[1702]: Removed session 24. Aug 13 07:12:52.783576 systemd[1]: Started sshd@22-10.200.20.42:22-10.200.16.10:45168.service - OpenSSH per-connection server daemon (10.200.16.10:45168). Aug 13 07:12:53.281532 sshd[4910]: Accepted publickey for core from 10.200.16.10 port 45168 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:53.282747 sshd-session[4910]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:53.286630 systemd-logind[1702]: New session 25 of user core. Aug 13 07:12:53.293451 systemd[1]: Started session-25.scope - Session 25 of User core. Aug 13 07:12:55.149445 systemd[1]: run-containerd-runc-k8s.io-9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810-runc.CgIj2e.mount: Deactivated successfully. Aug 13 07:12:55.157614 containerd[1741]: time="2025-08-13T07:12:55.157552649Z" level=info msg="StopContainer for \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\" with timeout 30 (s)" Aug 13 07:12:55.158094 containerd[1741]: time="2025-08-13T07:12:55.157893568Z" level=info msg="Stop container \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\" with signal terminated" Aug 13 07:12:55.165009 containerd[1741]: time="2025-08-13T07:12:55.164965481Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Aug 13 07:12:55.176396 systemd[1]: cri-containerd-0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407.scope: Deactivated successfully. Aug 13 07:12:55.181811 containerd[1741]: time="2025-08-13T07:12:55.181654383Z" level=info msg="StopContainer for \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\" with timeout 2 (s)" Aug 13 07:12:55.183139 containerd[1741]: time="2025-08-13T07:12:55.182321222Z" level=info msg="Stop container \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\" with signal terminated" Aug 13 07:12:55.189623 systemd-networkd[1609]: lxc_health: Link DOWN Aug 13 07:12:55.190636 systemd-networkd[1609]: lxc_health: Lost carrier Aug 13 07:12:55.206228 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407-rootfs.mount: Deactivated successfully. Aug 13 07:12:55.208056 systemd[1]: cri-containerd-9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810.scope: Deactivated successfully. Aug 13 07:12:55.208587 systemd[1]: cri-containerd-9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810.scope: Consumed 6.028s CPU time, 125.3M memory peak, 144K read from disk, 12.9M written to disk. Aug 13 07:12:55.227068 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810-rootfs.mount: Deactivated successfully. Aug 13 07:12:55.291609 containerd[1741]: time="2025-08-13T07:12:55.291435305Z" level=info msg="shim disconnected" id=9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810 namespace=k8s.io Aug 13 07:12:55.291609 containerd[1741]: time="2025-08-13T07:12:55.291555225Z" level=warning msg="cleaning up after shim disconnected" id=9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810 namespace=k8s.io Aug 13 07:12:55.291609 containerd[1741]: time="2025-08-13T07:12:55.291563865Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:12:55.291609 containerd[1741]: time="2025-08-13T07:12:55.291520665Z" level=info msg="shim disconnected" id=0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407 namespace=k8s.io Aug 13 07:12:55.291928 containerd[1741]: time="2025-08-13T07:12:55.291624625Z" level=warning msg="cleaning up after shim disconnected" id=0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407 namespace=k8s.io Aug 13 07:12:55.291928 containerd[1741]: time="2025-08-13T07:12:55.291634305Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:12:55.312023 containerd[1741]: time="2025-08-13T07:12:55.311979083Z" level=info msg="StopContainer for \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\" returns successfully" Aug 13 07:12:55.312627 containerd[1741]: time="2025-08-13T07:12:55.312598282Z" level=info msg="StopPodSandbox for \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\"" Aug 13 07:12:55.312709 containerd[1741]: time="2025-08-13T07:12:55.312638002Z" level=info msg="Container to stop \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:12:55.312709 containerd[1741]: time="2025-08-13T07:12:55.312650522Z" level=info msg="Container to stop \"b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:12:55.312709 containerd[1741]: time="2025-08-13T07:12:55.312660442Z" level=info msg="Container to stop \"4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:12:55.312709 containerd[1741]: time="2025-08-13T07:12:55.312668602Z" level=info msg="Container to stop \"165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:12:55.312709 containerd[1741]: time="2025-08-13T07:12:55.312677402Z" level=info msg="Container to stop \"e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:12:55.314485 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7-shm.mount: Deactivated successfully. Aug 13 07:12:55.316543 containerd[1741]: time="2025-08-13T07:12:55.316468198Z" level=info msg="StopContainer for \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\" returns successfully" Aug 13 07:12:55.317112 containerd[1741]: time="2025-08-13T07:12:55.316987398Z" level=info msg="StopPodSandbox for \"9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3\"" Aug 13 07:12:55.317112 containerd[1741]: time="2025-08-13T07:12:55.317019637Z" level=info msg="Container to stop \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Aug 13 07:12:55.320652 systemd[1]: cri-containerd-310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7.scope: Deactivated successfully. Aug 13 07:12:55.331588 systemd[1]: cri-containerd-9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3.scope: Deactivated successfully. Aug 13 07:12:55.358751 containerd[1741]: time="2025-08-13T07:12:55.358603393Z" level=info msg="shim disconnected" id=310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7 namespace=k8s.io Aug 13 07:12:55.358751 containerd[1741]: time="2025-08-13T07:12:55.358677113Z" level=warning msg="cleaning up after shim disconnected" id=310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7 namespace=k8s.io Aug 13 07:12:55.358751 containerd[1741]: time="2025-08-13T07:12:55.358685673Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:12:55.359183 containerd[1741]: time="2025-08-13T07:12:55.358944513Z" level=info msg="shim disconnected" id=9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3 namespace=k8s.io Aug 13 07:12:55.359183 containerd[1741]: time="2025-08-13T07:12:55.358971272Z" level=warning msg="cleaning up after shim disconnected" id=9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3 namespace=k8s.io Aug 13 07:12:55.359183 containerd[1741]: time="2025-08-13T07:12:55.358977952Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:12:55.371865 containerd[1741]: time="2025-08-13T07:12:55.371656339Z" level=info msg="TearDown network for sandbox \"9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3\" successfully" Aug 13 07:12:55.371865 containerd[1741]: time="2025-08-13T07:12:55.371691739Z" level=info msg="StopPodSandbox for \"9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3\" returns successfully" Aug 13 07:12:55.373003 containerd[1741]: time="2025-08-13T07:12:55.372855378Z" level=info msg="TearDown network for sandbox \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" successfully" Aug 13 07:12:55.373003 containerd[1741]: time="2025-08-13T07:12:55.372881618Z" level=info msg="StopPodSandbox for \"310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7\" returns successfully" Aug 13 07:12:55.576839 kubelet[3313]: I0813 07:12:55.576790 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-config-path\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.576839 kubelet[3313]: I0813 07:12:55.576838 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-xtables-lock\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577234 kubelet[3313]: I0813 07:12:55.576860 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pksj6\" (UniqueName: \"kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-kube-api-access-pksj6\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577234 kubelet[3313]: I0813 07:12:55.576874 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-hostproc\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577234 kubelet[3313]: I0813 07:12:55.576889 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-kernel\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577234 kubelet[3313]: I0813 07:12:55.576926 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-lib-modules\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577234 kubelet[3313]: I0813 07:12:55.576942 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/615a2f11-5eb4-45b2-96cd-bf70284e640c-clustermesh-secrets\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577234 kubelet[3313]: I0813 07:12:55.576957 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-hubble-tls\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577395 kubelet[3313]: I0813 07:12:55.576970 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-cgroup\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577395 kubelet[3313]: I0813 07:12:55.576986 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g6vd\" (UniqueName: \"kubernetes.io/projected/62b95808-a811-4451-90ed-261de488a09c-kube-api-access-5g6vd\") pod \"62b95808-a811-4451-90ed-261de488a09c\" (UID: \"62b95808-a811-4451-90ed-261de488a09c\") " Aug 13 07:12:55.577395 kubelet[3313]: I0813 07:12:55.577002 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-bpf-maps\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577395 kubelet[3313]: I0813 07:12:55.577015 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cni-path\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577395 kubelet[3313]: I0813 07:12:55.577029 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-net\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577395 kubelet[3313]: I0813 07:12:55.577045 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/62b95808-a811-4451-90ed-261de488a09c-cilium-config-path\") pod \"62b95808-a811-4451-90ed-261de488a09c\" (UID: \"62b95808-a811-4451-90ed-261de488a09c\") " Aug 13 07:12:55.577515 kubelet[3313]: I0813 07:12:55.577065 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-run\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577515 kubelet[3313]: I0813 07:12:55.577079 3313 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-etc-cni-netd\") pod \"615a2f11-5eb4-45b2-96cd-bf70284e640c\" (UID: \"615a2f11-5eb4-45b2-96cd-bf70284e640c\") " Aug 13 07:12:55.577515 kubelet[3313]: I0813 07:12:55.577164 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.579528 kubelet[3313]: I0813 07:12:55.578407 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.579528 kubelet[3313]: I0813 07:12:55.578554 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.579528 kubelet[3313]: I0813 07:12:55.579206 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-hostproc" (OuterVolumeSpecName: "hostproc") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.579528 kubelet[3313]: I0813 07:12:55.579259 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.579528 kubelet[3313]: I0813 07:12:55.579297 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.580371 kubelet[3313]: I0813 07:12:55.580344 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 07:12:55.582489 kubelet[3313]: I0813 07:12:55.580623 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.582489 kubelet[3313]: I0813 07:12:55.580641 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.582489 kubelet[3313]: I0813 07:12:55.580667 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cni-path" (OuterVolumeSpecName: "cni-path") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.582489 kubelet[3313]: I0813 07:12:55.581538 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Aug 13 07:12:55.583015 kubelet[3313]: I0813 07:12:55.582957 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 07:12:55.583514 kubelet[3313]: I0813 07:12:55.583478 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b95808-a811-4451-90ed-261de488a09c-kube-api-access-5g6vd" (OuterVolumeSpecName: "kube-api-access-5g6vd") pod "62b95808-a811-4451-90ed-261de488a09c" (UID: "62b95808-a811-4451-90ed-261de488a09c"). InnerVolumeSpecName "kube-api-access-5g6vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 07:12:55.584953 kubelet[3313]: I0813 07:12:55.584816 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/615a2f11-5eb4-45b2-96cd-bf70284e640c-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Aug 13 07:12:55.585858 kubelet[3313]: I0813 07:12:55.585740 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-kube-api-access-pksj6" (OuterVolumeSpecName: "kube-api-access-pksj6") pod "615a2f11-5eb4-45b2-96cd-bf70284e640c" (UID: "615a2f11-5eb4-45b2-96cd-bf70284e640c"). InnerVolumeSpecName "kube-api-access-pksj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Aug 13 07:12:55.586387 kubelet[3313]: I0813 07:12:55.586356 3313 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/62b95808-a811-4451-90ed-261de488a09c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "62b95808-a811-4451-90ed-261de488a09c" (UID: "62b95808-a811-4451-90ed-261de488a09c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Aug 13 07:12:55.677612 kubelet[3313]: I0813 07:12:55.677535 3313 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cni-path\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677744 kubelet[3313]: I0813 07:12:55.677706 3313 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-bpf-maps\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677744 kubelet[3313]: I0813 07:12:55.677722 3313 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-run\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677744 kubelet[3313]: I0813 07:12:55.677735 3313 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-etc-cni-netd\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677744 kubelet[3313]: I0813 07:12:55.677743 3313 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-net\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677837 kubelet[3313]: I0813 07:12:55.677753 3313 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/62b95808-a811-4451-90ed-261de488a09c-cilium-config-path\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677837 kubelet[3313]: I0813 07:12:55.677800 3313 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-config-path\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677837 kubelet[3313]: I0813 07:12:55.677810 3313 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-xtables-lock\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677837 kubelet[3313]: I0813 07:12:55.677818 3313 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pksj6\" (UniqueName: \"kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-kube-api-access-pksj6\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677837 kubelet[3313]: I0813 07:12:55.677826 3313 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-hostproc\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677837 kubelet[3313]: I0813 07:12:55.677834 3313 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-host-proc-sys-kernel\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677953 kubelet[3313]: I0813 07:12:55.677842 3313 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-lib-modules\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677953 kubelet[3313]: I0813 07:12:55.677863 3313 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/615a2f11-5eb4-45b2-96cd-bf70284e640c-clustermesh-secrets\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677953 kubelet[3313]: I0813 07:12:55.677889 3313 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/615a2f11-5eb4-45b2-96cd-bf70284e640c-hubble-tls\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677953 kubelet[3313]: I0813 07:12:55.677898 3313 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/615a2f11-5eb4-45b2-96cd-bf70284e640c-cilium-cgroup\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:55.677953 kubelet[3313]: I0813 07:12:55.677906 3313 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g6vd\" (UniqueName: \"kubernetes.io/projected/62b95808-a811-4451-90ed-261de488a09c-kube-api-access-5g6vd\") on node \"ci-4230.2.2-a-ad5fc6e7e6\" DevicePath \"\"" Aug 13 07:12:56.078771 kubelet[3313]: E0813 07:12:56.078713 3313 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Aug 13 07:12:56.147318 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3-rootfs.mount: Deactivated successfully. Aug 13 07:12:56.147400 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9b84e2193ca4d618ca64cad6b39c53aa5f0e35d376c6688581fbd1dbd031afd3-shm.mount: Deactivated successfully. Aug 13 07:12:56.147454 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-310fc55a29847bf2c8184a1d817549fffaade1f1b3a4fb9dee515d7ec323e4d7-rootfs.mount: Deactivated successfully. Aug 13 07:12:56.147511 systemd[1]: var-lib-kubelet-pods-615a2f11\x2d5eb4\x2d45b2\x2d96cd\x2dbf70284e640c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dpksj6.mount: Deactivated successfully. Aug 13 07:12:56.147563 systemd[1]: var-lib-kubelet-pods-62b95808\x2da811\x2d4451\x2d90ed\x2d261de488a09c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d5g6vd.mount: Deactivated successfully. Aug 13 07:12:56.147610 systemd[1]: var-lib-kubelet-pods-615a2f11\x2d5eb4\x2d45b2\x2d96cd\x2dbf70284e640c-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Aug 13 07:12:56.147657 systemd[1]: var-lib-kubelet-pods-615a2f11\x2d5eb4\x2d45b2\x2d96cd\x2dbf70284e640c-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Aug 13 07:12:56.360150 kubelet[3313]: I0813 07:12:56.358382 3313 scope.go:117] "RemoveContainer" containerID="0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407" Aug 13 07:12:56.360463 containerd[1741]: time="2025-08-13T07:12:56.360357718Z" level=info msg="RemoveContainer for \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\"" Aug 13 07:12:56.366431 systemd[1]: Removed slice kubepods-besteffort-pod62b95808_a811_4451_90ed_261de488a09c.slice - libcontainer container kubepods-besteffort-pod62b95808_a811_4451_90ed_261de488a09c.slice. Aug 13 07:12:56.369673 systemd[1]: Removed slice kubepods-burstable-pod615a2f11_5eb4_45b2_96cd_bf70284e640c.slice - libcontainer container kubepods-burstable-pod615a2f11_5eb4_45b2_96cd_bf70284e640c.slice. Aug 13 07:12:56.369790 systemd[1]: kubepods-burstable-pod615a2f11_5eb4_45b2_96cd_bf70284e640c.slice: Consumed 6.093s CPU time, 125.8M memory peak, 144K read from disk, 12.9M written to disk. Aug 13 07:12:56.377891 containerd[1741]: time="2025-08-13T07:12:56.377857899Z" level=info msg="RemoveContainer for \"0df62d5fa99354902625c82aa0c6e9ce6de3bc9e01eaf9b0c1104d6b2cdd0407\" returns successfully" Aug 13 07:12:56.378081 kubelet[3313]: I0813 07:12:56.378054 3313 scope.go:117] "RemoveContainer" containerID="9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810" Aug 13 07:12:56.379131 containerd[1741]: time="2025-08-13T07:12:56.378918618Z" level=info msg="RemoveContainer for \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\"" Aug 13 07:12:56.389061 containerd[1741]: time="2025-08-13T07:12:56.387903968Z" level=info msg="RemoveContainer for \"9bb9c83ab285eab0244d092f5a97f76eae0a58d519b4a7a36f59c824a2ade810\" returns successfully" Aug 13 07:12:56.389142 kubelet[3313]: I0813 07:12:56.388068 3313 scope.go:117] "RemoveContainer" containerID="165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb" Aug 13 07:12:56.391034 containerd[1741]: time="2025-08-13T07:12:56.390924765Z" level=info msg="RemoveContainer for \"165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb\"" Aug 13 07:12:56.403230 containerd[1741]: time="2025-08-13T07:12:56.403185832Z" level=info msg="RemoveContainer for \"165c189384f1ca4c842f5175bbc7428ecafcf28abc38f94521adab9bec05eeeb\" returns successfully" Aug 13 07:12:56.403985 kubelet[3313]: I0813 07:12:56.403489 3313 scope.go:117] "RemoveContainer" containerID="e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2" Aug 13 07:12:56.407441 containerd[1741]: time="2025-08-13T07:12:56.407409228Z" level=info msg="RemoveContainer for \"e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2\"" Aug 13 07:12:56.419200 containerd[1741]: time="2025-08-13T07:12:56.419128815Z" level=info msg="RemoveContainer for \"e4d459c08bc1e8edc336807f82ff4be7beb493104cc58ac296278518a6205dc2\" returns successfully" Aug 13 07:12:56.419652 kubelet[3313]: I0813 07:12:56.419355 3313 scope.go:117] "RemoveContainer" containerID="4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13" Aug 13 07:12:56.420532 containerd[1741]: time="2025-08-13T07:12:56.420505934Z" level=info msg="RemoveContainer for \"4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13\"" Aug 13 07:12:56.429925 containerd[1741]: time="2025-08-13T07:12:56.429895683Z" level=info msg="RemoveContainer for \"4af9642424dc2dd8749ab752d8a49997636bcb04dcad8cdd89d40ca1f43bcd13\" returns successfully" Aug 13 07:12:56.430096 kubelet[3313]: I0813 07:12:56.430077 3313 scope.go:117] "RemoveContainer" containerID="b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0" Aug 13 07:12:56.432166 containerd[1741]: time="2025-08-13T07:12:56.431788401Z" level=info msg="RemoveContainer for \"b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0\"" Aug 13 07:12:56.445075 containerd[1741]: time="2025-08-13T07:12:56.445018467Z" level=info msg="RemoveContainer for \"b1af2a082e28a412e4c5df8e3419e7132b2930c73e422337021dac5fa596d3e0\" returns successfully" Aug 13 07:12:56.978921 kubelet[3313]: I0813 07:12:56.978884 3313 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" path="/var/lib/kubelet/pods/615a2f11-5eb4-45b2-96cd-bf70284e640c/volumes" Aug 13 07:12:56.979446 kubelet[3313]: I0813 07:12:56.979421 3313 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b95808-a811-4451-90ed-261de488a09c" path="/var/lib/kubelet/pods/62b95808-a811-4451-90ed-261de488a09c/volumes" Aug 13 07:12:57.159118 sshd[4912]: Connection closed by 10.200.16.10 port 45168 Aug 13 07:12:57.159693 sshd-session[4910]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:57.163023 systemd[1]: sshd@22-10.200.20.42:22-10.200.16.10:45168.service: Deactivated successfully. Aug 13 07:12:57.164847 systemd[1]: session-25.scope: Deactivated successfully. Aug 13 07:12:57.165660 systemd-logind[1702]: Session 25 logged out. Waiting for processes to exit. Aug 13 07:12:57.166510 systemd-logind[1702]: Removed session 25. Aug 13 07:12:57.251507 systemd[1]: Started sshd@23-10.200.20.42:22-10.200.16.10:45180.service - OpenSSH per-connection server daemon (10.200.16.10:45180). Aug 13 07:12:57.744758 sshd[5077]: Accepted publickey for core from 10.200.16.10 port 45180 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:12:57.746245 sshd-session[5077]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:12:57.750781 systemd-logind[1702]: New session 26 of user core. Aug 13 07:12:57.757406 systemd[1]: Started session-26.scope - Session 26 of User core. Aug 13 07:12:59.741137 kubelet[3313]: E0813 07:12:59.741081 3313 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="62b95808-a811-4451-90ed-261de488a09c" containerName="cilium-operator" Aug 13 07:12:59.741137 kubelet[3313]: E0813 07:12:59.741109 3313 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" containerName="cilium-agent" Aug 13 07:12:59.741639 kubelet[3313]: E0813 07:12:59.741203 3313 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" containerName="mount-bpf-fs" Aug 13 07:12:59.741639 kubelet[3313]: E0813 07:12:59.741212 3313 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" containerName="clean-cilium-state" Aug 13 07:12:59.741639 kubelet[3313]: E0813 07:12:59.741219 3313 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" containerName="mount-cgroup" Aug 13 07:12:59.741639 kubelet[3313]: E0813 07:12:59.741224 3313 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" containerName="apply-sysctl-overwrites" Aug 13 07:12:59.741639 kubelet[3313]: I0813 07:12:59.741249 3313 memory_manager.go:354] "RemoveStaleState removing state" podUID="615a2f11-5eb4-45b2-96cd-bf70284e640c" containerName="cilium-agent" Aug 13 07:12:59.741639 kubelet[3313]: I0813 07:12:59.741257 3313 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b95808-a811-4451-90ed-261de488a09c" containerName="cilium-operator" Aug 13 07:12:59.750751 systemd[1]: Created slice kubepods-burstable-pod1ce1b10c_25a3_4c59_8876_f6cf9abb590b.slice - libcontainer container kubepods-burstable-pod1ce1b10c_25a3_4c59_8876_f6cf9abb590b.slice. Aug 13 07:12:59.758093 sshd[5079]: Connection closed by 10.200.16.10 port 45180 Aug 13 07:12:59.757811 sshd-session[5077]: pam_unix(sshd:session): session closed for user core Aug 13 07:12:59.764316 systemd[1]: sshd@23-10.200.20.42:22-10.200.16.10:45180.service: Deactivated successfully. Aug 13 07:12:59.767839 systemd[1]: session-26.scope: Deactivated successfully. Aug 13 07:12:59.769369 systemd[1]: session-26.scope: Consumed 1.581s CPU time, 23.8M memory peak. Aug 13 07:12:59.771446 systemd-logind[1702]: Session 26 logged out. Waiting for processes to exit. Aug 13 07:12:59.772787 systemd-logind[1702]: Removed session 26. Aug 13 07:12:59.845517 systemd[1]: Started sshd@24-10.200.20.42:22-10.200.16.10:45188.service - OpenSSH per-connection server daemon (10.200.16.10:45188). Aug 13 07:12:59.897450 kubelet[3313]: I0813 07:12:59.897410 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-clustermesh-secrets\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897450 kubelet[3313]: I0813 07:12:59.897454 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9thnp\" (UniqueName: \"kubernetes.io/projected/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-kube-api-access-9thnp\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897761 kubelet[3313]: I0813 07:12:59.897474 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-hubble-tls\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897761 kubelet[3313]: I0813 07:12:59.897491 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-cni-path\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897761 kubelet[3313]: I0813 07:12:59.897506 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-lib-modules\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897761 kubelet[3313]: I0813 07:12:59.897523 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-bpf-maps\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897761 kubelet[3313]: I0813 07:12:59.897540 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-host-proc-sys-net\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897761 kubelet[3313]: I0813 07:12:59.897558 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-host-proc-sys-kernel\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897887 kubelet[3313]: I0813 07:12:59.897609 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-xtables-lock\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.897887 kubelet[3313]: I0813 07:12:59.897635 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-cilium-config-path\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.898097 kubelet[3313]: I0813 07:12:59.897966 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-cilium-ipsec-secrets\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.898097 kubelet[3313]: I0813 07:12:59.897991 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-cilium-cgroup\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.898097 kubelet[3313]: I0813 07:12:59.898018 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-etc-cni-netd\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.898097 kubelet[3313]: I0813 07:12:59.898034 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-hostproc\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:12:59.898097 kubelet[3313]: I0813 07:12:59.898053 3313 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1ce1b10c-25a3-4c59-8876-f6cf9abb590b-cilium-run\") pod \"cilium-svl4z\" (UID: \"1ce1b10c-25a3-4c59-8876-f6cf9abb590b\") " pod="kube-system/cilium-svl4z" Aug 13 07:13:00.054288 containerd[1741]: time="2025-08-13T07:13:00.054233803Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-svl4z,Uid:1ce1b10c-25a3-4c59-8876-f6cf9abb590b,Namespace:kube-system,Attempt:0,}" Aug 13 07:13:00.132067 containerd[1741]: time="2025-08-13T07:13:00.131968972Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Aug 13 07:13:00.132400 containerd[1741]: time="2025-08-13T07:13:00.132217452Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Aug 13 07:13:00.132400 containerd[1741]: time="2025-08-13T07:13:00.132241492Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:00.132543 containerd[1741]: time="2025-08-13T07:13:00.132353572Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Aug 13 07:13:00.148481 systemd[1]: Started cri-containerd-04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15.scope - libcontainer container 04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15. Aug 13 07:13:00.168086 containerd[1741]: time="2025-08-13T07:13:00.168042059Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-svl4z,Uid:1ce1b10c-25a3-4c59-8876-f6cf9abb590b,Namespace:kube-system,Attempt:0,} returns sandbox id \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\"" Aug 13 07:13:00.172115 containerd[1741]: time="2025-08-13T07:13:00.172080936Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Aug 13 07:13:00.249370 containerd[1741]: time="2025-08-13T07:13:00.249325665Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33\"" Aug 13 07:13:00.249898 containerd[1741]: time="2025-08-13T07:13:00.249785505Z" level=info msg="StartContainer for \"e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33\"" Aug 13 07:13:00.270432 systemd[1]: Started cri-containerd-e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33.scope - libcontainer container e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33. Aug 13 07:13:00.296580 containerd[1741]: time="2025-08-13T07:13:00.296482542Z" level=info msg="StartContainer for \"e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33\" returns successfully" Aug 13 07:13:00.302213 systemd[1]: cri-containerd-e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33.scope: Deactivated successfully. Aug 13 07:13:00.345370 sshd[5090]: Accepted publickey for core from 10.200.16.10 port 45188 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:13:00.346970 sshd-session[5090]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:13:00.351664 systemd-logind[1702]: New session 27 of user core. Aug 13 07:13:00.354523 systemd[1]: Started session-27.scope - Session 27 of User core. Aug 13 07:13:00.397215 containerd[1741]: time="2025-08-13T07:13:00.397024851Z" level=info msg="shim disconnected" id=e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33 namespace=k8s.io Aug 13 07:13:00.397215 containerd[1741]: time="2025-08-13T07:13:00.397079411Z" level=warning msg="cleaning up after shim disconnected" id=e1a6fff738b907964d25455068ef076d6eead8c9a4d9bb2ba9adfdf01028ba33 namespace=k8s.io Aug 13 07:13:00.397215 containerd[1741]: time="2025-08-13T07:13:00.397088371Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:13:00.406795 containerd[1741]: time="2025-08-13T07:13:00.406748802Z" level=warning msg="cleanup warnings time=\"2025-08-13T07:13:00Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Aug 13 07:13:00.708387 sshd[5189]: Connection closed by 10.200.16.10 port 45188 Aug 13 07:13:00.709021 sshd-session[5090]: pam_unix(sshd:session): session closed for user core Aug 13 07:13:00.712181 systemd[1]: sshd@24-10.200.20.42:22-10.200.16.10:45188.service: Deactivated successfully. Aug 13 07:13:00.713774 systemd[1]: session-27.scope: Deactivated successfully. Aug 13 07:13:00.715132 systemd-logind[1702]: Session 27 logged out. Waiting for processes to exit. Aug 13 07:13:00.715991 systemd-logind[1702]: Removed session 27. Aug 13 07:13:00.803533 systemd[1]: Started sshd@25-10.200.20.42:22-10.200.16.10:53794.service - OpenSSH per-connection server daemon (10.200.16.10:53794). Aug 13 07:13:01.079683 kubelet[3313]: E0813 07:13:01.079645 3313 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Aug 13 07:13:01.295296 sshd[5209]: Accepted publickey for core from 10.200.16.10 port 53794 ssh2: RSA SHA256:mUTVkvCTqAM/q6yF06VEIEfaT11Wyv/ewAABhIXzqTw Aug 13 07:13:01.296679 sshd-session[5209]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Aug 13 07:13:01.304577 systemd-logind[1702]: New session 28 of user core. Aug 13 07:13:01.308421 systemd[1]: Started session-28.scope - Session 28 of User core. Aug 13 07:13:01.381679 containerd[1741]: time="2025-08-13T07:13:01.381312756Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Aug 13 07:13:01.415753 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2997835288.mount: Deactivated successfully. Aug 13 07:13:01.430094 containerd[1741]: time="2025-08-13T07:13:01.430046392Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400\"" Aug 13 07:13:01.430697 containerd[1741]: time="2025-08-13T07:13:01.430670071Z" level=info msg="StartContainer for \"d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400\"" Aug 13 07:13:01.455434 systemd[1]: Started cri-containerd-d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400.scope - libcontainer container d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400. Aug 13 07:13:01.482937 systemd[1]: cri-containerd-d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400.scope: Deactivated successfully. Aug 13 07:13:01.484356 containerd[1741]: time="2025-08-13T07:13:01.484236142Z" level=info msg="StartContainer for \"d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400\" returns successfully" Aug 13 07:13:01.527752 containerd[1741]: time="2025-08-13T07:13:01.527686103Z" level=info msg="shim disconnected" id=d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400 namespace=k8s.io Aug 13 07:13:01.527752 containerd[1741]: time="2025-08-13T07:13:01.527746143Z" level=warning msg="cleaning up after shim disconnected" id=d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400 namespace=k8s.io Aug 13 07:13:01.527752 containerd[1741]: time="2025-08-13T07:13:01.527754743Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:13:02.003017 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d00933b2d16a3f3674db537be83232661a95e72f04e53af436a6ee321cfa2400-rootfs.mount: Deactivated successfully. Aug 13 07:13:02.382902 containerd[1741]: time="2025-08-13T07:13:02.382651326Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Aug 13 07:13:02.455301 containerd[1741]: time="2025-08-13T07:13:02.455238820Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b\"" Aug 13 07:13:02.457325 containerd[1741]: time="2025-08-13T07:13:02.455732059Z" level=info msg="StartContainer for \"5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b\"" Aug 13 07:13:02.484441 systemd[1]: Started cri-containerd-5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b.scope - libcontainer container 5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b. Aug 13 07:13:02.511575 systemd[1]: cri-containerd-5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b.scope: Deactivated successfully. Aug 13 07:13:02.513421 containerd[1741]: time="2025-08-13T07:13:02.512964607Z" level=info msg="StartContainer for \"5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b\" returns successfully" Aug 13 07:13:02.567272 containerd[1741]: time="2025-08-13T07:13:02.567209038Z" level=info msg="shim disconnected" id=5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b namespace=k8s.io Aug 13 07:13:02.567272 containerd[1741]: time="2025-08-13T07:13:02.567264118Z" level=warning msg="cleaning up after shim disconnected" id=5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b namespace=k8s.io Aug 13 07:13:02.567775 containerd[1741]: time="2025-08-13T07:13:02.567352918Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:13:03.002092 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5f8fb842e69edd1e2b8addc175d3c7bd39b1a7374664c4956cafa8fcb8ec7b9b-rootfs.mount: Deactivated successfully. Aug 13 07:13:03.388391 containerd[1741]: time="2025-08-13T07:13:03.388269051Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Aug 13 07:13:03.480053 containerd[1741]: time="2025-08-13T07:13:03.480008208Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48\"" Aug 13 07:13:03.480768 containerd[1741]: time="2025-08-13T07:13:03.480613607Z" level=info msg="StartContainer for \"fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48\"" Aug 13 07:13:03.508426 systemd[1]: Started cri-containerd-fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48.scope - libcontainer container fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48. Aug 13 07:13:03.530942 systemd[1]: cri-containerd-fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48.scope: Deactivated successfully. Aug 13 07:13:03.540223 containerd[1741]: time="2025-08-13T07:13:03.540191353Z" level=info msg="StartContainer for \"fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48\" returns successfully" Aug 13 07:13:04.002136 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48-rootfs.mount: Deactivated successfully. Aug 13 07:13:04.333321 containerd[1741]: time="2025-08-13T07:13:04.333232272Z" level=info msg="shim disconnected" id=fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48 namespace=k8s.io Aug 13 07:13:04.333478 containerd[1741]: time="2025-08-13T07:13:04.333322832Z" level=warning msg="cleaning up after shim disconnected" id=fdfc7a55fbcab1453fef16e570ecf25eb319469511961da4b644e3ee0885de48 namespace=k8s.io Aug 13 07:13:04.333478 containerd[1741]: time="2025-08-13T07:13:04.333334232Z" level=info msg="cleaning up dead shim" namespace=k8s.io Aug 13 07:13:04.392310 containerd[1741]: time="2025-08-13T07:13:04.392230939Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Aug 13 07:13:04.543373 kubelet[3313]: I0813 07:13:04.542263 3313 setters.go:600] "Node became not ready" node="ci-4230.2.2-a-ad5fc6e7e6" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-08-13T07:13:04Z","lastTransitionTime":"2025-08-13T07:13:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Aug 13 07:13:04.692517 containerd[1741]: time="2025-08-13T07:13:04.692149946Z" level=info msg="CreateContainer within sandbox \"04f4af22474ee9d8776568e2756e4125b816c6e5b2f802361727a87ad5ae1b15\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"9b469deb13fc5d90d4ef8f507fc0e8ab65fb51651c0c6e42ce90000e8c1ec539\"" Aug 13 07:13:04.693891 containerd[1741]: time="2025-08-13T07:13:04.693854904Z" level=info msg="StartContainer for \"9b469deb13fc5d90d4ef8f507fc0e8ab65fb51651c0c6e42ce90000e8c1ec539\"" Aug 13 07:13:04.724483 systemd[1]: Started cri-containerd-9b469deb13fc5d90d4ef8f507fc0e8ab65fb51651c0c6e42ce90000e8c1ec539.scope - libcontainer container 9b469deb13fc5d90d4ef8f507fc0e8ab65fb51651c0c6e42ce90000e8c1ec539. Aug 13 07:13:04.750939 containerd[1741]: time="2025-08-13T07:13:04.750876372Z" level=info msg="StartContainer for \"9b469deb13fc5d90d4ef8f507fc0e8ab65fb51651c0c6e42ce90000e8c1ec539\" returns successfully" Aug 13 07:13:05.123303 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Aug 13 07:13:05.415554 kubelet[3313]: I0813 07:13:05.415130 3313 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-svl4z" podStartSLOduration=6.415113128 podStartE2EDuration="6.415113128s" podCreationTimestamp="2025-08-13 07:12:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-08-13 07:13:05.414739369 +0000 UTC m=+164.533053151" watchObservedRunningTime="2025-08-13 07:13:05.415113128 +0000 UTC m=+164.533426870" Aug 13 07:13:07.722402 systemd-networkd[1609]: lxc_health: Link UP Aug 13 07:13:07.731123 systemd-networkd[1609]: lxc_health: Gained carrier Aug 13 07:13:07.840268 systemd[1]: run-containerd-runc-k8s.io-9b469deb13fc5d90d4ef8f507fc0e8ab65fb51651c0c6e42ce90000e8c1ec539-runc.WMhkLK.mount: Deactivated successfully. Aug 13 07:13:09.566369 systemd-networkd[1609]: lxc_health: Gained IPv6LL Aug 13 07:13:10.033839 kubelet[3313]: E0813 07:13:10.033673 3313 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:45972->127.0.0.1:37811: write tcp 127.0.0.1:45972->127.0.0.1:37811: write: broken pipe Aug 13 07:13:14.317624 sshd[5211]: Connection closed by 10.200.16.10 port 53794 Aug 13 07:13:14.316975 sshd-session[5209]: pam_unix(sshd:session): session closed for user core Aug 13 07:13:14.320327 systemd[1]: sshd@25-10.200.20.42:22-10.200.16.10:53794.service: Deactivated successfully. Aug 13 07:13:14.322633 systemd[1]: session-28.scope: Deactivated successfully. Aug 13 07:13:14.325842 systemd-logind[1702]: Session 28 logged out. Waiting for processes to exit. Aug 13 07:13:14.327051 systemd-logind[1702]: Removed session 28.