Mar 2 12:53:57.055800 kernel: Booting Linux on physical CPU 0x0000000000 [0x410fd490] Mar 2 12:53:57.055818 kernel: Linux version 6.12.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.3.0 p8) 14.3.0, GNU ld (Gentoo 2.44 p4) 2.44.0) #1 SMP PREEMPT Mon Mar 2 10:44:26 -00 2026 Mar 2 12:53:57.055824 kernel: KASLR enabled Mar 2 12:53:57.055828 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Mar 2 12:53:57.055832 kernel: printk: legacy bootconsole [pl11] enabled Mar 2 12:53:57.055837 kernel: efi: EFI v2.7 by EDK II Mar 2 12:53:57.055842 kernel: efi: ACPI 2.0=0x3f979018 SMBIOS=0x3f8a0000 SMBIOS 3.0=0x3f880000 MEMATTR=0x3e89d018 RNG=0x3f979998 MEMRESERVE=0x3db83598 Mar 2 12:53:57.055846 kernel: random: crng init done Mar 2 12:53:57.055850 kernel: secureboot: Secure boot disabled Mar 2 12:53:57.055853 kernel: ACPI: Early table checksum verification disabled Mar 2 12:53:57.055857 kernel: ACPI: RSDP 0x000000003F979018 000024 (v02 VRTUAL) Mar 2 12:53:57.055861 kernel: ACPI: XSDT 0x000000003F979F18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055865 kernel: ACPI: FACP 0x000000003F979C18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055869 kernel: ACPI: DSDT 0x000000003F95A018 01E046 (v02 MSFTVM DSDT01 00000001 INTL 20230628) Mar 2 12:53:57.055875 kernel: ACPI: DBG2 0x000000003F979B18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055880 kernel: ACPI: GTDT 0x000000003F979D98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055884 kernel: ACPI: OEM0 0x000000003F979098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055888 kernel: ACPI: SPCR 0x000000003F979A98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055892 kernel: ACPI: APIC 0x000000003F979818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055898 kernel: ACPI: SRAT 0x000000003F979198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055902 kernel: ACPI: PPTT 0x000000003F979418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Mar 2 12:53:57.055906 kernel: ACPI: BGRT 0x000000003F979E98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Mar 2 12:53:57.055910 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Mar 2 12:53:57.055914 kernel: ACPI: Use ACPI SPCR as default console: Yes Mar 2 12:53:57.055919 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] hotplug Mar 2 12:53:57.055923 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] hotplug Mar 2 12:53:57.055927 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] hotplug Mar 2 12:53:57.055931 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] hotplug Mar 2 12:53:57.055935 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] hotplug Mar 2 12:53:57.055940 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] hotplug Mar 2 12:53:57.055945 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] hotplug Mar 2 12:53:57.055949 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] hotplug Mar 2 12:53:57.055953 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] hotplug Mar 2 12:53:57.055957 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] hotplug Mar 2 12:53:57.055961 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] hotplug Mar 2 12:53:57.055965 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] hotplug Mar 2 12:53:57.055969 kernel: NUMA: Node 0 [mem 0x00000000-0x3fffffff] + [mem 0x100000000-0x1bfffffff] -> [mem 0x00000000-0x1bfffffff] Mar 2 12:53:57.055973 kernel: NODE_DATA(0) allocated [mem 0x1bf7ffa00-0x1bf806fff] Mar 2 12:53:57.055978 kernel: Zone ranges: Mar 2 12:53:57.055982 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Mar 2 12:53:57.055989 kernel: DMA32 empty Mar 2 12:53:57.055993 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Mar 2 12:53:57.055997 kernel: Device empty Mar 2 12:53:57.056002 kernel: Movable zone start for each node Mar 2 12:53:57.056006 kernel: Early memory node ranges Mar 2 12:53:57.056010 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Mar 2 12:53:57.056015 kernel: node 0: [mem 0x0000000000824000-0x000000003f38ffff] Mar 2 12:53:57.056020 kernel: node 0: [mem 0x000000003f390000-0x000000003f93ffff] Mar 2 12:53:57.056024 kernel: node 0: [mem 0x000000003f940000-0x000000003f9effff] Mar 2 12:53:57.056029 kernel: node 0: [mem 0x000000003f9f0000-0x000000003fdeffff] Mar 2 12:53:57.056033 kernel: node 0: [mem 0x000000003fdf0000-0x000000003fffffff] Mar 2 12:53:57.056037 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Mar 2 12:53:57.056042 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Mar 2 12:53:57.056046 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Mar 2 12:53:57.056050 kernel: cma: Reserved 16 MiB at 0x000000003ca00000 on node -1 Mar 2 12:53:57.056055 kernel: psci: probing for conduit method from ACPI. Mar 2 12:53:57.056059 kernel: psci: PSCIv1.3 detected in firmware. Mar 2 12:53:57.056063 kernel: psci: Using standard PSCI v0.2 function IDs Mar 2 12:53:57.056068 kernel: psci: MIGRATE_INFO_TYPE not supported. Mar 2 12:53:57.056073 kernel: psci: SMC Calling Convention v1.4 Mar 2 12:53:57.056077 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Mar 2 12:53:57.056081 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Mar 2 12:53:57.056086 kernel: percpu: Embedded 33 pages/cpu s98200 r8192 d28776 u135168 Mar 2 12:53:57.056090 kernel: pcpu-alloc: s98200 r8192 d28776 u135168 alloc=33*4096 Mar 2 12:53:57.056095 kernel: pcpu-alloc: [0] 0 [0] 1 Mar 2 12:53:57.056099 kernel: Detected PIPT I-cache on CPU0 Mar 2 12:53:57.056104 kernel: CPU features: detected: Address authentication (architected QARMA5 algorithm) Mar 2 12:53:57.056108 kernel: CPU features: detected: GIC system register CPU interface Mar 2 12:53:57.056112 kernel: CPU features: detected: Spectre-v4 Mar 2 12:53:57.056117 kernel: CPU features: detected: Spectre-BHB Mar 2 12:53:57.056122 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 2 12:53:57.056126 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 2 12:53:57.056131 kernel: CPU features: detected: ARM erratum 2067961 or 2054223 Mar 2 12:53:57.056135 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 2 12:53:57.056139 kernel: alternatives: applying boot alternatives Mar 2 12:53:57.056145 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=75d5e65dbf56ddb5ea243beb025fcfbdb9b2a65e9b1b7d7db3d24aed3f0a168f Mar 2 12:53:57.056150 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 2 12:53:57.056154 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 2 12:53:57.056158 kernel: Fallback order for Node 0: 0 Mar 2 12:53:57.056163 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1048540 Mar 2 12:53:57.056168 kernel: Policy zone: Normal Mar 2 12:53:57.056172 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 2 12:53:57.056177 kernel: software IO TLB: area num 2. Mar 2 12:53:57.056181 kernel: software IO TLB: mapped [mem 0x0000000035900000-0x0000000039900000] (64MB) Mar 2 12:53:57.056185 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Mar 2 12:53:57.056190 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 2 12:53:57.056195 kernel: rcu: RCU event tracing is enabled. Mar 2 12:53:57.056199 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Mar 2 12:53:57.056204 kernel: Trampoline variant of Tasks RCU enabled. Mar 2 12:53:57.056208 kernel: Tracing variant of Tasks RCU enabled. Mar 2 12:53:57.056212 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 2 12:53:57.056217 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Mar 2 12:53:57.056222 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 2 12:53:57.056227 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Mar 2 12:53:57.056231 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 2 12:53:57.056235 kernel: GICv3: 960 SPIs implemented Mar 2 12:53:57.056240 kernel: GICv3: 0 Extended SPIs implemented Mar 2 12:53:57.056244 kernel: Root IRQ handler: gic_handle_irq Mar 2 12:53:57.056248 kernel: GICv3: GICv3 features: 16 PPIs, RSS Mar 2 12:53:57.056253 kernel: GICv3: GICD_CTRL.DS=0, SCR_EL3.FIQ=0 Mar 2 12:53:57.056257 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Mar 2 12:53:57.056262 kernel: ITS: No ITS available, not enabling LPIs Mar 2 12:53:57.056266 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Mar 2 12:53:57.056271 kernel: arch_timer: cp15 timer(s) running at 1000.00MHz (virt). Mar 2 12:53:57.056276 kernel: clocksource: arch_sys_counter: mask: 0x1fffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Mar 2 12:53:57.056280 kernel: sched_clock: 61 bits at 1000MHz, resolution 1ns, wraps every 4398046511103ns Mar 2 12:53:57.056285 kernel: Console: colour dummy device 80x25 Mar 2 12:53:57.056289 kernel: printk: legacy console [tty1] enabled Mar 2 12:53:57.056294 kernel: ACPI: Core revision 20240827 Mar 2 12:53:57.056299 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 2000.00 BogoMIPS (lpj=1000000) Mar 2 12:53:57.056303 kernel: pid_max: default: 32768 minimum: 301 Mar 2 12:53:57.056308 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,ima Mar 2 12:53:57.056312 kernel: landlock: Up and running. Mar 2 12:53:57.056318 kernel: SELinux: Initializing. Mar 2 12:53:57.056322 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 12:53:57.056327 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 2 12:53:57.056332 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3b8030, hints 0xa0000e, misc 0x31e1 Mar 2 12:53:57.056336 kernel: Hyper-V: Host Build 10.0.26102.1212-1-0 Mar 2 12:53:57.056344 kernel: Hyper-V: enabling crash_kexec_post_notifiers Mar 2 12:53:57.056349 kernel: rcu: Hierarchical SRCU implementation. Mar 2 12:53:57.056354 kernel: rcu: Max phase no-delay instances is 400. Mar 2 12:53:57.056359 kernel: Timer migration: 1 hierarchy levels; 8 children per group; 1 crossnode level Mar 2 12:53:57.056364 kernel: Remapping and enabling EFI services. Mar 2 12:53:57.056369 kernel: smp: Bringing up secondary CPUs ... Mar 2 12:53:57.056373 kernel: Detected PIPT I-cache on CPU1 Mar 2 12:53:57.056379 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Mar 2 12:53:57.056384 kernel: CPU1: Booted secondary processor 0x0000000001 [0x410fd490] Mar 2 12:53:57.056389 kernel: smp: Brought up 1 node, 2 CPUs Mar 2 12:53:57.056393 kernel: SMP: Total of 2 processors activated. Mar 2 12:53:57.056398 kernel: CPU: All CPU(s) started at EL1 Mar 2 12:53:57.056404 kernel: CPU features: detected: 32-bit EL0 Support Mar 2 12:53:57.056408 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Mar 2 12:53:57.056413 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 2 12:53:57.056418 kernel: CPU features: detected: Common not Private translations Mar 2 12:53:57.056423 kernel: CPU features: detected: CRC32 instructions Mar 2 12:53:57.056428 kernel: CPU features: detected: Generic authentication (architected QARMA5 algorithm) Mar 2 12:53:57.056432 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 2 12:53:57.056437 kernel: CPU features: detected: LSE atomic instructions Mar 2 12:53:57.056442 kernel: CPU features: detected: Privileged Access Never Mar 2 12:53:57.056448 kernel: CPU features: detected: Speculation barrier (SB) Mar 2 12:53:57.056452 kernel: CPU features: detected: TLB range maintenance instructions Mar 2 12:53:57.056457 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Mar 2 12:53:57.056473 kernel: CPU features: detected: Scalable Vector Extension Mar 2 12:53:57.056478 kernel: alternatives: applying system-wide alternatives Mar 2 12:53:57.056483 kernel: CPU features: detected: Hardware dirty bit management on CPU0-1 Mar 2 12:53:57.056488 kernel: SVE: maximum available vector length 16 bytes per vector Mar 2 12:53:57.056493 kernel: SVE: default vector length 16 bytes per vector Mar 2 12:53:57.056498 kernel: Memory: 3952828K/4194160K available (11200K kernel code, 2458K rwdata, 9088K rodata, 39552K init, 1038K bss, 220144K reserved, 16384K cma-reserved) Mar 2 12:53:57.056504 kernel: devtmpfs: initialized Mar 2 12:53:57.056509 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 2 12:53:57.056513 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Mar 2 12:53:57.056518 kernel: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Mar 2 12:53:57.056523 kernel: 0 pages in range for non-PLT usage Mar 2 12:53:57.056527 kernel: 508400 pages in range for PLT usage Mar 2 12:53:57.056532 kernel: pinctrl core: initialized pinctrl subsystem Mar 2 12:53:57.056537 kernel: SMBIOS 3.1.0 present. Mar 2 12:53:57.056543 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 06/10/2025 Mar 2 12:53:57.056547 kernel: DMI: Memory slots populated: 2/2 Mar 2 12:53:57.056552 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 2 12:53:57.056557 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 2 12:53:57.056562 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 2 12:53:57.056566 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 2 12:53:57.056571 kernel: audit: initializing netlink subsys (disabled) Mar 2 12:53:57.056576 kernel: audit: type=2000 audit(0.059:1): state=initialized audit_enabled=0 res=1 Mar 2 12:53:57.056581 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 2 12:53:57.056586 kernel: cpuidle: using governor menu Mar 2 12:53:57.056591 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 2 12:53:57.056596 kernel: ASID allocator initialised with 32768 entries Mar 2 12:53:57.056601 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 2 12:53:57.056605 kernel: Serial: AMBA PL011 UART driver Mar 2 12:53:57.056610 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Mar 2 12:53:57.056615 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Mar 2 12:53:57.056620 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Mar 2 12:53:57.056624 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Mar 2 12:53:57.056630 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Mar 2 12:53:57.056634 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Mar 2 12:53:57.056639 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Mar 2 12:53:57.056644 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Mar 2 12:53:57.056649 kernel: ACPI: Added _OSI(Module Device) Mar 2 12:53:57.056653 kernel: ACPI: Added _OSI(Processor Device) Mar 2 12:53:57.056658 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 2 12:53:57.056663 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 2 12:53:57.056667 kernel: ACPI: Interpreter enabled Mar 2 12:53:57.056673 kernel: ACPI: Using GIC for interrupt routing Mar 2 12:53:57.056678 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Mar 2 12:53:57.056682 kernel: printk: legacy console [ttyAMA0] enabled Mar 2 12:53:57.056687 kernel: printk: legacy bootconsole [pl11] disabled Mar 2 12:53:57.056692 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Mar 2 12:53:57.056697 kernel: ACPI: CPU0 has been hot-added Mar 2 12:53:57.056701 kernel: ACPI: CPU1 has been hot-added Mar 2 12:53:57.056706 kernel: iommu: Default domain type: Translated Mar 2 12:53:57.056711 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 2 12:53:57.056716 kernel: efivars: Registered efivars operations Mar 2 12:53:57.056721 kernel: vgaarb: loaded Mar 2 12:53:57.056726 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 2 12:53:57.056731 kernel: VFS: Disk quotas dquot_6.6.0 Mar 2 12:53:57.056735 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 2 12:53:57.056740 kernel: pnp: PnP ACPI init Mar 2 12:53:57.056745 kernel: pnp: PnP ACPI: found 0 devices Mar 2 12:53:57.056749 kernel: NET: Registered PF_INET protocol family Mar 2 12:53:57.056754 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 2 12:53:57.056759 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 2 12:53:57.056765 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 2 12:53:57.056769 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 2 12:53:57.056774 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Mar 2 12:53:57.056779 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 2 12:53:57.056784 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 12:53:57.056789 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 2 12:53:57.056793 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 2 12:53:57.056798 kernel: PCI: CLS 0 bytes, default 64 Mar 2 12:53:57.056803 kernel: kvm [1]: HYP mode not available Mar 2 12:53:57.056808 kernel: Initialise system trusted keyrings Mar 2 12:53:57.056813 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 2 12:53:57.056818 kernel: Key type asymmetric registered Mar 2 12:53:57.056823 kernel: Asymmetric key parser 'x509' registered Mar 2 12:53:57.056827 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Mar 2 12:53:57.056832 kernel: io scheduler mq-deadline registered Mar 2 12:53:57.056837 kernel: io scheduler kyber registered Mar 2 12:53:57.056842 kernel: io scheduler bfq registered Mar 2 12:53:57.056846 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 2 12:53:57.056852 kernel: thunder_xcv, ver 1.0 Mar 2 12:53:57.056857 kernel: thunder_bgx, ver 1.0 Mar 2 12:53:57.056861 kernel: nicpf, ver 1.0 Mar 2 12:53:57.056866 kernel: nicvf, ver 1.0 Mar 2 12:53:57.056973 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 2 12:53:57.057025 kernel: rtc-efi rtc-efi.0: setting system clock to 2026-03-02T12:53:56 UTC (1772456036) Mar 2 12:53:57.057031 kernel: efifb: probing for efifb Mar 2 12:53:57.057037 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Mar 2 12:53:57.057042 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Mar 2 12:53:57.057047 kernel: efifb: scrolling: redraw Mar 2 12:53:57.057052 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Mar 2 12:53:57.057056 kernel: Console: switching to colour frame buffer device 128x48 Mar 2 12:53:57.057061 kernel: fb0: EFI VGA frame buffer device Mar 2 12:53:57.057066 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Mar 2 12:53:57.057071 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 2 12:53:57.057076 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 (0,8000003f) counters available Mar 2 12:53:57.057081 kernel: watchdog: NMI not fully supported Mar 2 12:53:57.057086 kernel: watchdog: Hard watchdog permanently disabled Mar 2 12:53:57.057091 kernel: NET: Registered PF_INET6 protocol family Mar 2 12:53:57.057096 kernel: Segment Routing with IPv6 Mar 2 12:53:57.057101 kernel: In-situ OAM (IOAM) with IPv6 Mar 2 12:53:57.057105 kernel: NET: Registered PF_PACKET protocol family Mar 2 12:53:57.057110 kernel: Key type dns_resolver registered Mar 2 12:53:57.057115 kernel: registered taskstats version 1 Mar 2 12:53:57.057120 kernel: Loading compiled-in X.509 certificates Mar 2 12:53:57.057124 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.12.74-flatcar: 03854795d80c6b1eedd5f94f64a67d19428ce88e' Mar 2 12:53:57.057130 kernel: Demotion targets for Node 0: null Mar 2 12:53:57.057135 kernel: Key type .fscrypt registered Mar 2 12:53:57.057139 kernel: Key type fscrypt-provisioning registered Mar 2 12:53:57.057144 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 2 12:53:57.057149 kernel: ima: Allocated hash algorithm: sha1 Mar 2 12:53:57.057153 kernel: ima: No architecture policies found Mar 2 12:53:57.057158 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 2 12:53:57.057163 kernel: clk: Disabling unused clocks Mar 2 12:53:57.057168 kernel: PM: genpd: Disabling unused power domains Mar 2 12:53:57.057173 kernel: Warning: unable to open an initial console. Mar 2 12:53:57.057178 kernel: Freeing unused kernel memory: 39552K Mar 2 12:53:57.057183 kernel: Run /init as init process Mar 2 12:53:57.057188 kernel: with arguments: Mar 2 12:53:57.057192 kernel: /init Mar 2 12:53:57.057197 kernel: with environment: Mar 2 12:53:57.057202 kernel: HOME=/ Mar 2 12:53:57.057206 kernel: TERM=linux Mar 2 12:53:57.057212 systemd[1]: Successfully made /usr/ read-only. Mar 2 12:53:57.057220 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 12:53:57.057226 systemd[1]: Detected virtualization microsoft. Mar 2 12:53:57.057231 systemd[1]: Detected architecture arm64. Mar 2 12:53:57.057236 systemd[1]: Running in initrd. Mar 2 12:53:57.057241 systemd[1]: No hostname configured, using default hostname. Mar 2 12:53:57.057246 systemd[1]: Hostname set to . Mar 2 12:53:57.057251 systemd[1]: Initializing machine ID from random generator. Mar 2 12:53:57.057257 systemd[1]: Queued start job for default target initrd.target. Mar 2 12:53:57.057262 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 12:53:57.057267 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 12:53:57.057273 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Mar 2 12:53:57.057278 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 12:53:57.057283 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Mar 2 12:53:57.057289 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Mar 2 12:53:57.057295 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Mar 2 12:53:57.057301 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Mar 2 12:53:57.057306 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 12:53:57.057311 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 12:53:57.057316 systemd[1]: Reached target paths.target - Path Units. Mar 2 12:53:57.057321 systemd[1]: Reached target slices.target - Slice Units. Mar 2 12:53:57.057326 systemd[1]: Reached target swap.target - Swaps. Mar 2 12:53:57.057331 systemd[1]: Reached target timers.target - Timer Units. Mar 2 12:53:57.057337 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 12:53:57.057342 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 12:53:57.057348 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Mar 2 12:53:57.057353 systemd[1]: Listening on systemd-journald.socket - Journal Sockets. Mar 2 12:53:57.057358 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 12:53:57.057363 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 12:53:57.057368 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 12:53:57.057373 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 12:53:57.057379 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Mar 2 12:53:57.057384 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 12:53:57.057390 systemd[1]: Finished network-cleanup.service - Network Cleanup. Mar 2 12:53:57.057395 systemd[1]: systemd-battery-check.service - Check battery level during early boot was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/class/power_supply). Mar 2 12:53:57.057400 systemd[1]: Starting systemd-fsck-usr.service... Mar 2 12:53:57.057406 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 12:53:57.057411 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 12:53:57.057434 systemd-journald[225]: Collecting audit messages is disabled. Mar 2 12:53:57.057448 systemd-journald[225]: Journal started Mar 2 12:53:57.057469 systemd-journald[225]: Runtime Journal (/run/log/journal/79800782aa1244e2b421ffd9da39fa4c) is 8M, max 78.3M, 70.3M free. Mar 2 12:53:57.065501 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:53:57.070666 systemd-modules-load[227]: Inserted module 'overlay' Mar 2 12:53:57.092517 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 2 12:53:57.092558 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 12:53:57.099082 systemd-modules-load[227]: Inserted module 'br_netfilter' Mar 2 12:53:57.102656 kernel: Bridge firewalling registered Mar 2 12:53:57.101484 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Mar 2 12:53:57.111741 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 12:53:57.117753 systemd[1]: Finished systemd-fsck-usr.service. Mar 2 12:53:57.128486 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 12:53:57.133836 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:53:57.144358 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Mar 2 12:53:57.158673 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 12:53:57.171634 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Mar 2 12:53:57.178953 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 12:53:57.190810 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:53:57.201878 systemd-tmpfiles[247]: /usr/lib/tmpfiles.d/var.conf:14: Duplicate line for path "/var/log", ignoring. Mar 2 12:53:57.213605 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 12:53:57.226486 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Mar 2 12:53:57.235329 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 12:53:57.247282 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Mar 2 12:53:57.265589 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 12:53:57.276106 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 12:53:57.287187 dracut-cmdline[261]: Using kernel command line parameters: rd.driver.pre=btrfs SYSTEMD_SULOGIN_FORCE=1 BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=75d5e65dbf56ddb5ea243beb025fcfbdb9b2a65e9b1b7d7db3d24aed3f0a168f Mar 2 12:53:57.324293 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 12:53:57.333163 systemd-resolved[263]: Positive Trust Anchors: Mar 2 12:53:57.333171 systemd-resolved[263]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 12:53:57.333190 systemd-resolved[263]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 12:53:57.334780 systemd-resolved[263]: Defaulting to hostname 'linux'. Mar 2 12:53:57.336119 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 12:53:57.341730 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 12:53:57.445488 kernel: SCSI subsystem initialized Mar 2 12:53:57.450473 kernel: Loading iSCSI transport class v2.0-870. Mar 2 12:53:57.458491 kernel: iscsi: registered transport (tcp) Mar 2 12:53:57.471054 kernel: iscsi: registered transport (qla4xxx) Mar 2 12:53:57.471066 kernel: QLogic iSCSI HBA Driver Mar 2 12:53:57.485026 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 12:53:57.508052 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 12:53:57.519407 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 12:53:57.563617 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Mar 2 12:53:57.570581 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Mar 2 12:53:57.628482 kernel: raid6: neonx8 gen() 18552 MB/s Mar 2 12:53:57.647471 kernel: raid6: neonx4 gen() 18537 MB/s Mar 2 12:53:57.666471 kernel: raid6: neonx2 gen() 17068 MB/s Mar 2 12:53:57.686473 kernel: raid6: neonx1 gen() 15073 MB/s Mar 2 12:53:57.705472 kernel: raid6: int64x8 gen() 10551 MB/s Mar 2 12:53:57.724472 kernel: raid6: int64x4 gen() 10608 MB/s Mar 2 12:53:57.744566 kernel: raid6: int64x2 gen() 8985 MB/s Mar 2 12:53:57.765955 kernel: raid6: int64x1 gen() 7009 MB/s Mar 2 12:53:57.766036 kernel: raid6: using algorithm neonx8 gen() 18552 MB/s Mar 2 12:53:57.788648 kernel: raid6: .... xor() 14904 MB/s, rmw enabled Mar 2 12:53:57.788701 kernel: raid6: using neon recovery algorithm Mar 2 12:53:57.796426 kernel: xor: measuring software checksum speed Mar 2 12:53:57.796443 kernel: 8regs : 28641 MB/sec Mar 2 12:53:57.798875 kernel: 32regs : 28763 MB/sec Mar 2 12:53:57.801401 kernel: arm64_neon : 37511 MB/sec Mar 2 12:53:57.804320 kernel: xor: using function: arm64_neon (37511 MB/sec) Mar 2 12:53:57.842480 kernel: Btrfs loaded, zoned=no, fsverity=no Mar 2 12:53:57.848212 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Mar 2 12:53:57.857156 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 12:53:57.886657 systemd-udevd[475]: Using default interface naming scheme 'v255'. Mar 2 12:53:57.891586 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 12:53:57.904508 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Mar 2 12:53:57.927713 dracut-pre-trigger[487]: rd.md=0: removing MD RAID activation Mar 2 12:53:57.947127 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 12:53:57.952620 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 12:53:58.001378 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 12:53:58.016163 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Mar 2 12:53:58.071499 kernel: hv_vmbus: Vmbus version:5.3 Mar 2 12:53:58.078047 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:53:58.082052 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:53:58.104030 kernel: pps_core: LinuxPPS API ver. 1 registered Mar 2 12:53:58.104050 kernel: hv_vmbus: registering driver hv_storvsc Mar 2 12:53:58.104066 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Mar 2 12:53:58.095924 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:53:58.139971 kernel: scsi host0: storvsc_host_t Mar 2 12:53:58.140127 kernel: hv_vmbus: registering driver hv_netvsc Mar 2 12:53:58.140135 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Mar 2 12:53:58.140215 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 5 Mar 2 12:53:58.140283 kernel: scsi host1: storvsc_host_t Mar 2 12:53:58.140342 kernel: PTP clock support registered Mar 2 12:53:58.140348 kernel: hv_vmbus: registering driver hid_hyperv Mar 2 12:53:58.125934 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:53:58.161307 kernel: hv_utils: Registering HyperV Utility Driver Mar 2 12:53:58.161349 kernel: hv_vmbus: registering driver hv_utils Mar 2 12:53:58.156572 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:53:58.172220 kernel: hv_vmbus: registering driver hyperv_keyboard Mar 2 12:53:58.172241 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input0 Mar 2 12:53:58.156655 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:53:57.951529 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Mar 2 12:53:57.964183 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Mar 2 12:53:57.964335 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Mar 2 12:53:57.964496 kernel: hv_utils: Heartbeat IC version 3.0 Mar 2 12:53:57.964518 kernel: hv_utils: Shutdown IC version 3.2 Mar 2 12:53:57.964538 kernel: sd 0:0:0:0: [sda] Write Protect is off Mar 2 12:53:57.964689 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Mar 2 12:53:57.964754 kernel: hv_utils: TimeSync IC version 4.0 Mar 2 12:53:57.964760 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input1 Mar 2 12:53:57.964768 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Mar 2 12:53:57.964828 kernel: hv_netvsc 7ced8d88-5e8d-7ced-8d88-5e8d7ced8d88 eth0: VF slot 1 added Mar 2 12:53:57.964898 systemd-journald[225]: Time jumped backwards, rotating. Mar 2 12:53:57.964929 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 2 12:53:57.964935 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Mar 2 12:53:58.181834 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:53:58.183108 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:53:57.924139 systemd-resolved[263]: Clock change detected. Flushing caches. Mar 2 12:53:57.993769 kernel: hv_vmbus: registering driver hv_pci Mar 2 12:53:57.993802 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Mar 2 12:53:57.993964 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Mar 2 12:53:57.994437 kernel: hv_pci 04a712f1-5234-4bc1-99d3-62fe35243ccd: PCI VMBus probing: Using version 0x10004 Mar 2 12:53:57.999449 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Mar 2 12:53:57.999735 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:53:58.022600 kernel: hv_pci 04a712f1-5234-4bc1-99d3-62fe35243ccd: PCI host bridge to bus 5234:00 Mar 2 12:53:58.022761 kernel: pci_bus 5234:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Mar 2 12:53:58.022850 kernel: pci_bus 5234:00: No busn resource found for root bus, will use [bus 00-ff] Mar 2 12:53:58.028499 kernel: pci 5234:00:02.0: [15b3:101a] type 00 class 0x020000 PCIe Endpoint Mar 2 12:53:58.034435 kernel: pci 5234:00:02.0: BAR 0 [mem 0xfc0000000-0xfc00fffff 64bit pref] Mar 2 12:53:58.039419 kernel: pci 5234:00:02.0: enabling Extended Tags Mar 2 12:53:58.053454 kernel: pci 5234:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 5234:00:02.0 (capable of 252.048 Gb/s with 16.0 GT/s PCIe x16 link) Mar 2 12:53:58.061963 kernel: pci_bus 5234:00: busn_res: [bus 00-ff] end is updated to 00 Mar 2 12:53:58.062132 kernel: pci 5234:00:02.0: BAR 0 [mem 0xfc0000000-0xfc00fffff 64bit pref]: assigned Mar 2 12:53:58.084466 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#98 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 2 12:53:58.104472 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#73 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 2 12:53:58.124822 kernel: mlx5_core 5234:00:02.0: enabling device (0000 -> 0002) Mar 2 12:53:58.133460 kernel: mlx5_core 5234:00:02.0: PTM is not supported by PCIe Mar 2 12:53:58.133669 kernel: mlx5_core 5234:00:02.0: firmware version: 16.30.5026 Mar 2 12:53:58.309257 kernel: hv_netvsc 7ced8d88-5e8d-7ced-8d88-5e8d7ced8d88 eth0: VF registering: eth1 Mar 2 12:53:58.309474 kernel: mlx5_core 5234:00:02.0 eth1: joined to eth0 Mar 2 12:53:58.314147 kernel: mlx5_core 5234:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Mar 2 12:53:58.324682 kernel: mlx5_core 5234:00:02.0 enP21044s1: renamed from eth1 Mar 2 12:53:58.439919 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Mar 2 12:53:58.511678 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 2 12:53:58.562705 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Mar 2 12:53:58.568043 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Mar 2 12:53:58.578919 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Mar 2 12:53:58.608819 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Mar 2 12:53:58.618750 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Mar 2 12:53:58.625294 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 12:53:58.630338 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 12:53:58.649326 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 12:53:58.654424 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Mar 2 12:53:58.673932 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 2 12:53:58.687480 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Mar 2 12:53:59.689166 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Mar 2 12:53:59.689298 disk-uuid[660]: The operation has completed successfully. Mar 2 12:53:59.759265 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 2 12:53:59.759380 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Mar 2 12:53:59.782990 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Mar 2 12:53:59.803721 sh[828]: Success Mar 2 12:53:59.837537 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 2 12:53:59.837600 kernel: device-mapper: uevent: version 1.0.3 Mar 2 12:53:59.842370 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@lists.linux.dev Mar 2 12:53:59.852616 kernel: device-mapper: verity: sha256 using shash "sha256-ce" Mar 2 12:54:00.117295 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Mar 2 12:54:00.124670 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Mar 2 12:54:00.141286 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Mar 2 12:54:00.159431 kernel: BTRFS: device fsid da6bd89d-75a6-483e-9a3e-89df5ed9b6c4 devid 1 transid 36 /dev/mapper/usr (254:0) scanned by mount (846) Mar 2 12:54:00.168967 kernel: BTRFS info (device dm-0): first mount of filesystem da6bd89d-75a6-483e-9a3e-89df5ed9b6c4 Mar 2 12:54:00.169002 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:54:00.490146 kernel: BTRFS info (device dm-0 state E): disabling log replay at mount time Mar 2 12:54:00.490223 kernel: BTRFS info (device dm-0 state E): enabling free space tree Mar 2 12:54:00.521669 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Mar 2 12:54:00.525332 systemd[1]: Reached target initrd-usr-fs.target - Initrd /usr File System. Mar 2 12:54:00.532724 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Mar 2 12:54:00.533361 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Mar 2 12:54:00.554418 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Mar 2 12:54:00.581475 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (869) Mar 2 12:54:00.591727 kernel: BTRFS info (device sda6): first mount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:54:00.591764 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:54:00.615760 kernel: BTRFS info (device sda6): turning on async discard Mar 2 12:54:00.615817 kernel: BTRFS info (device sda6): enabling free space tree Mar 2 12:54:00.624481 kernel: BTRFS info (device sda6): last unmount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:54:00.625674 systemd[1]: Finished ignition-setup.service - Ignition (setup). Mar 2 12:54:00.631221 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Mar 2 12:54:00.670154 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 12:54:00.680564 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 12:54:00.711712 systemd-networkd[1015]: lo: Link UP Mar 2 12:54:00.711724 systemd-networkd[1015]: lo: Gained carrier Mar 2 12:54:00.712440 systemd-networkd[1015]: Enumeration completed Mar 2 12:54:00.714488 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 12:54:00.719547 systemd[1]: Reached target network.target - Network. Mar 2 12:54:00.722285 systemd-networkd[1015]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:54:00.722288 systemd-networkd[1015]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 12:54:00.791423 kernel: mlx5_core 5234:00:02.0 enP21044s1: Link up Mar 2 12:54:00.823442 kernel: hv_netvsc 7ced8d88-5e8d-7ced-8d88-5e8d7ced8d88 eth0: Data path switched to VF: enP21044s1 Mar 2 12:54:00.823278 systemd-networkd[1015]: enP21044s1: Link UP Mar 2 12:54:00.823342 systemd-networkd[1015]: eth0: Link UP Mar 2 12:54:00.824320 systemd-networkd[1015]: eth0: Gained carrier Mar 2 12:54:00.824336 systemd-networkd[1015]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:54:00.832594 systemd-networkd[1015]: enP21044s1: Gained carrier Mar 2 12:54:00.851438 systemd-networkd[1015]: eth0: DHCPv4 address 10.200.20.30/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 2 12:54:01.719886 ignition[964]: Ignition 2.22.0 Mar 2 12:54:01.719897 ignition[964]: Stage: fetch-offline Mar 2 12:54:01.724248 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 12:54:01.719997 ignition[964]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:01.732137 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Mar 2 12:54:01.720004 ignition[964]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:01.720086 ignition[964]: parsed url from cmdline: "" Mar 2 12:54:01.720089 ignition[964]: no config URL provided Mar 2 12:54:01.720095 ignition[964]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 12:54:01.720100 ignition[964]: no config at "/usr/lib/ignition/user.ign" Mar 2 12:54:01.720104 ignition[964]: failed to fetch config: resource requires networking Mar 2 12:54:01.720217 ignition[964]: Ignition finished successfully Mar 2 12:54:01.771915 ignition[1026]: Ignition 2.22.0 Mar 2 12:54:01.771932 ignition[1026]: Stage: fetch Mar 2 12:54:01.772091 ignition[1026]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:01.772098 ignition[1026]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:01.772160 ignition[1026]: parsed url from cmdline: "" Mar 2 12:54:01.772163 ignition[1026]: no config URL provided Mar 2 12:54:01.772166 ignition[1026]: reading system config file "/usr/lib/ignition/user.ign" Mar 2 12:54:01.772171 ignition[1026]: no config at "/usr/lib/ignition/user.ign" Mar 2 12:54:01.772186 ignition[1026]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Mar 2 12:54:01.842450 ignition[1026]: GET result: OK Mar 2 12:54:01.842516 ignition[1026]: config has been read from IMDS userdata Mar 2 12:54:01.842536 ignition[1026]: parsing config with SHA512: a1eb05ad9192b210ae15381dc99e307a9b7b5006eeb399e6b7d4542f1017ae6c054fd18a44b589fa7179c1082aff8bae165b2196b7a16a469c704d33a2e9c426 Mar 2 12:54:01.845500 unknown[1026]: fetched base config from "system" Mar 2 12:54:01.845725 ignition[1026]: fetch: fetch complete Mar 2 12:54:01.845506 unknown[1026]: fetched base config from "system" Mar 2 12:54:01.845728 ignition[1026]: fetch: fetch passed Mar 2 12:54:01.845509 unknown[1026]: fetched user config from "azure" Mar 2 12:54:01.845768 ignition[1026]: Ignition finished successfully Mar 2 12:54:01.849642 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Mar 2 12:54:01.857070 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Mar 2 12:54:01.894282 ignition[1032]: Ignition 2.22.0 Mar 2 12:54:01.894293 ignition[1032]: Stage: kargs Mar 2 12:54:01.894559 ignition[1032]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:01.900128 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Mar 2 12:54:01.894567 ignition[1032]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:01.905767 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Mar 2 12:54:01.895062 ignition[1032]: kargs: kargs passed Mar 2 12:54:01.895101 ignition[1032]: Ignition finished successfully Mar 2 12:54:01.938206 ignition[1038]: Ignition 2.22.0 Mar 2 12:54:01.938225 ignition[1038]: Stage: disks Mar 2 12:54:01.941995 systemd[1]: Finished ignition-disks.service - Ignition (disks). Mar 2 12:54:01.938393 ignition[1038]: no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:01.947787 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Mar 2 12:54:01.938401 ignition[1038]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:01.955947 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Mar 2 12:54:01.938927 ignition[1038]: disks: disks passed Mar 2 12:54:01.963979 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 12:54:01.938963 ignition[1038]: Ignition finished successfully Mar 2 12:54:01.972037 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 12:54:01.980223 systemd[1]: Reached target basic.target - Basic System. Mar 2 12:54:01.989130 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Mar 2 12:54:02.112737 systemd-fsck[1046]: ROOT: clean, 15/7326000 files, 477845/7359488 blocks Mar 2 12:54:02.120795 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Mar 2 12:54:02.127317 systemd[1]: Mounting sysroot.mount - /sysroot... Mar 2 12:54:02.290668 systemd-networkd[1015]: eth0: Gained IPv6LL Mar 2 12:54:02.355426 kernel: EXT4-fs (sda9): mounted filesystem 6408ffd3-d563-490c-803b-1f4582ee0319 r/w with ordered data mode. Quota mode: none. Mar 2 12:54:02.355834 systemd[1]: Mounted sysroot.mount - /sysroot. Mar 2 12:54:02.359449 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Mar 2 12:54:02.382044 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 12:54:02.395914 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Mar 2 12:54:02.411424 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1060) Mar 2 12:54:02.412372 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Mar 2 12:54:02.426535 kernel: BTRFS info (device sda6): first mount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:54:02.426566 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:54:02.430883 kernel: BTRFS info (device sda6): turning on async discard Mar 2 12:54:02.437313 kernel: BTRFS info (device sda6): enabling free space tree Mar 2 12:54:02.437435 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 2 12:54:02.437542 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 12:54:02.453104 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 12:54:02.460074 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Mar 2 12:54:02.471178 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Mar 2 12:54:02.994400 coreos-metadata[1062]: Mar 02 12:54:02.994 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 2 12:54:03.001927 coreos-metadata[1062]: Mar 02 12:54:03.001 INFO Fetch successful Mar 2 12:54:03.005572 coreos-metadata[1062]: Mar 02 12:54:03.001 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Mar 2 12:54:03.014703 coreos-metadata[1062]: Mar 02 12:54:03.014 INFO Fetch successful Mar 2 12:54:03.037538 coreos-metadata[1062]: Mar 02 12:54:03.037 INFO wrote hostname ci-4459.2.101-f64d7d9b2a to /sysroot/etc/hostname Mar 2 12:54:03.044767 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 2 12:54:03.207135 initrd-setup-root[1090]: cut: /sysroot/etc/passwd: No such file or directory Mar 2 12:54:03.242206 initrd-setup-root[1097]: cut: /sysroot/etc/group: No such file or directory Mar 2 12:54:03.248937 initrd-setup-root[1104]: cut: /sysroot/etc/shadow: No such file or directory Mar 2 12:54:03.267927 initrd-setup-root[1111]: cut: /sysroot/etc/gshadow: No such file or directory Mar 2 12:54:04.264221 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Mar 2 12:54:04.269886 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Mar 2 12:54:04.286970 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Mar 2 12:54:04.301277 kernel: BTRFS info (device sda6): last unmount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:54:04.297040 systemd[1]: sysroot-oem.mount: Deactivated successfully. Mar 2 12:54:04.322476 ignition[1179]: INFO : Ignition 2.22.0 Mar 2 12:54:04.322476 ignition[1179]: INFO : Stage: mount Mar 2 12:54:04.332310 ignition[1179]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:04.332310 ignition[1179]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:04.332310 ignition[1179]: INFO : mount: mount passed Mar 2 12:54:04.332310 ignition[1179]: INFO : Ignition finished successfully Mar 2 12:54:04.325638 systemd[1]: Finished ignition-mount.service - Ignition (mount). Mar 2 12:54:04.331481 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Mar 2 12:54:04.337643 systemd[1]: Starting ignition-files.service - Ignition (files)... Mar 2 12:54:04.359500 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Mar 2 12:54:04.388422 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 (8:6) scanned by mount (1190) Mar 2 12:54:04.398337 kernel: BTRFS info (device sda6): first mount of filesystem fce10a79-d373-45d9-9854-55ae8d2c9f36 Mar 2 12:54:04.398346 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Mar 2 12:54:04.407117 kernel: BTRFS info (device sda6): turning on async discard Mar 2 12:54:04.407146 kernel: BTRFS info (device sda6): enabling free space tree Mar 2 12:54:04.408899 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Mar 2 12:54:04.443376 ignition[1208]: INFO : Ignition 2.22.0 Mar 2 12:54:04.443376 ignition[1208]: INFO : Stage: files Mar 2 12:54:04.449701 ignition[1208]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:04.449701 ignition[1208]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:04.449701 ignition[1208]: DEBUG : files: compiled without relabeling support, skipping Mar 2 12:54:04.482449 ignition[1208]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 2 12:54:04.482449 ignition[1208]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 2 12:54:04.543672 ignition[1208]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 2 12:54:04.549156 ignition[1208]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 2 12:54:04.549156 ignition[1208]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 2 12:54:04.544176 unknown[1208]: wrote ssh authorized keys file for user: core Mar 2 12:54:04.596834 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 2 12:54:04.604783 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.3-linux-arm64.tar.gz: attempt #1 Mar 2 12:54:04.641043 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 2 12:54:04.779377 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.3-linux-arm64.tar.gz" Mar 2 12:54:04.786961 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 2 12:54:04.786961 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Mar 2 12:54:05.003762 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 2 12:54:05.070181 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 12:54:05.077355 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 2 12:54:05.132811 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 12:54:05.132811 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 2 12:54:05.132811 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 2 12:54:05.132811 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 2 12:54:05.132811 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 2 12:54:05.132811 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://extensions.flatcar.org/extensions/kubernetes-v1.33.8-arm64.raw: attempt #1 Mar 2 12:54:05.620528 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 2 12:54:06.521162 ignition[1208]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.33.8-arm64.raw" Mar 2 12:54:06.521162 ignition[1208]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Mar 2 12:54:06.578995 ignition[1208]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 12:54:06.592521 ignition[1208]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 2 12:54:06.592521 ignition[1208]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Mar 2 12:54:06.606097 ignition[1208]: INFO : files: op(e): [started] setting preset to enabled for "prepare-helm.service" Mar 2 12:54:06.606097 ignition[1208]: INFO : files: op(e): [finished] setting preset to enabled for "prepare-helm.service" Mar 2 12:54:06.606097 ignition[1208]: INFO : files: createResultFile: createFiles: op(f): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 2 12:54:06.606097 ignition[1208]: INFO : files: createResultFile: createFiles: op(f): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 2 12:54:06.606097 ignition[1208]: INFO : files: files passed Mar 2 12:54:06.606097 ignition[1208]: INFO : Ignition finished successfully Mar 2 12:54:06.604178 systemd[1]: Finished ignition-files.service - Ignition (files). Mar 2 12:54:06.611257 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Mar 2 12:54:06.637831 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Mar 2 12:54:06.649624 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 2 12:54:06.659607 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Mar 2 12:54:06.685985 initrd-setup-root-after-ignition[1237]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 12:54:06.692593 initrd-setup-root-after-ignition[1237]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Mar 2 12:54:06.698944 initrd-setup-root-after-ignition[1241]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 2 12:54:06.692868 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 12:54:06.704193 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Mar 2 12:54:06.714879 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Mar 2 12:54:06.755650 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 2 12:54:06.755754 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Mar 2 12:54:06.765701 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Mar 2 12:54:06.774580 systemd[1]: Reached target initrd.target - Initrd Default Target. Mar 2 12:54:06.782466 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Mar 2 12:54:06.782988 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Mar 2 12:54:06.812228 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 12:54:06.818311 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Mar 2 12:54:06.846247 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Mar 2 12:54:06.851176 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 12:54:06.860256 systemd[1]: Stopped target timers.target - Timer Units. Mar 2 12:54:06.868623 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 2 12:54:06.868714 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Mar 2 12:54:06.880499 systemd[1]: Stopped target initrd.target - Initrd Default Target. Mar 2 12:54:06.885118 systemd[1]: Stopped target basic.target - Basic System. Mar 2 12:54:06.893090 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Mar 2 12:54:06.901354 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Mar 2 12:54:06.909350 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Mar 2 12:54:06.918091 systemd[1]: Stopped target initrd-usr-fs.target - Initrd /usr File System. Mar 2 12:54:06.926925 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Mar 2 12:54:06.935272 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Mar 2 12:54:06.944607 systemd[1]: Stopped target sysinit.target - System Initialization. Mar 2 12:54:06.952762 systemd[1]: Stopped target local-fs.target - Local File Systems. Mar 2 12:54:06.961757 systemd[1]: Stopped target swap.target - Swaps. Mar 2 12:54:06.968987 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 2 12:54:06.969095 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Mar 2 12:54:06.980333 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Mar 2 12:54:06.984801 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 12:54:06.993545 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Mar 2 12:54:06.997323 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 12:54:07.002382 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 2 12:54:07.002469 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Mar 2 12:54:07.015161 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 2 12:54:07.015240 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Mar 2 12:54:07.020152 systemd[1]: ignition-files.service: Deactivated successfully. Mar 2 12:54:07.020218 systemd[1]: Stopped ignition-files.service - Ignition (files). Mar 2 12:54:07.027772 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Mar 2 12:54:07.027833 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Mar 2 12:54:07.093528 ignition[1261]: INFO : Ignition 2.22.0 Mar 2 12:54:07.093528 ignition[1261]: INFO : Stage: umount Mar 2 12:54:07.093528 ignition[1261]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 2 12:54:07.093528 ignition[1261]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Mar 2 12:54:07.042559 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Mar 2 12:54:07.131452 ignition[1261]: INFO : umount: umount passed Mar 2 12:54:07.131452 ignition[1261]: INFO : Ignition finished successfully Mar 2 12:54:07.056258 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 2 12:54:07.056362 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 12:54:07.075551 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Mar 2 12:54:07.084917 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 2 12:54:07.085129 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 12:54:07.100140 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 2 12:54:07.100219 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Mar 2 12:54:07.109960 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 2 12:54:07.110039 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Mar 2 12:54:07.118841 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 2 12:54:07.119110 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Mar 2 12:54:07.128485 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 2 12:54:07.129458 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 2 12:54:07.129499 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Mar 2 12:54:07.135702 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 2 12:54:07.135737 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Mar 2 12:54:07.142851 systemd[1]: ignition-fetch.service: Deactivated successfully. Mar 2 12:54:07.142889 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Mar 2 12:54:07.150499 systemd[1]: Stopped target network.target - Network. Mar 2 12:54:07.157656 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 2 12:54:07.157694 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Mar 2 12:54:07.166961 systemd[1]: Stopped target paths.target - Path Units. Mar 2 12:54:07.174028 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 2 12:54:07.182421 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 12:54:07.191420 systemd[1]: Stopped target slices.target - Slice Units. Mar 2 12:54:07.199439 systemd[1]: Stopped target sockets.target - Socket Units. Mar 2 12:54:07.207120 systemd[1]: iscsid.socket: Deactivated successfully. Mar 2 12:54:07.207154 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Mar 2 12:54:07.215303 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 2 12:54:07.215324 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Mar 2 12:54:07.223395 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 2 12:54:07.223444 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Mar 2 12:54:07.232646 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Mar 2 12:54:07.232689 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Mar 2 12:54:07.240956 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Mar 2 12:54:07.249652 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Mar 2 12:54:07.261588 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 2 12:54:07.261727 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Mar 2 12:54:07.274574 systemd[1]: run-credentials-systemd\x2dresolved.service.mount: Deactivated successfully. Mar 2 12:54:07.274784 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 2 12:54:07.274866 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Mar 2 12:54:07.285330 systemd[1]: run-credentials-systemd\x2dnetworkd.service.mount: Deactivated successfully. Mar 2 12:54:07.285750 systemd[1]: Stopped target network-pre.target - Preparation for Network. Mar 2 12:54:07.470558 kernel: hv_netvsc 7ced8d88-5e8d-7ced-8d88-5e8d7ced8d88 eth0: Data path switched from VF: enP21044s1 Mar 2 12:54:07.293998 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 2 12:54:07.294033 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Mar 2 12:54:07.303307 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Mar 2 12:54:07.311999 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 2 12:54:07.312047 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Mar 2 12:54:07.321039 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 12:54:07.321085 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:54:07.332607 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 2 12:54:07.332642 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Mar 2 12:54:07.344366 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Mar 2 12:54:07.344418 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 12:54:07.356361 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 12:54:07.364205 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 2 12:54:07.364252 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:54:07.366880 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 2 12:54:07.366953 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Mar 2 12:54:07.379623 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 2 12:54:07.379723 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 12:54:07.388702 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 2 12:54:07.388769 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Mar 2 12:54:07.399843 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 2 12:54:07.399868 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 12:54:07.409004 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 2 12:54:07.409048 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Mar 2 12:54:07.420896 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 2 12:54:07.420940 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Mar 2 12:54:07.433854 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 2 12:54:07.433896 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Mar 2 12:54:07.442568 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 2 12:54:07.442618 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Mar 2 12:54:07.451523 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Mar 2 12:54:07.465825 systemd[1]: systemd-network-generator.service: Deactivated successfully. Mar 2 12:54:07.465893 systemd[1]: Stopped systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 12:54:07.478963 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 2 12:54:07.478999 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 12:54:07.488275 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:54:07.488314 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:54:07.501678 systemd[1]: run-credentials-systemd\x2dnetwork\x2dgenerator.service.mount: Deactivated successfully. Mar 2 12:54:07.501720 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 2 12:54:07.501746 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:54:07.502032 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 2 12:54:07.502112 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Mar 2 12:54:07.700361 systemd-journald[225]: Received SIGTERM from PID 1 (systemd). Mar 2 12:54:07.577999 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 2 12:54:07.578160 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Mar 2 12:54:07.586125 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Mar 2 12:54:07.594838 systemd[1]: Starting initrd-switch-root.service - Switch Root... Mar 2 12:54:07.613527 systemd[1]: Switching root. Mar 2 12:54:07.718035 systemd-journald[225]: Journal stopped Mar 2 12:54:12.571011 kernel: SELinux: policy capability network_peer_controls=1 Mar 2 12:54:12.571028 kernel: SELinux: policy capability open_perms=1 Mar 2 12:54:12.571035 kernel: SELinux: policy capability extended_socket_class=1 Mar 2 12:54:12.571040 kernel: SELinux: policy capability always_check_network=0 Mar 2 12:54:12.571045 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 2 12:54:12.571051 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 2 12:54:12.571057 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 2 12:54:12.571063 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 2 12:54:12.571068 kernel: SELinux: policy capability userspace_initial_context=0 Mar 2 12:54:12.571073 kernel: audit: type=1403 audit(1772456049.010:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 2 12:54:12.571080 systemd[1]: Successfully loaded SELinux policy in 167.647ms. Mar 2 12:54:12.571087 systemd[1]: Relabeled /dev/, /dev/shm/, /run/ in 4.320ms. Mar 2 12:54:12.571094 systemd[1]: systemd 256.8 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP -GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBCRYPTSETUP_PLUGINS +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT +LIBARCHIVE) Mar 2 12:54:12.571100 systemd[1]: Detected virtualization microsoft. Mar 2 12:54:12.571106 systemd[1]: Detected architecture arm64. Mar 2 12:54:12.571112 systemd[1]: Detected first boot. Mar 2 12:54:12.571119 systemd[1]: Hostname set to . Mar 2 12:54:12.571125 systemd[1]: Initializing machine ID from random generator. Mar 2 12:54:12.571131 zram_generator::config[1305]: No configuration found. Mar 2 12:54:12.571137 kernel: NET: Registered PF_VSOCK protocol family Mar 2 12:54:12.571143 systemd[1]: Populated /etc with preset unit settings. Mar 2 12:54:12.571149 systemd[1]: run-credentials-systemd\x2djournald.service.mount: Deactivated successfully. Mar 2 12:54:12.571155 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 2 12:54:12.571163 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Mar 2 12:54:12.571169 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 2 12:54:12.571175 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Mar 2 12:54:12.571181 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Mar 2 12:54:12.571187 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Mar 2 12:54:12.571193 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Mar 2 12:54:12.571199 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Mar 2 12:54:12.571206 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Mar 2 12:54:12.571212 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Mar 2 12:54:12.571218 systemd[1]: Created slice user.slice - User and Session Slice. Mar 2 12:54:12.571224 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Mar 2 12:54:12.571230 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Mar 2 12:54:12.571236 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Mar 2 12:54:12.571242 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Mar 2 12:54:12.571248 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Mar 2 12:54:12.571255 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Mar 2 12:54:12.571261 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Mar 2 12:54:12.571268 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Mar 2 12:54:12.571274 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Mar 2 12:54:12.571281 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Mar 2 12:54:12.571287 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Mar 2 12:54:12.571294 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Mar 2 12:54:12.571300 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Mar 2 12:54:12.571307 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Mar 2 12:54:12.571313 systemd[1]: Reached target remote-fs.target - Remote File Systems. Mar 2 12:54:12.571319 systemd[1]: Reached target slices.target - Slice Units. Mar 2 12:54:12.571325 systemd[1]: Reached target swap.target - Swaps. Mar 2 12:54:12.571331 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Mar 2 12:54:12.571337 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Mar 2 12:54:12.571345 systemd[1]: Listening on systemd-creds.socket - Credential Encryption/Decryption. Mar 2 12:54:12.571351 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Mar 2 12:54:12.571357 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Mar 2 12:54:12.571363 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Mar 2 12:54:12.571369 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Mar 2 12:54:12.571375 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Mar 2 12:54:12.571381 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Mar 2 12:54:12.571388 systemd[1]: Mounting media.mount - External Media Directory... Mar 2 12:54:12.571394 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Mar 2 12:54:12.571401 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Mar 2 12:54:12.571424 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Mar 2 12:54:12.571431 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 2 12:54:12.571437 systemd[1]: Reached target machines.target - Containers. Mar 2 12:54:12.571444 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Mar 2 12:54:12.571450 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:54:12.571458 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Mar 2 12:54:12.571464 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Mar 2 12:54:12.571470 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:54:12.571476 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 12:54:12.571482 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:54:12.571488 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Mar 2 12:54:12.571495 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:54:12.571501 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 2 12:54:12.571507 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 2 12:54:12.571514 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Mar 2 12:54:12.571520 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 2 12:54:12.571527 systemd[1]: Stopped systemd-fsck-usr.service. Mar 2 12:54:12.571533 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:54:12.571539 systemd[1]: Starting systemd-journald.service - Journal Service... Mar 2 12:54:12.571545 kernel: fuse: init (API version 7.41) Mar 2 12:54:12.571551 kernel: loop: module loaded Mar 2 12:54:12.571556 kernel: ACPI: bus type drm_connector registered Mar 2 12:54:12.571575 systemd-journald[1409]: Collecting audit messages is disabled. Mar 2 12:54:12.571589 systemd-journald[1409]: Journal started Mar 2 12:54:12.571605 systemd-journald[1409]: Runtime Journal (/run/log/journal/78d8d773026e485fa3e0fd9b25fc7bb7) is 8M, max 78.3M, 70.3M free. Mar 2 12:54:11.855580 systemd[1]: Queued start job for default target multi-user.target. Mar 2 12:54:11.866847 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Mar 2 12:54:11.867214 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 2 12:54:11.867484 systemd[1]: systemd-journald.service: Consumed 2.389s CPU time. Mar 2 12:54:12.581883 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Mar 2 12:54:12.598703 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Mar 2 12:54:12.610919 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Mar 2 12:54:12.626155 systemd[1]: Starting systemd-udev-load-credentials.service - Load udev Rules from Credentials... Mar 2 12:54:12.636961 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Mar 2 12:54:12.644620 systemd[1]: verity-setup.service: Deactivated successfully. Mar 2 12:54:12.644745 systemd[1]: Stopped verity-setup.service. Mar 2 12:54:12.659964 systemd[1]: Started systemd-journald.service - Journal Service. Mar 2 12:54:12.660660 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Mar 2 12:54:12.666199 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Mar 2 12:54:12.670918 systemd[1]: Mounted media.mount - External Media Directory. Mar 2 12:54:12.674823 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Mar 2 12:54:12.679187 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Mar 2 12:54:12.684024 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Mar 2 12:54:12.688360 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Mar 2 12:54:12.693249 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Mar 2 12:54:12.698923 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 2 12:54:12.700439 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Mar 2 12:54:12.705295 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:54:12.705437 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:54:12.710040 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 12:54:12.710153 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 12:54:12.715292 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:54:12.715573 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:54:12.720746 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 2 12:54:12.720886 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Mar 2 12:54:12.725458 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:54:12.725588 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:54:12.731588 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Mar 2 12:54:12.736342 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Mar 2 12:54:12.742004 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Mar 2 12:54:12.747535 systemd[1]: Finished systemd-udev-load-credentials.service - Load udev Rules from Credentials. Mar 2 12:54:12.753127 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Mar 2 12:54:12.766819 systemd[1]: Reached target network-pre.target - Preparation for Network. Mar 2 12:54:12.772480 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Mar 2 12:54:12.785493 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Mar 2 12:54:12.790170 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 2 12:54:12.790200 systemd[1]: Reached target local-fs.target - Local File Systems. Mar 2 12:54:12.795188 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management. Mar 2 12:54:12.801171 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Mar 2 12:54:12.805272 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:54:12.823544 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Mar 2 12:54:12.830591 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Mar 2 12:54:12.835451 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 12:54:12.836171 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Mar 2 12:54:12.841522 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 12:54:12.847320 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 12:54:12.854415 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Mar 2 12:54:12.860546 systemd[1]: Starting systemd-sysusers.service - Create System Users... Mar 2 12:54:12.867701 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Mar 2 12:54:12.874657 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Mar 2 12:54:12.882987 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Mar 2 12:54:12.891175 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Mar 2 12:54:12.897345 systemd[1]: Starting systemd-machine-id-commit.service - Save Transient machine-id to Disk... Mar 2 12:54:12.912427 kernel: loop0: detected capacity change from 0 to 119840 Mar 2 12:54:12.918664 systemd-journald[1409]: Time spent on flushing to /var/log/journal/78d8d773026e485fa3e0fd9b25fc7bb7 is 65.580ms for 939 entries. Mar 2 12:54:12.918664 systemd-journald[1409]: System Journal (/var/log/journal/78d8d773026e485fa3e0fd9b25fc7bb7) is 11.8M, max 2.6G, 2.6G free. Mar 2 12:54:13.046788 systemd-journald[1409]: Received client request to flush runtime journal. Mar 2 12:54:13.046847 systemd-journald[1409]: /var/log/journal/78d8d773026e485fa3e0fd9b25fc7bb7/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Mar 2 12:54:13.046865 systemd-journald[1409]: Rotating system journal. Mar 2 12:54:12.961653 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:54:13.015422 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 2 12:54:13.017080 systemd[1]: Finished systemd-machine-id-commit.service - Save Transient machine-id to Disk. Mar 2 12:54:13.049441 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Mar 2 12:54:13.064681 systemd[1]: Finished systemd-sysusers.service - Create System Users. Mar 2 12:54:13.070587 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Mar 2 12:54:13.212748 systemd-tmpfiles[1460]: ACLs are not supported, ignoring. Mar 2 12:54:13.212760 systemd-tmpfiles[1460]: ACLs are not supported, ignoring. Mar 2 12:54:13.216130 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Mar 2 12:54:13.364578 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 2 12:54:13.367729 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Mar 2 12:54:13.373842 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Mar 2 12:54:13.397423 kernel: loop1: detected capacity change from 0 to 100632 Mar 2 12:54:13.401837 systemd-udevd[1466]: Using default interface naming scheme 'v255'. Mar 2 12:54:13.621261 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Mar 2 12:54:13.632843 systemd[1]: Starting systemd-networkd.service - Network Configuration... Mar 2 12:54:13.691475 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Mar 2 12:54:13.719020 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Mar 2 12:54:13.750128 systemd[1]: Started systemd-userdbd.service - User Database Manager. Mar 2 12:54:13.778463 kernel: mousedev: PS/2 mouse device common for all mice Mar 2 12:54:13.802902 kernel: hv_vmbus: registering driver hv_balloon Mar 2 12:54:13.802977 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Mar 2 12:54:13.806103 kernel: hv_balloon: Memory hot add disabled on ARM64 Mar 2 12:54:13.835535 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#244 cmd 0x85 status: scsi 0x2 srb 0x6 hv 0xc0000001 Mar 2 12:54:13.868274 systemd-networkd[1485]: lo: Link UP Mar 2 12:54:13.868279 systemd-networkd[1485]: lo: Gained carrier Mar 2 12:54:13.869170 systemd-networkd[1485]: Enumeration completed Mar 2 12:54:13.869396 systemd-networkd[1485]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:54:13.869399 systemd-networkd[1485]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 12:54:13.870463 systemd[1]: Started systemd-networkd.service - Network Configuration. Mar 2 12:54:13.878591 kernel: hv_vmbus: registering driver hyperv_fb Mar 2 12:54:13.878651 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Mar 2 12:54:13.878575 systemd[1]: Starting systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd... Mar 2 12:54:13.891719 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Mar 2 12:54:13.894773 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Mar 2 12:54:13.902535 kernel: Console: switching to colour dummy device 80x25 Mar 2 12:54:13.906425 kernel: Console: switching to colour frame buffer device 128x48 Mar 2 12:54:13.936287 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:54:13.947423 kernel: mlx5_core 5234:00:02.0 enP21044s1: Link up Mar 2 12:54:13.948163 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:54:13.949537 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:54:13.960521 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:54:13.964425 kernel: loop2: detected capacity change from 0 to 209336 Mar 2 12:54:13.977189 kernel: hv_netvsc 7ced8d88-5e8d-7ced-8d88-5e8d7ced8d88 eth0: Data path switched to VF: enP21044s1 Mar 2 12:54:13.973146 systemd-networkd[1485]: enP21044s1: Link UP Mar 2 12:54:13.973242 systemd-networkd[1485]: eth0: Link UP Mar 2 12:54:13.973244 systemd-networkd[1485]: eth0: Gained carrier Mar 2 12:54:13.973263 systemd-networkd[1485]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:54:13.976811 systemd-networkd[1485]: enP21044s1: Gained carrier Mar 2 12:54:13.980851 systemd[1]: Finished systemd-networkd-persistent-storage.service - Enable Persistent Storage in systemd-networkd. Mar 2 12:54:13.988522 systemd-networkd[1485]: eth0: DHCPv4 address 10.200.20.30/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 2 12:54:13.990331 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 2 12:54:13.990518 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:54:13.995909 systemd[1]: run-credentials-systemd\x2dvconsole\x2dsetup.service.mount: Deactivated successfully. Mar 2 12:54:13.996876 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Mar 2 12:54:14.024514 kernel: MACsec IEEE 802.1AE Mar 2 12:54:14.051453 kernel: loop3: detected capacity change from 0 to 27936 Mar 2 12:54:14.091336 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Mar 2 12:54:14.097605 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Mar 2 12:54:14.162482 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Mar 2 12:54:14.338215 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Mar 2 12:54:14.467433 kernel: loop4: detected capacity change from 0 to 119840 Mar 2 12:54:14.480429 kernel: loop5: detected capacity change from 0 to 100632 Mar 2 12:54:14.492435 kernel: loop6: detected capacity change from 0 to 209336 Mar 2 12:54:14.514439 kernel: loop7: detected capacity change from 0 to 27936 Mar 2 12:54:14.522506 (sd-merge)[1618]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Mar 2 12:54:14.522914 (sd-merge)[1618]: Merged extensions into '/usr'. Mar 2 12:54:14.526953 systemd[1]: Reload requested from client PID 1445 ('systemd-sysext') (unit systemd-sysext.service)... Mar 2 12:54:14.527068 systemd[1]: Reloading... Mar 2 12:54:14.580439 zram_generator::config[1647]: No configuration found. Mar 2 12:54:14.743703 systemd[1]: Reloading finished in 216 ms. Mar 2 12:54:14.760176 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Mar 2 12:54:14.779231 systemd[1]: Starting ensure-sysext.service... Mar 2 12:54:14.785519 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Mar 2 12:54:14.797630 systemd[1]: Reload requested from client PID 1702 ('systemctl') (unit ensure-sysext.service)... Mar 2 12:54:14.797721 systemd[1]: Reloading... Mar 2 12:54:14.797911 systemd-tmpfiles[1703]: /usr/lib/tmpfiles.d/nfs-utils.conf:6: Duplicate line for path "/var/lib/nfs/sm", ignoring. Mar 2 12:54:14.797928 systemd-tmpfiles[1703]: /usr/lib/tmpfiles.d/nfs-utils.conf:7: Duplicate line for path "/var/lib/nfs/sm.bak", ignoring. Mar 2 12:54:14.798066 systemd-tmpfiles[1703]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 2 12:54:14.798196 systemd-tmpfiles[1703]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Mar 2 12:54:14.798608 systemd-tmpfiles[1703]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 2 12:54:14.798741 systemd-tmpfiles[1703]: ACLs are not supported, ignoring. Mar 2 12:54:14.798768 systemd-tmpfiles[1703]: ACLs are not supported, ignoring. Mar 2 12:54:14.802903 systemd-tmpfiles[1703]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 12:54:14.802914 systemd-tmpfiles[1703]: Skipping /boot Mar 2 12:54:14.809160 systemd-tmpfiles[1703]: Detected autofs mount point /boot during canonicalization of boot. Mar 2 12:54:14.809174 systemd-tmpfiles[1703]: Skipping /boot Mar 2 12:54:14.857544 zram_generator::config[1737]: No configuration found. Mar 2 12:54:15.003986 systemd[1]: Reloading finished in 206 ms. Mar 2 12:54:15.019331 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Mar 2 12:54:15.038231 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 12:54:15.049182 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Mar 2 12:54:15.056157 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Mar 2 12:54:15.063625 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Mar 2 12:54:15.071537 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Mar 2 12:54:15.078705 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:54:15.081632 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:54:15.090788 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:54:15.100835 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:54:15.105958 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:54:15.106100 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:54:15.108379 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:54:15.108632 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:54:15.114237 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:54:15.114372 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:54:15.120469 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:54:15.120593 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:54:15.133304 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Mar 2 12:54:15.135578 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Mar 2 12:54:15.143805 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Mar 2 12:54:15.152162 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Mar 2 12:54:15.161700 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Mar 2 12:54:15.165970 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Mar 2 12:54:15.166471 systemd[1]: systemd-hibernate-clear.service - Clear Stale Hibernate Storage Info was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/HibernateLocation-8cf2644b-4b0b-428f-9387-6d876050dc67). Mar 2 12:54:15.166996 systemd[1]: Reached target time-set.target - System Time Set. Mar 2 12:54:15.175582 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Mar 2 12:54:15.183202 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Mar 2 12:54:15.189844 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 2 12:54:15.190046 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Mar 2 12:54:15.190532 systemd-resolved[1794]: Positive Trust Anchors: Mar 2 12:54:15.190772 systemd-resolved[1794]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 2 12:54:15.190796 systemd-resolved[1794]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Mar 2 12:54:15.195659 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 2 12:54:15.195779 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Mar 2 12:54:15.200274 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 2 12:54:15.200397 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Mar 2 12:54:15.205867 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 2 12:54:15.206004 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Mar 2 12:54:15.214448 systemd[1]: Finished ensure-sysext.service. Mar 2 12:54:15.219869 systemd-resolved[1794]: Using system hostname 'ci-4459.2.101-f64d7d9b2a'. Mar 2 12:54:15.221858 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Mar 2 12:54:15.227179 systemd[1]: Reached target network.target - Network. Mar 2 12:54:15.230770 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Mar 2 12:54:15.235456 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 2 12:54:15.235505 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Mar 2 12:54:15.271163 augenrules[1833]: No rules Mar 2 12:54:15.272218 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 12:54:15.272415 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 12:54:15.474520 systemd-networkd[1485]: eth0: Gained IPv6LL Mar 2 12:54:15.479232 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Mar 2 12:54:15.485082 systemd[1]: Reached target network-online.target - Network is Online. Mar 2 12:54:15.727139 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Mar 2 12:54:15.732611 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 2 12:54:18.117712 ldconfig[1439]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 2 12:54:18.130720 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Mar 2 12:54:18.137024 systemd[1]: Starting systemd-update-done.service - Update is Completed... Mar 2 12:54:18.150092 systemd[1]: Finished systemd-update-done.service - Update is Completed. Mar 2 12:54:18.154902 systemd[1]: Reached target sysinit.target - System Initialization. Mar 2 12:54:18.159260 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Mar 2 12:54:18.164216 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Mar 2 12:54:18.169739 systemd[1]: Started logrotate.timer - Daily rotation of log files. Mar 2 12:54:18.174031 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Mar 2 12:54:18.179021 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Mar 2 12:54:18.184114 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 2 12:54:18.184142 systemd[1]: Reached target paths.target - Path Units. Mar 2 12:54:18.187803 systemd[1]: Reached target timers.target - Timer Units. Mar 2 12:54:18.193105 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Mar 2 12:54:18.198926 systemd[1]: Starting docker.socket - Docker Socket for the API... Mar 2 12:54:18.204652 systemd[1]: Listening on sshd-unix-local.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_UNIX Local). Mar 2 12:54:18.209776 systemd[1]: Listening on sshd-vsock.socket - OpenSSH Server Socket (systemd-ssh-generator, AF_VSOCK). Mar 2 12:54:18.214708 systemd[1]: Reached target ssh-access.target - SSH Access Available. Mar 2 12:54:18.220881 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Mar 2 12:54:18.225165 systemd[1]: Listening on systemd-hostnamed.socket - Hostname Service Socket. Mar 2 12:54:18.230340 systemd[1]: Listening on docker.socket - Docker Socket for the API. Mar 2 12:54:18.234871 systemd[1]: Reached target sockets.target - Socket Units. Mar 2 12:54:18.238629 systemd[1]: Reached target basic.target - Basic System. Mar 2 12:54:18.242590 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Mar 2 12:54:18.242611 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Mar 2 12:54:18.244763 systemd[1]: Starting chronyd.service - NTP client/server... Mar 2 12:54:18.258498 systemd[1]: Starting containerd.service - containerd container runtime... Mar 2 12:54:18.269615 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Mar 2 12:54:18.278532 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Mar 2 12:54:18.285998 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Mar 2 12:54:18.318397 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Mar 2 12:54:18.325350 chronyd[1846]: chronyd version 4.7 starting (+CMDMON +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +NTS +SECHASH +IPV6 -DEBUG) Mar 2 12:54:18.325525 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Mar 2 12:54:18.329568 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Mar 2 12:54:18.335393 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Mar 2 12:54:18.341098 KVP[1856]: KVP starting; pid is:1856 Mar 2 12:54:18.341703 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Mar 2 12:54:18.342901 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:54:18.349260 jq[1854]: false Mar 2 12:54:18.349442 kernel: hv_utils: KVP IC version 4.0 Mar 2 12:54:18.351515 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Mar 2 12:54:18.351516 KVP[1856]: KVP LIC Version: 3.1 Mar 2 12:54:18.359600 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Mar 2 12:54:18.366320 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Mar 2 12:54:18.370653 chronyd[1846]: Timezone right/UTC failed leap second check, ignoring Mar 2 12:54:18.370937 chronyd[1846]: Loaded seccomp filter (level 2) Mar 2 12:54:18.373564 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Mar 2 12:54:18.380289 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Mar 2 12:54:18.387874 systemd[1]: Starting systemd-logind.service - User Login Management... Mar 2 12:54:18.392615 extend-filesystems[1855]: Found /dev/sda6 Mar 2 12:54:18.395060 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 2 12:54:18.400914 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 2 12:54:18.401610 systemd[1]: Starting update-engine.service - Update Engine... Mar 2 12:54:18.407830 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Mar 2 12:54:18.412546 extend-filesystems[1855]: Found /dev/sda9 Mar 2 12:54:18.412546 extend-filesystems[1855]: Checking size of /dev/sda9 Mar 2 12:54:18.414738 systemd[1]: Started chronyd.service - NTP client/server. Mar 2 12:54:18.423639 jq[1872]: true Mar 2 12:54:18.429464 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Mar 2 12:54:18.436870 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 2 12:54:18.440616 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Mar 2 12:54:18.442456 systemd[1]: motdgen.service: Deactivated successfully. Mar 2 12:54:18.442638 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Mar 2 12:54:18.449546 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 2 12:54:18.450476 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Mar 2 12:54:18.462984 extend-filesystems[1855]: Old size kept for /dev/sda9 Mar 2 12:54:18.471505 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 2 12:54:18.472543 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Mar 2 12:54:18.476134 update_engine[1871]: I20260302 12:54:18.476065 1871 main.cc:92] Flatcar Update Engine starting Mar 2 12:54:18.479497 jq[1886]: true Mar 2 12:54:18.490648 (ntainerd)[1890]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Mar 2 12:54:18.518681 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Mar 2 12:54:18.527166 tar[1885]: linux-arm64/LICENSE Mar 2 12:54:18.527166 tar[1885]: linux-arm64/helm Mar 2 12:54:18.531024 systemd-logind[1868]: New seat seat0. Mar 2 12:54:18.535032 systemd-logind[1868]: Watching system buttons on /dev/input/event1 (AT Translated Set 2 keyboard) Mar 2 12:54:18.535205 systemd[1]: Started systemd-logind.service - User Login Management. Mar 2 12:54:18.606794 bash[1919]: Updated "/home/core/.ssh/authorized_keys" Mar 2 12:54:18.610396 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Mar 2 12:54:18.622864 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Mar 2 12:54:18.640183 dbus-daemon[1849]: [system] SELinux support is enabled Mar 2 12:54:18.640562 systemd[1]: Started dbus.service - D-Bus System Message Bus. Mar 2 12:54:18.648198 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 2 12:54:18.648229 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Mar 2 12:54:18.654823 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 2 12:54:18.655231 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Mar 2 12:54:18.660267 update_engine[1871]: I20260302 12:54:18.660132 1871 update_check_scheduler.cc:74] Next update check in 8m40s Mar 2 12:54:18.664924 systemd[1]: Started update-engine.service - Update Engine. Mar 2 12:54:18.665192 dbus-daemon[1849]: [system] Successfully activated service 'org.freedesktop.systemd1' Mar 2 12:54:18.682465 systemd[1]: Started locksmithd.service - Cluster reboot manager. Mar 2 12:54:18.751280 coreos-metadata[1848]: Mar 02 12:54:18.750 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Mar 2 12:54:18.751280 coreos-metadata[1848]: Mar 02 12:54:18.750 INFO Fetch successful Mar 2 12:54:18.751280 coreos-metadata[1848]: Mar 02 12:54:18.750 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Mar 2 12:54:18.751280 coreos-metadata[1848]: Mar 02 12:54:18.750 INFO Fetch successful Mar 2 12:54:18.751280 coreos-metadata[1848]: Mar 02 12:54:18.750 INFO Fetching http://168.63.129.16/machine/fc44bbaf-c808-40cb-8aa7-e9cb8df0838c/335b77e5%2D8db0%2D4e71%2D94dc%2D89fb27d7e522.%5Fci%2D4459.2.101%2Df64d7d9b2a?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Mar 2 12:54:18.754091 coreos-metadata[1848]: Mar 02 12:54:18.754 INFO Fetch successful Mar 2 12:54:18.754091 coreos-metadata[1848]: Mar 02 12:54:18.754 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Mar 2 12:54:18.764929 coreos-metadata[1848]: Mar 02 12:54:18.764 INFO Fetch successful Mar 2 12:54:18.815873 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Mar 2 12:54:18.823091 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Mar 2 12:54:18.945545 sshd_keygen[1880]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 2 12:54:18.963439 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Mar 2 12:54:18.974584 systemd[1]: Starting issuegen.service - Generate /run/issue... Mar 2 12:54:18.986939 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Mar 2 12:54:18.993602 locksmithd[1968]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 2 12:54:19.003542 systemd[1]: issuegen.service: Deactivated successfully. Mar 2 12:54:19.003721 systemd[1]: Finished issuegen.service - Generate /run/issue. Mar 2 12:54:19.017647 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Mar 2 12:54:19.038759 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Mar 2 12:54:19.046896 tar[1885]: linux-arm64/README.md Mar 2 12:54:19.049565 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Mar 2 12:54:19.060758 systemd[1]: Started getty@tty1.service - Getty on tty1. Mar 2 12:54:19.073984 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Mar 2 12:54:19.081642 systemd[1]: Reached target getty.target - Login Prompts. Mar 2 12:54:19.086270 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Mar 2 12:54:19.182423 containerd[1890]: time="2026-03-02T12:54:19Z" level=warning msg="Ignoring unknown key in TOML" column=1 error="strict mode: fields in the document are missing in the target struct" file=/usr/share/containerd/config.toml key=subreaper row=8 Mar 2 12:54:19.183427 containerd[1890]: time="2026-03-02T12:54:19.183383172Z" level=info msg="starting containerd" revision=4ac6c20c7bbf8177f29e46bbdc658fec02ffb8ad version=v2.0.7 Mar 2 12:54:19.189311 containerd[1890]: time="2026-03-02T12:54:19.189281052Z" level=warning msg="Configuration migrated from version 2, use `containerd config migrate` to avoid migration" t="6.928µs" Mar 2 12:54:19.189671 containerd[1890]: time="2026-03-02T12:54:19.189540964Z" level=info msg="loading plugin" id=io.containerd.image-verifier.v1.bindir type=io.containerd.image-verifier.v1 Mar 2 12:54:19.189713 containerd[1890]: time="2026-03-02T12:54:19.189678524Z" level=info msg="loading plugin" id=io.containerd.internal.v1.opt type=io.containerd.internal.v1 Mar 2 12:54:19.190500 containerd[1890]: time="2026-03-02T12:54:19.190451996Z" level=info msg="loading plugin" id=io.containerd.warning.v1.deprecations type=io.containerd.warning.v1 Mar 2 12:54:19.190500 containerd[1890]: time="2026-03-02T12:54:19.190474156Z" level=info msg="loading plugin" id=io.containerd.content.v1.content type=io.containerd.content.v1 Mar 2 12:54:19.190500 containerd[1890]: time="2026-03-02T12:54:19.190492084Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 12:54:19.190585 containerd[1890]: time="2026-03-02T12:54:19.190529876Z" level=info msg="skip loading plugin" error="no scratch file generator: skip plugin" id=io.containerd.snapshotter.v1.blockfile type=io.containerd.snapshotter.v1 Mar 2 12:54:19.190585 containerd[1890]: time="2026-03-02T12:54:19.190537036Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 12:54:19.191185 containerd[1890]: time="2026-03-02T12:54:19.191159332Z" level=info msg="skip loading plugin" error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" id=io.containerd.snapshotter.v1.btrfs type=io.containerd.snapshotter.v1 Mar 2 12:54:19.191185 containerd[1890]: time="2026-03-02T12:54:19.191182020Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 12:54:19.191221 containerd[1890]: time="2026-03-02T12:54:19.191193556Z" level=info msg="skip loading plugin" error="devmapper not configured: skip plugin" id=io.containerd.snapshotter.v1.devmapper type=io.containerd.snapshotter.v1 Mar 2 12:54:19.191221 containerd[1890]: time="2026-03-02T12:54:19.191199796Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.native type=io.containerd.snapshotter.v1 Mar 2 12:54:19.191295 containerd[1890]: time="2026-03-02T12:54:19.191279460Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.overlayfs type=io.containerd.snapshotter.v1 Mar 2 12:54:19.192382 containerd[1890]: time="2026-03-02T12:54:19.192357556Z" level=info msg="loading plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 12:54:19.192428 containerd[1890]: time="2026-03-02T12:54:19.192396868Z" level=info msg="skip loading plugin" error="lstat /var/lib/containerd/io.containerd.snapshotter.v1.zfs: no such file or directory: skip plugin" id=io.containerd.snapshotter.v1.zfs type=io.containerd.snapshotter.v1 Mar 2 12:54:19.192428 containerd[1890]: time="2026-03-02T12:54:19.192425452Z" level=info msg="loading plugin" id=io.containerd.event.v1.exchange type=io.containerd.event.v1 Mar 2 12:54:19.192461 containerd[1890]: time="2026-03-02T12:54:19.192448388Z" level=info msg="loading plugin" id=io.containerd.monitor.task.v1.cgroups type=io.containerd.monitor.task.v1 Mar 2 12:54:19.192830 containerd[1890]: time="2026-03-02T12:54:19.192805908Z" level=info msg="loading plugin" id=io.containerd.metadata.v1.bolt type=io.containerd.metadata.v1 Mar 2 12:54:19.192900 containerd[1890]: time="2026-03-02T12:54:19.192883220Z" level=info msg="metadata content store policy set" policy=shared Mar 2 12:54:19.208758 containerd[1890]: time="2026-03-02T12:54:19.208681812Z" level=info msg="loading plugin" id=io.containerd.gc.v1.scheduler type=io.containerd.gc.v1 Mar 2 12:54:19.209272 containerd[1890]: time="2026-03-02T12:54:19.209248932Z" level=info msg="loading plugin" id=io.containerd.differ.v1.walking type=io.containerd.differ.v1 Mar 2 12:54:19.209355 containerd[1890]: time="2026-03-02T12:54:19.209342180Z" level=info msg="loading plugin" id=io.containerd.lease.v1.manager type=io.containerd.lease.v1 Mar 2 12:54:19.209428 containerd[1890]: time="2026-03-02T12:54:19.209390788Z" level=info msg="loading plugin" id=io.containerd.service.v1.containers-service type=io.containerd.service.v1 Mar 2 12:54:19.209479 containerd[1890]: time="2026-03-02T12:54:19.209467412Z" level=info msg="loading plugin" id=io.containerd.service.v1.content-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209568364Z" level=info msg="loading plugin" id=io.containerd.service.v1.diff-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209586188Z" level=info msg="loading plugin" id=io.containerd.service.v1.images-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209595300Z" level=info msg="loading plugin" id=io.containerd.service.v1.introspection-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209603564Z" level=info msg="loading plugin" id=io.containerd.service.v1.namespaces-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209610300Z" level=info msg="loading plugin" id=io.containerd.service.v1.snapshots-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209617668Z" level=info msg="loading plugin" id=io.containerd.shim.v1.manager type=io.containerd.shim.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209627108Z" level=info msg="loading plugin" id=io.containerd.runtime.v2.task type=io.containerd.runtime.v2 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209738876Z" level=info msg="loading plugin" id=io.containerd.service.v1.tasks-service type=io.containerd.service.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209755028Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.containers type=io.containerd.grpc.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209768004Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.content type=io.containerd.grpc.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209775268Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.diff type=io.containerd.grpc.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209781884Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.events type=io.containerd.grpc.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209789316Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.images type=io.containerd.grpc.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209797044Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.introspection type=io.containerd.grpc.v1 Mar 2 12:54:19.211246 containerd[1890]: time="2026-03-02T12:54:19.209804348Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.leases type=io.containerd.grpc.v1 Mar 2 12:54:19.211492 containerd[1890]: time="2026-03-02T12:54:19.209812044Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.namespaces type=io.containerd.grpc.v1 Mar 2 12:54:19.211492 containerd[1890]: time="2026-03-02T12:54:19.209818828Z" level=info msg="loading plugin" id=io.containerd.sandbox.store.v1.local type=io.containerd.sandbox.store.v1 Mar 2 12:54:19.211492 containerd[1890]: time="2026-03-02T12:54:19.209827340Z" level=info msg="loading plugin" id=io.containerd.cri.v1.images type=io.containerd.cri.v1 Mar 2 12:54:19.211492 containerd[1890]: time="2026-03-02T12:54:19.209868188Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\" for snapshotter \"overlayfs\"" Mar 2 12:54:19.211492 containerd[1890]: time="2026-03-02T12:54:19.209879740Z" level=info msg="Start snapshots syncer" Mar 2 12:54:19.211492 containerd[1890]: time="2026-03-02T12:54:19.209907996Z" level=info msg="loading plugin" id=io.containerd.cri.v1.runtime type=io.containerd.cri.v1 Mar 2 12:54:19.211568 containerd[1890]: time="2026-03-02T12:54:19.210078996Z" level=info msg="starting cri plugin" config="{\"containerd\":{\"defaultRuntimeName\":\"runc\",\"runtimes\":{\"runc\":{\"runtimeType\":\"io.containerd.runc.v2\",\"runtimePath\":\"\",\"PodAnnotations\":null,\"ContainerAnnotations\":null,\"options\":{\"BinaryName\":\"\",\"CriuImagePath\":\"\",\"CriuWorkPath\":\"\",\"IoGid\":0,\"IoUid\":0,\"NoNewKeyring\":false,\"Root\":\"\",\"ShimCgroup\":\"\",\"SystemdCgroup\":true},\"privileged_without_host_devices\":false,\"privileged_without_host_devices_all_devices_allowed\":false,\"baseRuntimeSpec\":\"\",\"cniConfDir\":\"\",\"cniMaxConfNum\":0,\"snapshotter\":\"\",\"sandboxer\":\"podsandbox\",\"io_type\":\"\"}},\"ignoreBlockIONotEnabledErrors\":false,\"ignoreRdtNotEnabledErrors\":false},\"cni\":{\"binDir\":\"/opt/cni/bin\",\"confDir\":\"/etc/cni/net.d\",\"maxConfNum\":1,\"setupSerially\":false,\"confTemplate\":\"\",\"ipPref\":\"\",\"useInternalLoopback\":false},\"enableSelinux\":true,\"selinuxCategoryRange\":1024,\"maxContainerLogSize\":16384,\"disableApparmor\":false,\"restrictOOMScoreAdj\":false,\"disableProcMount\":false,\"unsetSeccompProfile\":\"\",\"tolerateMissingHugetlbController\":true,\"disableHugetlbController\":true,\"device_ownership_from_security_context\":false,\"ignoreImageDefinedVolumes\":false,\"netnsMountsUnderStateDir\":false,\"enableUnprivilegedPorts\":true,\"enableUnprivilegedICMP\":true,\"enableCDI\":true,\"cdiSpecDirs\":[\"/etc/cdi\",\"/var/run/cdi\"],\"drainExecSyncIOTimeout\":\"0s\",\"ignoreDeprecationWarnings\":null,\"containerdRootDir\":\"/var/lib/containerd\",\"containerdEndpoint\":\"/run/containerd/containerd.sock\",\"rootDir\":\"/var/lib/containerd/io.containerd.grpc.v1.cri\",\"stateDir\":\"/run/containerd/io.containerd.grpc.v1.cri\"}" Mar 2 12:54:19.211568 containerd[1890]: time="2026-03-02T12:54:19.210111780Z" level=info msg="loading plugin" id=io.containerd.podsandbox.controller.v1.podsandbox type=io.containerd.podsandbox.controller.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210166028Z" level=info msg="loading plugin" id=io.containerd.sandbox.controller.v1.shim type=io.containerd.sandbox.controller.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210254068Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandbox-controllers type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210274652Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.sandboxes type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210281836Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.snapshots type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210288084Z" level=info msg="loading plugin" id=io.containerd.streaming.v1.manager type=io.containerd.streaming.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210296004Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.streaming type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210305268Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.tasks type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210312236Z" level=info msg="loading plugin" id=io.containerd.transfer.v1.local type=io.containerd.transfer.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210330524Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.transfer type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210337324Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.version type=io.containerd.grpc.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210343556Z" level=info msg="loading plugin" id=io.containerd.monitor.container.v1.restart type=io.containerd.monitor.container.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210376460Z" level=info msg="loading plugin" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210386588Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.tracing.processor.v1.otlp type=io.containerd.tracing.processor.v1 Mar 2 12:54:19.211648 containerd[1890]: time="2026-03-02T12:54:19.210391740Z" level=info msg="loading plugin" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210397844Z" level=info msg="skip loading plugin" error="skip plugin: tracing endpoint not configured" id=io.containerd.internal.v1.tracing type=io.containerd.internal.v1 Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210418996Z" level=info msg="loading plugin" id=io.containerd.ttrpc.v1.otelttrpc type=io.containerd.ttrpc.v1 Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210424740Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.healthcheck type=io.containerd.grpc.v1 Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210435564Z" level=info msg="loading plugin" id=io.containerd.nri.v1.nri type=io.containerd.nri.v1 Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210446892Z" level=info msg="runtime interface created" Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210450260Z" level=info msg="created NRI interface" Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210455172Z" level=info msg="loading plugin" id=io.containerd.grpc.v1.cri type=io.containerd.grpc.v1 Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210462860Z" level=info msg="Connect containerd service" Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210478676Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Mar 2 12:54:19.211796 containerd[1890]: time="2026-03-02T12:54:19.210995612Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 12:54:19.329496 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:54:19.444394 (kubelet)[2043]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:54:19.559011 containerd[1890]: time="2026-03-02T12:54:19.558811612Z" level=info msg="Start subscribing containerd event" Mar 2 12:54:19.559011 containerd[1890]: time="2026-03-02T12:54:19.558922940Z" level=info msg="Start recovering state" Mar 2 12:54:19.559011 containerd[1890]: time="2026-03-02T12:54:19.558954372Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 2 12:54:19.559011 containerd[1890]: time="2026-03-02T12:54:19.558999972Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559205804Z" level=info msg="Start event monitor" Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559223988Z" level=info msg="Start cni network conf syncer for default" Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559451028Z" level=info msg="Start streaming server" Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559460188Z" level=info msg="Registered namespace \"k8s.io\" with NRI" Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559468364Z" level=info msg="runtime interface starting up..." Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559472244Z" level=info msg="starting plugins..." Mar 2 12:54:19.559592 containerd[1890]: time="2026-03-02T12:54:19.559486820Z" level=info msg="Synchronizing NRI (plugin) with current runtime state" Mar 2 12:54:19.559842 systemd[1]: Started containerd.service - containerd container runtime. Mar 2 12:54:19.564386 containerd[1890]: time="2026-03-02T12:54:19.564357716Z" level=info msg="containerd successfully booted in 0.382270s" Mar 2 12:54:19.565679 systemd[1]: Reached target multi-user.target - Multi-User System. Mar 2 12:54:19.571214 systemd[1]: Startup finished in 1.707s (kernel) + 12.495s (initrd) + 10.726s (userspace) = 24.928s. Mar 2 12:54:19.855506 kubelet[2043]: E0302 12:54:19.854687 2043 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:54:19.856893 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:54:19.857094 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:54:19.857495 systemd[1]: kubelet.service: Consumed 522ms CPU time, 256.5M memory peak. Mar 2 12:54:19.900796 login[2025]: pam_lastlog(login:session): file /var/log/lastlog is locked/write, retrying Mar 2 12:54:19.902074 login[2026]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:19.907781 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Mar 2 12:54:19.910344 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Mar 2 12:54:19.915037 systemd-logind[1868]: New session 1 of user core. Mar 2 12:54:19.937837 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Mar 2 12:54:19.939732 systemd[1]: Starting user@500.service - User Manager for UID 500... Mar 2 12:54:19.993339 (systemd)[2060]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 2 12:54:19.995295 systemd-logind[1868]: New session c1 of user core. Mar 2 12:54:20.178445 systemd[2060]: Queued start job for default target default.target. Mar 2 12:54:20.184128 systemd[2060]: Created slice app.slice - User Application Slice. Mar 2 12:54:20.184153 systemd[2060]: Reached target paths.target - Paths. Mar 2 12:54:20.184182 systemd[2060]: Reached target timers.target - Timers. Mar 2 12:54:20.185268 systemd[2060]: Starting dbus.socket - D-Bus User Message Bus Socket... Mar 2 12:54:20.192387 systemd[2060]: Listening on dbus.socket - D-Bus User Message Bus Socket. Mar 2 12:54:20.192448 systemd[2060]: Reached target sockets.target - Sockets. Mar 2 12:54:20.192477 systemd[2060]: Reached target basic.target - Basic System. Mar 2 12:54:20.192497 systemd[2060]: Reached target default.target - Main User Target. Mar 2 12:54:20.192515 systemd[2060]: Startup finished in 192ms. Mar 2 12:54:20.192602 systemd[1]: Started user@500.service - User Manager for UID 500. Mar 2 12:54:20.194418 systemd[1]: Started session-1.scope - Session 1 of User core. Mar 2 12:54:20.696985 waagent[2021]: 2026-03-02T12:54:20.696913Z INFO Daemon Daemon Azure Linux Agent Version: 2.12.0.4 Mar 2 12:54:20.705744 waagent[2021]: 2026-03-02T12:54:20.702232Z INFO Daemon Daemon OS: flatcar 4459.2.101 Mar 2 12:54:20.706058 waagent[2021]: 2026-03-02T12:54:20.706023Z INFO Daemon Daemon Python: 3.11.13 Mar 2 12:54:20.709722 waagent[2021]: 2026-03-02T12:54:20.709686Z INFO Daemon Daemon Run daemon Mar 2 12:54:20.712899 waagent[2021]: 2026-03-02T12:54:20.712867Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4459.2.101' Mar 2 12:54:20.720149 waagent[2021]: 2026-03-02T12:54:20.720117Z INFO Daemon Daemon Using waagent for provisioning Mar 2 12:54:20.724398 waagent[2021]: 2026-03-02T12:54:20.724367Z INFO Daemon Daemon Activate resource disk Mar 2 12:54:20.728384 waagent[2021]: 2026-03-02T12:54:20.728352Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Mar 2 12:54:20.737026 waagent[2021]: 2026-03-02T12:54:20.736990Z INFO Daemon Daemon Found device: None Mar 2 12:54:20.740417 waagent[2021]: 2026-03-02T12:54:20.740373Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Mar 2 12:54:20.746692 waagent[2021]: 2026-03-02T12:54:20.746653Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Mar 2 12:54:20.755964 waagent[2021]: 2026-03-02T12:54:20.755925Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 2 12:54:20.760308 waagent[2021]: 2026-03-02T12:54:20.760274Z INFO Daemon Daemon Running default provisioning handler Mar 2 12:54:20.769704 waagent[2021]: 2026-03-02T12:54:20.769673Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Mar 2 12:54:20.780030 waagent[2021]: 2026-03-02T12:54:20.779999Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Mar 2 12:54:20.787587 waagent[2021]: 2026-03-02T12:54:20.787553Z INFO Daemon Daemon cloud-init is enabled: False Mar 2 12:54:20.791374 waagent[2021]: 2026-03-02T12:54:20.791345Z INFO Daemon Daemon Copying ovf-env.xml Mar 2 12:54:20.889861 waagent[2021]: 2026-03-02T12:54:20.889795Z INFO Daemon Daemon Successfully mounted dvd Mar 2 12:54:20.901133 login[2025]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:20.904668 systemd-logind[1868]: New session 2 of user core. Mar 2 12:54:20.910529 systemd[1]: Started session-2.scope - Session 2 of User core. Mar 2 12:54:20.915856 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Mar 2 12:54:20.920369 waagent[2021]: 2026-03-02T12:54:20.916775Z INFO Daemon Daemon Detect protocol endpoint Mar 2 12:54:20.920653 waagent[2021]: 2026-03-02T12:54:20.920616Z INFO Daemon Daemon Clean protocol and wireserver endpoint Mar 2 12:54:20.925568 waagent[2021]: 2026-03-02T12:54:20.925532Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Mar 2 12:54:20.930912 waagent[2021]: 2026-03-02T12:54:20.930875Z INFO Daemon Daemon Test for route to 168.63.129.16 Mar 2 12:54:20.935251 waagent[2021]: 2026-03-02T12:54:20.935215Z INFO Daemon Daemon Route to 168.63.129.16 exists Mar 2 12:54:20.939608 waagent[2021]: 2026-03-02T12:54:20.939575Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Mar 2 12:54:20.983198 waagent[2021]: 2026-03-02T12:54:20.983083Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Mar 2 12:54:20.988576 waagent[2021]: 2026-03-02T12:54:20.988548Z INFO Daemon Daemon Wire protocol version:2012-11-30 Mar 2 12:54:20.992759 waagent[2021]: 2026-03-02T12:54:20.992724Z INFO Daemon Daemon Server preferred version:2015-04-05 Mar 2 12:54:21.145537 waagent[2021]: 2026-03-02T12:54:21.145469Z INFO Daemon Daemon Initializing goal state during protocol detection Mar 2 12:54:21.151036 waagent[2021]: 2026-03-02T12:54:21.150983Z INFO Daemon Daemon Forcing an update of the goal state. Mar 2 12:54:21.158720 waagent[2021]: 2026-03-02T12:54:21.158681Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 2 12:54:21.176761 waagent[2021]: 2026-03-02T12:54:21.176723Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.179 Mar 2 12:54:21.181130 waagent[2021]: 2026-03-02T12:54:21.181097Z INFO Daemon Mar 2 12:54:21.183267 waagent[2021]: 2026-03-02T12:54:21.183236Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: ac286f06-e8d0-4857-a98c-a7f4d7c0511b eTag: 6227309068230716841 source: Fabric] Mar 2 12:54:21.191845 waagent[2021]: 2026-03-02T12:54:21.191810Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Mar 2 12:54:21.196588 waagent[2021]: 2026-03-02T12:54:21.196557Z INFO Daemon Mar 2 12:54:21.198651 waagent[2021]: 2026-03-02T12:54:21.198622Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Mar 2 12:54:21.207756 waagent[2021]: 2026-03-02T12:54:21.207725Z INFO Daemon Daemon Downloading artifacts profile blob Mar 2 12:54:21.276769 waagent[2021]: 2026-03-02T12:54:21.276681Z INFO Daemon Downloaded certificate {'thumbprint': '1D0A33F885FC94563A1C9F224FD87A9CA5D0A371', 'hasPrivateKey': True} Mar 2 12:54:21.283843 waagent[2021]: 2026-03-02T12:54:21.283809Z INFO Daemon Fetch goal state completed Mar 2 12:54:21.295193 waagent[2021]: 2026-03-02T12:54:21.295163Z INFO Daemon Daemon Starting provisioning Mar 2 12:54:21.299075 waagent[2021]: 2026-03-02T12:54:21.299037Z INFO Daemon Daemon Handle ovf-env.xml. Mar 2 12:54:21.303007 waagent[2021]: 2026-03-02T12:54:21.302976Z INFO Daemon Daemon Set hostname [ci-4459.2.101-f64d7d9b2a] Mar 2 12:54:21.322701 waagent[2021]: 2026-03-02T12:54:21.322662Z INFO Daemon Daemon Publish hostname [ci-4459.2.101-f64d7d9b2a] Mar 2 12:54:21.327199 waagent[2021]: 2026-03-02T12:54:21.327160Z INFO Daemon Daemon Examine /proc/net/route for primary interface Mar 2 12:54:21.331702 waagent[2021]: 2026-03-02T12:54:21.331665Z INFO Daemon Daemon Primary interface is [eth0] Mar 2 12:54:21.341585 systemd-networkd[1485]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Mar 2 12:54:21.341788 systemd-networkd[1485]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 2 12:54:21.342230 systemd-networkd[1485]: eth0: DHCP lease lost Mar 2 12:54:21.342659 waagent[2021]: 2026-03-02T12:54:21.342613Z INFO Daemon Daemon Create user account if not exists Mar 2 12:54:21.346529 waagent[2021]: 2026-03-02T12:54:21.346495Z INFO Daemon Daemon User core already exists, skip useradd Mar 2 12:54:21.351016 waagent[2021]: 2026-03-02T12:54:21.350988Z INFO Daemon Daemon Configure sudoer Mar 2 12:54:21.358466 waagent[2021]: 2026-03-02T12:54:21.358426Z INFO Daemon Daemon Configure sshd Mar 2 12:54:21.365350 waagent[2021]: 2026-03-02T12:54:21.365302Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Mar 2 12:54:21.366447 systemd-networkd[1485]: eth0: DHCPv4 address 10.200.20.30/24, gateway 10.200.20.1 acquired from 168.63.129.16 Mar 2 12:54:21.375182 waagent[2021]: 2026-03-02T12:54:21.375124Z INFO Daemon Daemon Deploy ssh public key. Mar 2 12:54:22.482394 waagent[2021]: 2026-03-02T12:54:22.482338Z INFO Daemon Daemon Provisioning complete Mar 2 12:54:22.497836 waagent[2021]: 2026-03-02T12:54:22.497799Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Mar 2 12:54:22.502408 waagent[2021]: 2026-03-02T12:54:22.502373Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Mar 2 12:54:22.509505 waagent[2021]: 2026-03-02T12:54:22.509470Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.12.0.4 is the most current agent Mar 2 12:54:22.609441 waagent[2110]: 2026-03-02T12:54:22.609311Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.12.0.4) Mar 2 12:54:22.610442 waagent[2110]: 2026-03-02T12:54:22.609798Z INFO ExtHandler ExtHandler OS: flatcar 4459.2.101 Mar 2 12:54:22.610442 waagent[2110]: 2026-03-02T12:54:22.609865Z INFO ExtHandler ExtHandler Python: 3.11.13 Mar 2 12:54:22.610442 waagent[2110]: 2026-03-02T12:54:22.609905Z INFO ExtHandler ExtHandler CPU Arch: aarch64 Mar 2 12:54:22.662154 waagent[2110]: 2026-03-02T12:54:22.662083Z INFO ExtHandler ExtHandler Distro: flatcar-4459.2.101; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.13; Arch: aarch64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.22.0; Mar 2 12:54:22.662506 waagent[2110]: 2026-03-02T12:54:22.662471Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 2 12:54:22.662659 waagent[2110]: 2026-03-02T12:54:22.662636Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 2 12:54:22.668649 waagent[2110]: 2026-03-02T12:54:22.668602Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Mar 2 12:54:22.674080 waagent[2110]: 2026-03-02T12:54:22.674048Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.179 Mar 2 12:54:22.674579 waagent[2110]: 2026-03-02T12:54:22.674548Z INFO ExtHandler Mar 2 12:54:22.674702 waagent[2110]: 2026-03-02T12:54:22.674679Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 0396d2da-2079-4a25-bc87-b9216663630b eTag: 6227309068230716841 source: Fabric] Mar 2 12:54:22.674987 waagent[2110]: 2026-03-02T12:54:22.674960Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Mar 2 12:54:22.675544 waagent[2110]: 2026-03-02T12:54:22.675510Z INFO ExtHandler Mar 2 12:54:22.675676 waagent[2110]: 2026-03-02T12:54:22.675653Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Mar 2 12:54:22.679229 waagent[2110]: 2026-03-02T12:54:22.679201Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Mar 2 12:54:22.743209 waagent[2110]: 2026-03-02T12:54:22.743107Z INFO ExtHandler Downloaded certificate {'thumbprint': '1D0A33F885FC94563A1C9F224FD87A9CA5D0A371', 'hasPrivateKey': True} Mar 2 12:54:22.743756 waagent[2110]: 2026-03-02T12:54:22.743721Z INFO ExtHandler Fetch goal state completed Mar 2 12:54:22.756188 waagent[2110]: 2026-03-02T12:54:22.756152Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.4.4 27 Jan 2026 (Library: OpenSSL 3.4.4 27 Jan 2026) Mar 2 12:54:22.759706 waagent[2110]: 2026-03-02T12:54:22.759667Z INFO ExtHandler ExtHandler WALinuxAgent-2.12.0.4 running as process 2110 Mar 2 12:54:22.759918 waagent[2110]: 2026-03-02T12:54:22.759887Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Mar 2 12:54:22.760224 waagent[2110]: 2026-03-02T12:54:22.760195Z INFO ExtHandler ExtHandler ******** AutoUpdate.UpdateToLatestVersion is set to False, not processing the operation ******** Mar 2 12:54:22.761450 waagent[2110]: 2026-03-02T12:54:22.761386Z INFO ExtHandler ExtHandler [CGI] Cgroup monitoring is not supported on ['flatcar', '4459.2.101', '', 'Flatcar Container Linux by Kinvolk'] Mar 2 12:54:22.761853 waagent[2110]: 2026-03-02T12:54:22.761820Z INFO ExtHandler ExtHandler [CGI] Agent will reset the quotas in case distro: ['flatcar', '4459.2.101', '', 'Flatcar Container Linux by Kinvolk'] went from supported to unsupported Mar 2 12:54:22.762042 waagent[2110]: 2026-03-02T12:54:22.762015Z INFO ExtHandler ExtHandler [CGI] Agent cgroups enabled: False Mar 2 12:54:22.762564 waagent[2110]: 2026-03-02T12:54:22.762532Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Mar 2 12:54:22.807084 waagent[2110]: 2026-03-02T12:54:22.807037Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Mar 2 12:54:22.807625 waagent[2110]: 2026-03-02T12:54:22.807374Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Mar 2 12:54:22.812011 waagent[2110]: 2026-03-02T12:54:22.811990Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Mar 2 12:54:22.816845 systemd[1]: Reload requested from client PID 2125 ('systemctl') (unit waagent.service)... Mar 2 12:54:22.816860 systemd[1]: Reloading... Mar 2 12:54:22.886455 zram_generator::config[2165]: No configuration found. Mar 2 12:54:23.025977 systemd[1]: Reloading finished in 208 ms. Mar 2 12:54:23.038327 waagent[2110]: 2026-03-02T12:54:23.037209Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Mar 2 12:54:23.038327 waagent[2110]: 2026-03-02T12:54:23.037356Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Mar 2 12:54:24.282329 waagent[2110]: 2026-03-02T12:54:24.282245Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Mar 2 12:54:24.282644 waagent[2110]: 2026-03-02T12:54:24.282577Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: 1. configuration enabled [True], 2. cgroups v1 enabled [False] OR cgroups v2 is in use and v2 resource limiting configuration enabled [False], 3. python supported: [True] Mar 2 12:54:24.283245 waagent[2110]: 2026-03-02T12:54:24.283204Z INFO ExtHandler ExtHandler Starting env monitor service. Mar 2 12:54:24.283519 waagent[2110]: 2026-03-02T12:54:24.283483Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Mar 2 12:54:24.284319 waagent[2110]: 2026-03-02T12:54:24.283716Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 2 12:54:24.284319 waagent[2110]: 2026-03-02T12:54:24.283784Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 2 12:54:24.284319 waagent[2110]: 2026-03-02T12:54:24.283950Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Mar 2 12:54:24.284319 waagent[2110]: 2026-03-02T12:54:24.284086Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Mar 2 12:54:24.284319 waagent[2110]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Mar 2 12:54:24.284319 waagent[2110]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Mar 2 12:54:24.284319 waagent[2110]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Mar 2 12:54:24.284319 waagent[2110]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Mar 2 12:54:24.284319 waagent[2110]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 2 12:54:24.284319 waagent[2110]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Mar 2 12:54:24.284623 waagent[2110]: 2026-03-02T12:54:24.284586Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Mar 2 12:54:24.284683 waagent[2110]: 2026-03-02T12:54:24.284642Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Mar 2 12:54:24.284948 waagent[2110]: 2026-03-02T12:54:24.284921Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Mar 2 12:54:24.285070 waagent[2110]: 2026-03-02T12:54:24.285044Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Mar 2 12:54:24.285237 waagent[2110]: 2026-03-02T12:54:24.285210Z INFO EnvHandler ExtHandler Configure routes Mar 2 12:54:24.285343 waagent[2110]: 2026-03-02T12:54:24.285321Z INFO EnvHandler ExtHandler Gateway:None Mar 2 12:54:24.285488 waagent[2110]: 2026-03-02T12:54:24.285460Z INFO EnvHandler ExtHandler Routes:None Mar 2 12:54:24.285627 waagent[2110]: 2026-03-02T12:54:24.285592Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Mar 2 12:54:24.285712 waagent[2110]: 2026-03-02T12:54:24.285690Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Mar 2 12:54:24.286306 waagent[2110]: 2026-03-02T12:54:24.286273Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Mar 2 12:54:24.292131 waagent[2110]: 2026-03-02T12:54:24.292094Z INFO ExtHandler ExtHandler Mar 2 12:54:24.292262 waagent[2110]: 2026-03-02T12:54:24.292236Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 8c7c583c-b070-4967-8bda-f2f9446d12b8 correlation 140713f7-88ad-4ae9-af8e-9e1153fdf9c3 created: 2026-03-02T12:53:24.313828Z] Mar 2 12:54:24.292647 waagent[2110]: 2026-03-02T12:54:24.292614Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Mar 2 12:54:24.293137 waagent[2110]: 2026-03-02T12:54:24.293105Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 1 ms] Mar 2 12:54:24.316225 waagent[2110]: 2026-03-02T12:54:24.316188Z WARNING ExtHandler ExtHandler Failed to get firewall packets: 'iptables -w -t security -L OUTPUT --zero OUTPUT -nxv' failed: 2 (iptables v1.8.11 (nf_tables): Illegal option `--numeric' with this command Mar 2 12:54:24.316225 waagent[2110]: Try `iptables -h' or 'iptables --help' for more information.) Mar 2 12:54:24.316690 waagent[2110]: 2026-03-02T12:54:24.316661Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.12.0.4 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 8B5767F6-971A-488C-A227-E92FB46C7F42;DroppedPackets: -1;UpdateGSErrors: 0;AutoUpdate: 0;UpdateMode: SelfUpdate;] Mar 2 12:54:24.335109 waagent[2110]: 2026-03-02T12:54:24.335060Z INFO MonitorHandler ExtHandler Network interfaces: Mar 2 12:54:24.335109 waagent[2110]: Executing ['ip', '-a', '-o', 'link']: Mar 2 12:54:24.335109 waagent[2110]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Mar 2 12:54:24.335109 waagent[2110]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:88:5e:8d brd ff:ff:ff:ff:ff:ff Mar 2 12:54:24.335109 waagent[2110]: 3: enP21044s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 7c:ed:8d:88:5e:8d brd ff:ff:ff:ff:ff:ff\ altname enP21044p0s2 Mar 2 12:54:24.335109 waagent[2110]: Executing ['ip', '-4', '-a', '-o', 'address']: Mar 2 12:54:24.335109 waagent[2110]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Mar 2 12:54:24.335109 waagent[2110]: 2: eth0 inet 10.200.20.30/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Mar 2 12:54:24.335109 waagent[2110]: Executing ['ip', '-6', '-a', '-o', 'address']: Mar 2 12:54:24.335109 waagent[2110]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Mar 2 12:54:24.335109 waagent[2110]: 2: eth0 inet6 fe80::7eed:8dff:fe88:5e8d/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Mar 2 12:54:24.385774 waagent[2110]: 2026-03-02T12:54:24.385734Z INFO EnvHandler ExtHandler Created firewall rules for the Azure Fabric: Mar 2 12:54:24.385774 waagent[2110]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:54:24.385774 waagent[2110]: pkts bytes target prot opt in out source destination Mar 2 12:54:24.385774 waagent[2110]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:54:24.385774 waagent[2110]: pkts bytes target prot opt in out source destination Mar 2 12:54:24.385774 waagent[2110]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:54:24.385774 waagent[2110]: pkts bytes target prot opt in out source destination Mar 2 12:54:24.385774 waagent[2110]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 2 12:54:24.385774 waagent[2110]: 1 52 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 2 12:54:24.385774 waagent[2110]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 2 12:54:24.388280 waagent[2110]: 2026-03-02T12:54:24.388247Z INFO EnvHandler ExtHandler Current Firewall rules: Mar 2 12:54:24.388280 waagent[2110]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:54:24.388280 waagent[2110]: pkts bytes target prot opt in out source destination Mar 2 12:54:24.388280 waagent[2110]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:54:24.388280 waagent[2110]: pkts bytes target prot opt in out source destination Mar 2 12:54:24.388280 waagent[2110]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Mar 2 12:54:24.388280 waagent[2110]: pkts bytes target prot opt in out source destination Mar 2 12:54:24.388280 waagent[2110]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Mar 2 12:54:24.388280 waagent[2110]: 1 52 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Mar 2 12:54:24.388280 waagent[2110]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Mar 2 12:54:24.388715 waagent[2110]: 2026-03-02T12:54:24.388691Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Mar 2 12:54:29.933114 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 2 12:54:29.934913 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:54:30.031655 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:54:30.039672 (kubelet)[2259]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:54:30.142719 kubelet[2259]: E0302 12:54:30.142650 2259 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:54:30.145674 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:54:30.145888 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:54:30.146477 systemd[1]: kubelet.service: Consumed 184ms CPU time, 105.1M memory peak. Mar 2 12:54:40.182439 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Mar 2 12:54:40.185549 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:54:40.447562 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Mar 2 12:54:40.449657 systemd[1]: Started sshd@0-10.200.20.30:22-10.200.16.10:46980.service - OpenSSH per-connection server daemon (10.200.16.10:46980). Mar 2 12:54:42.164225 chronyd[1846]: Selected source PHC0 Mar 2 12:54:43.588964 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:54:43.593768 (kubelet)[2277]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:54:43.620171 kubelet[2277]: E0302 12:54:43.620140 2277 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:54:43.622271 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:54:43.622373 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:54:43.622743 systemd[1]: kubelet.service: Consumed 103ms CPU time, 104.7M memory peak. Mar 2 12:54:44.027938 sshd[2270]: Accepted publickey for core from 10.200.16.10 port 46980 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:44.028988 sshd-session[2270]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:44.032616 systemd-logind[1868]: New session 3 of user core. Mar 2 12:54:44.043797 systemd[1]: Started session-3.scope - Session 3 of User core. Mar 2 12:54:44.355605 systemd[1]: Started sshd@1-10.200.20.30:22-10.200.16.10:46996.service - OpenSSH per-connection server daemon (10.200.16.10:46996). Mar 2 12:54:44.779551 sshd[2288]: Accepted publickey for core from 10.200.16.10 port 46996 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:44.780637 sshd-session[2288]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:44.784098 systemd-logind[1868]: New session 4 of user core. Mar 2 12:54:44.791522 systemd[1]: Started session-4.scope - Session 4 of User core. Mar 2 12:54:45.014260 sshd[2291]: Connection closed by 10.200.16.10 port 46996 Mar 2 12:54:45.014161 sshd-session[2288]: pam_unix(sshd:session): session closed for user core Mar 2 12:54:45.016953 systemd[1]: sshd@1-10.200.20.30:22-10.200.16.10:46996.service: Deactivated successfully. Mar 2 12:54:45.018643 systemd[1]: session-4.scope: Deactivated successfully. Mar 2 12:54:45.019805 systemd-logind[1868]: Session 4 logged out. Waiting for processes to exit. Mar 2 12:54:45.020880 systemd-logind[1868]: Removed session 4. Mar 2 12:54:45.100975 systemd[1]: Started sshd@2-10.200.20.30:22-10.200.16.10:47006.service - OpenSSH per-connection server daemon (10.200.16.10:47006). Mar 2 12:54:45.520567 sshd[2297]: Accepted publickey for core from 10.200.16.10 port 47006 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:45.521698 sshd-session[2297]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:45.525779 systemd-logind[1868]: New session 5 of user core. Mar 2 12:54:45.531539 systemd[1]: Started session-5.scope - Session 5 of User core. Mar 2 12:54:45.751749 sshd[2300]: Connection closed by 10.200.16.10 port 47006 Mar 2 12:54:45.752287 sshd-session[2297]: pam_unix(sshd:session): session closed for user core Mar 2 12:54:45.756014 systemd-logind[1868]: Session 5 logged out. Waiting for processes to exit. Mar 2 12:54:45.756286 systemd[1]: sshd@2-10.200.20.30:22-10.200.16.10:47006.service: Deactivated successfully. Mar 2 12:54:45.758611 systemd[1]: session-5.scope: Deactivated successfully. Mar 2 12:54:45.759992 systemd-logind[1868]: Removed session 5. Mar 2 12:54:45.841910 systemd[1]: Started sshd@3-10.200.20.30:22-10.200.16.10:47016.service - OpenSSH per-connection server daemon (10.200.16.10:47016). Mar 2 12:54:46.262437 sshd[2306]: Accepted publickey for core from 10.200.16.10 port 47016 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:46.263145 sshd-session[2306]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:46.266512 systemd-logind[1868]: New session 6 of user core. Mar 2 12:54:46.275537 systemd[1]: Started session-6.scope - Session 6 of User core. Mar 2 12:54:46.495288 sshd[2309]: Connection closed by 10.200.16.10 port 47016 Mar 2 12:54:46.495832 sshd-session[2306]: pam_unix(sshd:session): session closed for user core Mar 2 12:54:46.498665 systemd[1]: sshd@3-10.200.20.30:22-10.200.16.10:47016.service: Deactivated successfully. Mar 2 12:54:46.500020 systemd[1]: session-6.scope: Deactivated successfully. Mar 2 12:54:46.500611 systemd-logind[1868]: Session 6 logged out. Waiting for processes to exit. Mar 2 12:54:46.501759 systemd-logind[1868]: Removed session 6. Mar 2 12:54:46.582727 systemd[1]: Started sshd@4-10.200.20.30:22-10.200.16.10:47026.service - OpenSSH per-connection server daemon (10.200.16.10:47026). Mar 2 12:54:47.005546 sshd[2315]: Accepted publickey for core from 10.200.16.10 port 47026 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:47.006590 sshd-session[2315]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:47.009790 systemd-logind[1868]: New session 7 of user core. Mar 2 12:54:47.017683 systemd[1]: Started session-7.scope - Session 7 of User core. Mar 2 12:54:47.250002 sudo[2319]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Mar 2 12:54:47.250207 sudo[2319]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:54:47.262983 sudo[2319]: pam_unix(sudo:session): session closed for user root Mar 2 12:54:47.340511 sshd[2318]: Connection closed by 10.200.16.10 port 47026 Mar 2 12:54:47.341102 sshd-session[2315]: pam_unix(sshd:session): session closed for user core Mar 2 12:54:47.344854 systemd[1]: sshd@4-10.200.20.30:22-10.200.16.10:47026.service: Deactivated successfully. Mar 2 12:54:47.346132 systemd[1]: session-7.scope: Deactivated successfully. Mar 2 12:54:47.347224 systemd-logind[1868]: Session 7 logged out. Waiting for processes to exit. Mar 2 12:54:47.348112 systemd-logind[1868]: Removed session 7. Mar 2 12:54:47.426728 systemd[1]: Started sshd@5-10.200.20.30:22-10.200.16.10:47040.service - OpenSSH per-connection server daemon (10.200.16.10:47040). Mar 2 12:54:47.849293 sshd[2325]: Accepted publickey for core from 10.200.16.10 port 47040 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:47.850010 sshd-session[2325]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:47.853251 systemd-logind[1868]: New session 8 of user core. Mar 2 12:54:47.864523 systemd[1]: Started session-8.scope - Session 8 of User core. Mar 2 12:54:48.006056 sudo[2330]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Mar 2 12:54:48.006604 sudo[2330]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:54:48.015681 sudo[2330]: pam_unix(sudo:session): session closed for user root Mar 2 12:54:48.019056 sudo[2329]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Mar 2 12:54:48.019491 sudo[2329]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:54:48.026799 systemd[1]: Starting audit-rules.service - Load Audit Rules... Mar 2 12:54:48.057190 augenrules[2352]: No rules Mar 2 12:54:48.058261 systemd[1]: audit-rules.service: Deactivated successfully. Mar 2 12:54:48.058484 systemd[1]: Finished audit-rules.service - Load Audit Rules. Mar 2 12:54:48.061234 sudo[2329]: pam_unix(sudo:session): session closed for user root Mar 2 12:54:48.139052 sshd[2328]: Connection closed by 10.200.16.10 port 47040 Mar 2 12:54:48.139543 sshd-session[2325]: pam_unix(sshd:session): session closed for user core Mar 2 12:54:48.143328 systemd[1]: sshd@5-10.200.20.30:22-10.200.16.10:47040.service: Deactivated successfully. Mar 2 12:54:48.145078 systemd[1]: session-8.scope: Deactivated successfully. Mar 2 12:54:48.146938 systemd-logind[1868]: Session 8 logged out. Waiting for processes to exit. Mar 2 12:54:48.148339 systemd-logind[1868]: Removed session 8. Mar 2 12:54:48.226601 systemd[1]: Started sshd@6-10.200.20.30:22-10.200.16.10:47050.service - OpenSSH per-connection server daemon (10.200.16.10:47050). Mar 2 12:54:48.653419 sshd[2361]: Accepted publickey for core from 10.200.16.10 port 47050 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:54:48.654108 sshd-session[2361]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:54:48.657946 systemd-logind[1868]: New session 9 of user core. Mar 2 12:54:48.660517 systemd[1]: Started session-9.scope - Session 9 of User core. Mar 2 12:54:48.809460 sudo[2365]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 2 12:54:48.809665 sudo[2365]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Mar 2 12:54:50.314374 systemd[1]: Starting docker.service - Docker Application Container Engine... Mar 2 12:54:50.324827 (dockerd)[2383]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Mar 2 12:54:51.060420 dockerd[2383]: time="2026-03-02T12:54:51.060111187Z" level=info msg="Starting up" Mar 2 12:54:51.061616 dockerd[2383]: time="2026-03-02T12:54:51.061592746Z" level=info msg="OTEL tracing is not configured, using no-op tracer provider" Mar 2 12:54:51.068846 dockerd[2383]: time="2026-03-02T12:54:51.068818772Z" level=info msg="Creating a containerd client" address=/var/run/docker/libcontainerd/docker-containerd.sock timeout=1m0s Mar 2 12:54:51.102782 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2063666313-merged.mount: Deactivated successfully. Mar 2 12:54:51.155945 systemd[1]: var-lib-docker-metacopy\x2dcheck1263564730-merged.mount: Deactivated successfully. Mar 2 12:54:51.171129 dockerd[2383]: time="2026-03-02T12:54:51.171092309Z" level=info msg="Loading containers: start." Mar 2 12:54:51.231632 kernel: Initializing XFRM netlink socket Mar 2 12:54:51.601502 systemd-networkd[1485]: docker0: Link UP Mar 2 12:54:51.616278 dockerd[2383]: time="2026-03-02T12:54:51.616241176Z" level=info msg="Loading containers: done." Mar 2 12:54:51.642470 dockerd[2383]: time="2026-03-02T12:54:51.642435535Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 2 12:54:51.642581 dockerd[2383]: time="2026-03-02T12:54:51.642500713Z" level=info msg="Docker daemon" commit=6430e49a55babd9b8f4d08e70ecb2b68900770fe containerd-snapshotter=false storage-driver=overlay2 version=28.0.4 Mar 2 12:54:51.642581 dockerd[2383]: time="2026-03-02T12:54:51.642573683Z" level=info msg="Initializing buildkit" Mar 2 12:54:51.692257 dockerd[2383]: time="2026-03-02T12:54:51.692222323Z" level=info msg="Completed buildkit initialization" Mar 2 12:54:51.696718 dockerd[2383]: time="2026-03-02T12:54:51.696687894Z" level=info msg="Daemon has completed initialization" Mar 2 12:54:51.696791 dockerd[2383]: time="2026-03-02T12:54:51.696731400Z" level=info msg="API listen on /run/docker.sock" Mar 2 12:54:51.697821 systemd[1]: Started docker.service - Docker Application Container Engine. Mar 2 12:54:52.066423 containerd[1890]: time="2026-03-02T12:54:52.066286528Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\"" Mar 2 12:54:52.097384 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2851459443-merged.mount: Deactivated successfully. Mar 2 12:54:52.899020 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1518195447.mount: Deactivated successfully. Mar 2 12:54:53.682447 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Mar 2 12:54:53.683738 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:54:53.791691 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:54:53.801638 (kubelet)[2651]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:54:53.994747 kubelet[2651]: E0302 12:54:53.994615 2651 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:54:53.997075 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:54:53.997369 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:54:53.997745 systemd[1]: kubelet.service: Consumed 103ms CPU time, 107.1M memory peak. Mar 2 12:54:54.670842 containerd[1890]: time="2026-03-02T12:54:54.670784677Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:54.674132 containerd[1890]: time="2026-03-02T12:54:54.674106331Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.33.9: active requests=0, bytes read=27390174" Mar 2 12:54:54.677716 containerd[1890]: time="2026-03-02T12:54:54.677688736Z" level=info msg="ImageCreate event name:\"sha256:6dbc3c6e88c8bca1294fa5fafe73dbe01fb58d40e562dbfc8b8b4195940270c8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:54.684544 containerd[1890]: time="2026-03-02T12:54:54.684515561Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:54.685527 containerd[1890]: time="2026-03-02T12:54:54.685504683Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.33.9\" with image id \"sha256:6dbc3c6e88c8bca1294fa5fafe73dbe01fb58d40e562dbfc8b8b4195940270c8\", repo tag \"registry.k8s.io/kube-apiserver:v1.33.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:a1fe354f8b36dbce37fef26c3731e2376fb8eb7375e7df3068df7ad11656f022\", size \"27386773\" in 2.61917325s" Mar 2 12:54:54.685561 containerd[1890]: time="2026-03-02T12:54:54.685532675Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.33.9\" returns image reference \"sha256:6dbc3c6e88c8bca1294fa5fafe73dbe01fb58d40e562dbfc8b8b4195940270c8\"" Mar 2 12:54:54.686021 containerd[1890]: time="2026-03-02T12:54:54.685998672Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\"" Mar 2 12:54:55.965633 containerd[1890]: time="2026-03-02T12:54:55.965572262Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:55.969844 containerd[1890]: time="2026-03-02T12:54:55.969816460Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.33.9: active requests=0, bytes read=23552106" Mar 2 12:54:55.973536 containerd[1890]: time="2026-03-02T12:54:55.973513692Z" level=info msg="ImageCreate event name:\"sha256:c58be92c40cc41b6c83c361b92110b587104386f93c5b7a9fc66dffdd1523d17\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:55.978585 containerd[1890]: time="2026-03-02T12:54:55.978543574Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:55.979163 containerd[1890]: time="2026-03-02T12:54:55.979058779Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.33.9\" with image id \"sha256:c58be92c40cc41b6c83c361b92110b587104386f93c5b7a9fc66dffdd1523d17\", repo tag \"registry.k8s.io/kube-controller-manager:v1.33.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:a495c9f30cfd4d57ae6c27cb21e477b9b1ddebdace61762e80a06fe264a0d61a\", size \"25136510\" in 1.293030595s" Mar 2 12:54:55.979163 containerd[1890]: time="2026-03-02T12:54:55.979084812Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.33.9\" returns image reference \"sha256:c58be92c40cc41b6c83c361b92110b587104386f93c5b7a9fc66dffdd1523d17\"" Mar 2 12:54:55.979698 containerd[1890]: time="2026-03-02T12:54:55.979668803Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\"" Mar 2 12:54:57.817630 containerd[1890]: time="2026-03-02T12:54:57.817571907Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:57.821956 containerd[1890]: time="2026-03-02T12:54:57.821930436Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.33.9: active requests=0, bytes read=18301305" Mar 2 12:54:57.825399 containerd[1890]: time="2026-03-02T12:54:57.825371245Z" level=info msg="ImageCreate event name:\"sha256:5dcd4a0c93d95bd92241ba240a130ffbde67814e2b417a13c25738a7b0204e95\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:57.830526 containerd[1890]: time="2026-03-02T12:54:57.830483289Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:57.831045 containerd[1890]: time="2026-03-02T12:54:57.830797466Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.33.9\" with image id \"sha256:5dcd4a0c93d95bd92241ba240a130ffbde67814e2b417a13c25738a7b0204e95\", repo tag \"registry.k8s.io/kube-scheduler:v1.33.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:d1533368d3acd772e3d11225337a61be319b5ecf7523adeff7ebfe4107ab05b5\", size \"19885727\" in 1.851043636s" Mar 2 12:54:57.831045 containerd[1890]: time="2026-03-02T12:54:57.830823642Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.33.9\" returns image reference \"sha256:5dcd4a0c93d95bd92241ba240a130ffbde67814e2b417a13c25738a7b0204e95\"" Mar 2 12:54:57.831595 containerd[1890]: time="2026-03-02T12:54:57.831551685Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\"" Mar 2 12:54:58.877371 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2629568719.mount: Deactivated successfully. Mar 2 12:54:59.153208 containerd[1890]: time="2026-03-02T12:54:59.152719868Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.33.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:59.156789 containerd[1890]: time="2026-03-02T12:54:59.156761326Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.33.9: active requests=0, bytes read=28148870" Mar 2 12:54:59.160467 containerd[1890]: time="2026-03-02T12:54:59.160425991Z" level=info msg="ImageCreate event name:\"sha256:fb4f3cb8cccaec5975890c2ee802236a557e3f108da9c3c66ebec335ac73dcc9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:59.164418 containerd[1890]: time="2026-03-02T12:54:59.164375709Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:54:59.164820 containerd[1890]: time="2026-03-02T12:54:59.164675010Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.33.9\" with image id \"sha256:fb4f3cb8cccaec5975890c2ee802236a557e3f108da9c3c66ebec335ac73dcc9\", repo tag \"registry.k8s.io/kube-proxy:v1.33.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:079ba0e77e457dbf755e78bf3a6d736b7eb73d021fe53b853a0b82bbb2c17322\", size \"28147889\" in 1.333075196s" Mar 2 12:54:59.164820 containerd[1890]: time="2026-03-02T12:54:59.164704779Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.33.9\" returns image reference \"sha256:fb4f3cb8cccaec5975890c2ee802236a557e3f108da9c3c66ebec335ac73dcc9\"" Mar 2 12:54:59.165162 containerd[1890]: time="2026-03-02T12:54:59.165143606Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\"" Mar 2 12:54:59.836349 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount81219546.mount: Deactivated successfully. Mar 2 12:55:01.560572 containerd[1890]: time="2026-03-02T12:55:01.560520736Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.12.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:01.569275 containerd[1890]: time="2026-03-02T12:55:01.569244624Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.12.0: active requests=0, bytes read=19152117" Mar 2 12:55:01.575345 containerd[1890]: time="2026-03-02T12:55:01.575316963Z" level=info msg="ImageCreate event name:\"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:01.584031 containerd[1890]: time="2026-03-02T12:55:01.584002489Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:01.585111 containerd[1890]: time="2026-03-02T12:55:01.585083864Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.12.0\" with image id \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\", repo tag \"registry.k8s.io/coredns/coredns:v1.12.0\", repo digest \"registry.k8s.io/coredns/coredns@sha256:40384aa1f5ea6bfdc77997d243aec73da05f27aed0c5e9d65bfa98933c519d97\", size \"19148915\" in 2.419859038s" Mar 2 12:55:01.585147 containerd[1890]: time="2026-03-02T12:55:01.585120114Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.12.0\" returns image reference \"sha256:f72407be9e08c3a1b29a88318cbfee87b9f2da489f84015a5090b1e386e4dbc1\"" Mar 2 12:55:01.585551 containerd[1890]: time="2026-03-02T12:55:01.585526300Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Mar 2 12:55:01.928575 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Mar 2 12:55:02.149040 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3136581478.mount: Deactivated successfully. Mar 2 12:55:02.172942 containerd[1890]: time="2026-03-02T12:55:02.172893714Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:55:02.175576 containerd[1890]: time="2026-03-02T12:55:02.175424890Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268703" Mar 2 12:55:02.178393 containerd[1890]: time="2026-03-02T12:55:02.178366019Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:55:02.182351 containerd[1890]: time="2026-03-02T12:55:02.182301728Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Mar 2 12:55:02.183112 containerd[1890]: time="2026-03-02T12:55:02.182617974Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 597.063817ms" Mar 2 12:55:02.183112 containerd[1890]: time="2026-03-02T12:55:02.182652295Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Mar 2 12:55:02.183455 containerd[1890]: time="2026-03-02T12:55:02.183421393Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\"" Mar 2 12:55:02.900056 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount25089928.mount: Deactivated successfully. Mar 2 12:55:03.650858 update_engine[1871]: I20260302 12:55:03.650426 1871 update_attempter.cc:509] Updating boot flags... Mar 2 12:55:04.141600 containerd[1890]: time="2026-03-02T12:55:04.140934572Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.24-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:04.157482 containerd[1890]: time="2026-03-02T12:55:04.157435379Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.24-0: active requests=0, bytes read=21885780" Mar 2 12:55:04.161311 containerd[1890]: time="2026-03-02T12:55:04.161284830Z" level=info msg="ImageCreate event name:\"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:04.165700 containerd[1890]: time="2026-03-02T12:55:04.165676337Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:04.166201 containerd[1890]: time="2026-03-02T12:55:04.166172647Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.24-0\" with image id \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\", repo tag \"registry.k8s.io/etcd:3.5.24-0\", repo digest \"registry.k8s.io/etcd@sha256:251e7e490f64859d329cd963bc879dc04acf3d7195bb52c4c50b4a07bedf37d6\", size \"21882972\" in 1.982605152s" Mar 2 12:55:04.166201 containerd[1890]: time="2026-03-02T12:55:04.166199728Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.24-0\" returns image reference \"sha256:1211402d28f5813ed906916bfcdd0a7404c2f9048ef5bb54387a6745bc410eca\"" Mar 2 12:55:04.182761 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Mar 2 12:55:04.184112 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:55:04.315328 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:55:04.321736 (kubelet)[2926]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Mar 2 12:55:04.394440 kubelet[2926]: E0302 12:55:04.393719 2926 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 2 12:55:04.396351 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 2 12:55:04.396472 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 2 12:55:04.396726 systemd[1]: kubelet.service: Consumed 102ms CPU time, 107M memory peak. Mar 2 12:55:06.740311 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:55:06.740434 systemd[1]: kubelet.service: Consumed 102ms CPU time, 107M memory peak. Mar 2 12:55:06.742243 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:55:06.763898 systemd[1]: Reload requested from client PID 2956 ('systemctl') (unit session-9.scope)... Mar 2 12:55:06.763988 systemd[1]: Reloading... Mar 2 12:55:06.834450 zram_generator::config[2999]: No configuration found. Mar 2 12:55:06.981129 systemd[1]: Reloading finished in 216 ms. Mar 2 12:55:07.002920 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Mar 2 12:55:07.002986 systemd[1]: kubelet.service: Failed with result 'signal'. Mar 2 12:55:07.003298 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:55:07.003357 systemd[1]: kubelet.service: Consumed 64ms CPU time, 94M memory peak. Mar 2 12:55:07.004858 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:55:07.501912 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:55:07.504735 (kubelet)[3067]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 12:55:07.528432 kubelet[3067]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 12:55:07.528432 kubelet[3067]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 12:55:07.528432 kubelet[3067]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 12:55:07.528432 kubelet[3067]: I0302 12:55:07.527230 3067 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 12:55:07.815176 kubelet[3067]: I0302 12:55:07.814819 3067 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 2 12:55:07.815287 kubelet[3067]: I0302 12:55:07.815275 3067 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 12:55:07.815570 kubelet[3067]: I0302 12:55:07.815557 3067 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 12:55:07.831341 kubelet[3067]: I0302 12:55:07.831318 3067 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 12:55:07.832001 kubelet[3067]: E0302 12:55:07.831529 3067 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://10.200.20.30:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.200.20.30:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Mar 2 12:55:07.840306 kubelet[3067]: I0302 12:55:07.840292 3067 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 12:55:07.843622 kubelet[3067]: I0302 12:55:07.843606 3067 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 2 12:55:07.845158 kubelet[3067]: I0302 12:55:07.845133 3067 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 12:55:07.845377 kubelet[3067]: I0302 12:55:07.845264 3067 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459.2.101-f64d7d9b2a","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 12:55:07.845520 kubelet[3067]: I0302 12:55:07.845509 3067 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 12:55:07.845611 kubelet[3067]: I0302 12:55:07.845587 3067 container_manager_linux.go:303] "Creating device plugin manager" Mar 2 12:55:07.845769 kubelet[3067]: I0302 12:55:07.845758 3067 state_mem.go:36] "Initialized new in-memory state store" Mar 2 12:55:07.848257 kubelet[3067]: I0302 12:55:07.848243 3067 kubelet.go:480] "Attempting to sync node with API server" Mar 2 12:55:07.848336 kubelet[3067]: I0302 12:55:07.848327 3067 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 12:55:07.848397 kubelet[3067]: I0302 12:55:07.848390 3067 kubelet.go:386] "Adding apiserver pod source" Mar 2 12:55:07.849411 kubelet[3067]: I0302 12:55:07.849390 3067 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 12:55:07.852425 kubelet[3067]: E0302 12:55:07.852387 3067 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://10.200.20.30:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4459.2.101-f64d7d9b2a&limit=500&resourceVersion=0\": dial tcp 10.200.20.30:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Mar 2 12:55:07.852501 kubelet[3067]: I0302 12:55:07.852486 3067 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 2 12:55:07.852837 kubelet[3067]: I0302 12:55:07.852823 3067 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 12:55:07.852885 kubelet[3067]: W0302 12:55:07.852872 3067 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 2 12:55:07.854718 kubelet[3067]: I0302 12:55:07.854695 3067 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 2 12:55:07.854776 kubelet[3067]: I0302 12:55:07.854735 3067 server.go:1289] "Started kubelet" Mar 2 12:55:07.856306 kubelet[3067]: E0302 12:55:07.856284 3067 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://10.200.20.30:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.200.20.30:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Mar 2 12:55:07.856520 kubelet[3067]: I0302 12:55:07.856485 3067 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 12:55:07.857142 kubelet[3067]: I0302 12:55:07.857127 3067 server.go:317] "Adding debug handlers to kubelet server" Mar 2 12:55:07.858199 kubelet[3067]: I0302 12:55:07.858147 3067 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 12:55:07.858467 kubelet[3067]: I0302 12:55:07.858448 3067 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 12:55:07.859423 kubelet[3067]: E0302 12:55:07.858541 3067 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.30:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.30:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4459.2.101-f64d7d9b2a.1899077070a6c69c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4459.2.101-f64d7d9b2a,UID:ci-4459.2.101-f64d7d9b2a,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4459.2.101-f64d7d9b2a,},FirstTimestamp:2026-03-02 12:55:07.854710428 +0000 UTC m=+0.347019511,LastTimestamp:2026-03-02 12:55:07.854710428 +0000 UTC m=+0.347019511,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4459.2.101-f64d7d9b2a,}" Mar 2 12:55:07.860524 kubelet[3067]: I0302 12:55:07.860501 3067 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 12:55:07.862008 kubelet[3067]: I0302 12:55:07.860800 3067 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 12:55:07.863134 kubelet[3067]: E0302 12:55:07.863121 3067 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 12:55:07.863394 kubelet[3067]: E0302 12:55:07.863361 3067 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" Mar 2 12:55:07.863490 kubelet[3067]: I0302 12:55:07.863481 3067 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 2 12:55:07.863670 kubelet[3067]: I0302 12:55:07.863656 3067 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 2 12:55:07.863833 kubelet[3067]: I0302 12:55:07.863824 3067 reconciler.go:26] "Reconciler: start to sync state" Mar 2 12:55:07.864996 kubelet[3067]: I0302 12:55:07.864983 3067 factory.go:223] Registration of the systemd container factory successfully Mar 2 12:55:07.865150 kubelet[3067]: I0302 12:55:07.865135 3067 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 12:55:07.865497 kubelet[3067]: E0302 12:55:07.865480 3067 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://10.200.20.30:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.200.20.30:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Mar 2 12:55:07.866304 kubelet[3067]: I0302 12:55:07.866291 3067 factory.go:223] Registration of the containerd container factory successfully Mar 2 12:55:07.887155 kubelet[3067]: E0302 12:55:07.887112 3067 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.101-f64d7d9b2a?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="200ms" Mar 2 12:55:07.890310 kubelet[3067]: I0302 12:55:07.890292 3067 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 12:55:07.890310 kubelet[3067]: I0302 12:55:07.890304 3067 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 12:55:07.890474 kubelet[3067]: I0302 12:55:07.890318 3067 state_mem.go:36] "Initialized new in-memory state store" Mar 2 12:55:07.964524 kubelet[3067]: E0302 12:55:07.964479 3067 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" Mar 2 12:55:08.046589 kubelet[3067]: I0302 12:55:08.046347 3067 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 2 12:55:08.047491 kubelet[3067]: I0302 12:55:08.047158 3067 policy_none.go:49] "None policy: Start" Mar 2 12:55:08.047491 kubelet[3067]: I0302 12:55:08.047188 3067 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 2 12:55:08.047491 kubelet[3067]: I0302 12:55:08.047198 3067 state_mem.go:35] "Initializing new in-memory state store" Mar 2 12:55:08.047491 kubelet[3067]: I0302 12:55:08.047444 3067 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 2 12:55:08.047594 kubelet[3067]: I0302 12:55:08.047516 3067 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 2 12:55:08.047594 kubelet[3067]: I0302 12:55:08.047537 3067 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 12:55:08.047594 kubelet[3067]: I0302 12:55:08.047543 3067 kubelet.go:2436] "Starting kubelet main sync loop" Mar 2 12:55:08.047676 kubelet[3067]: E0302 12:55:08.047654 3067 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 12:55:08.049161 kubelet[3067]: E0302 12:55:08.049130 3067 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://10.200.20.30:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.200.20.30:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Mar 2 12:55:08.064542 kubelet[3067]: E0302 12:55:08.064526 3067 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" Mar 2 12:55:08.066794 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Mar 2 12:55:08.075806 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Mar 2 12:55:08.078686 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Mar 2 12:55:08.087974 kubelet[3067]: E0302 12:55:08.087945 3067 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.101-f64d7d9b2a?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="400ms" Mar 2 12:55:08.088147 kubelet[3067]: E0302 12:55:08.088065 3067 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 12:55:08.088258 kubelet[3067]: I0302 12:55:08.088220 3067 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 12:55:08.088495 kubelet[3067]: I0302 12:55:08.088236 3067 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 12:55:08.088634 kubelet[3067]: I0302 12:55:08.088616 3067 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 12:55:08.090497 kubelet[3067]: E0302 12:55:08.090442 3067 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 12:55:08.090497 kubelet[3067]: E0302 12:55:08.090475 3067 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4459.2.101-f64d7d9b2a\" not found" Mar 2 12:55:08.160931 systemd[1]: Created slice kubepods-burstable-podba983be96769664dba45422d6e3f75ec.slice - libcontainer container kubepods-burstable-podba983be96769664dba45422d6e3f75ec.slice. Mar 2 12:55:08.168553 kubelet[3067]: E0302 12:55:08.168467 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.171327 systemd[1]: Created slice kubepods-burstable-pod4ac4fd9df0728a504f752485c4a74e8c.slice - libcontainer container kubepods-burstable-pod4ac4fd9df0728a504f752485c4a74e8c.slice. Mar 2 12:55:08.174282 kubelet[3067]: E0302 12:55:08.174036 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.184561 systemd[1]: Created slice kubepods-burstable-podf6be864e25f17e0423e7d96ea3571b69.slice - libcontainer container kubepods-burstable-podf6be864e25f17e0423e7d96ea3571b69.slice. Mar 2 12:55:08.185830 kubelet[3067]: E0302 12:55:08.185816 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.189708 kubelet[3067]: I0302 12:55:08.189687 3067 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.190015 kubelet[3067]: E0302 12:55:08.189994 3067 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265575 kubelet[3067]: I0302 12:55:08.265439 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-k8s-certs\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265575 kubelet[3067]: I0302 12:55:08.265466 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ba983be96769664dba45422d6e3f75ec-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" (UID: \"ba983be96769664dba45422d6e3f75ec\") " pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265575 kubelet[3067]: I0302 12:55:08.265481 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-ca-certs\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265575 kubelet[3067]: I0302 12:55:08.265492 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-kubeconfig\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265575 kubelet[3067]: I0302 12:55:08.265503 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265719 kubelet[3067]: I0302 12:55:08.265515 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f6be864e25f17e0423e7d96ea3571b69-kubeconfig\") pod \"kube-scheduler-ci-4459.2.101-f64d7d9b2a\" (UID: \"f6be864e25f17e0423e7d96ea3571b69\") " pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265719 kubelet[3067]: I0302 12:55:08.265524 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ba983be96769664dba45422d6e3f75ec-ca-certs\") pod \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" (UID: \"ba983be96769664dba45422d6e3f75ec\") " pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265719 kubelet[3067]: I0302 12:55:08.265532 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ba983be96769664dba45422d6e3f75ec-k8s-certs\") pod \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" (UID: \"ba983be96769664dba45422d6e3f75ec\") " pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.265719 kubelet[3067]: I0302 12:55:08.265559 3067 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-flexvolume-dir\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.392240 kubelet[3067]: I0302 12:55:08.391926 3067 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.392645 kubelet[3067]: E0302 12:55:08.392620 3067 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.471532 containerd[1890]: time="2026-03-02T12:55:08.471493949Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459.2.101-f64d7d9b2a,Uid:ba983be96769664dba45422d6e3f75ec,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:08.475044 containerd[1890]: time="2026-03-02T12:55:08.474903302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459.2.101-f64d7d9b2a,Uid:4ac4fd9df0728a504f752485c4a74e8c,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:08.486562 containerd[1890]: time="2026-03-02T12:55:08.486534859Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459.2.101-f64d7d9b2a,Uid:f6be864e25f17e0423e7d96ea3571b69,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:08.489163 kubelet[3067]: E0302 12:55:08.489133 3067 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.30:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4459.2.101-f64d7d9b2a?timeout=10s\": dial tcp 10.200.20.30:6443: connect: connection refused" interval="800ms" Mar 2 12:55:08.563257 containerd[1890]: time="2026-03-02T12:55:08.563219089Z" level=info msg="connecting to shim de60cd9ec6566576d54024cc9c20a9bad3de352785a41a36bb3c606e9b243fe9" address="unix:///run/containerd/s/1f9440a98799fea3e38987c624c42f7f9794691b8c6eb06541c00845253a9be0" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:08.581525 systemd[1]: Started cri-containerd-de60cd9ec6566576d54024cc9c20a9bad3de352785a41a36bb3c606e9b243fe9.scope - libcontainer container de60cd9ec6566576d54024cc9c20a9bad3de352785a41a36bb3c606e9b243fe9. Mar 2 12:55:08.587472 containerd[1890]: time="2026-03-02T12:55:08.587428541Z" level=info msg="connecting to shim d05720edf9d8978aef4b5e74a2e0532d7750fe223706ed5c8a4f4188d36c996c" address="unix:///run/containerd/s/6fa989cea3d3706e69b0613ebdadf4fa896d5197178b805d802d72eba63373d6" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:08.593433 containerd[1890]: time="2026-03-02T12:55:08.592862886Z" level=info msg="connecting to shim 2d20e9e61bc6a449baf449d004d4d67949717f9e6a015921739b7b223740d24c" address="unix:///run/containerd/s/470272d77aed5fccd8f395896cbd4655c209857f7a8b411d3532aa160478c0b0" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:08.620635 systemd[1]: Started cri-containerd-2d20e9e61bc6a449baf449d004d4d67949717f9e6a015921739b7b223740d24c.scope - libcontainer container 2d20e9e61bc6a449baf449d004d4d67949717f9e6a015921739b7b223740d24c. Mar 2 12:55:08.624753 systemd[1]: Started cri-containerd-d05720edf9d8978aef4b5e74a2e0532d7750fe223706ed5c8a4f4188d36c996c.scope - libcontainer container d05720edf9d8978aef4b5e74a2e0532d7750fe223706ed5c8a4f4188d36c996c. Mar 2 12:55:08.641506 containerd[1890]: time="2026-03-02T12:55:08.641478118Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4459.2.101-f64d7d9b2a,Uid:ba983be96769664dba45422d6e3f75ec,Namespace:kube-system,Attempt:0,} returns sandbox id \"de60cd9ec6566576d54024cc9c20a9bad3de352785a41a36bb3c606e9b243fe9\"" Mar 2 12:55:08.656292 containerd[1890]: time="2026-03-02T12:55:08.656213586Z" level=info msg="CreateContainer within sandbox \"de60cd9ec6566576d54024cc9c20a9bad3de352785a41a36bb3c606e9b243fe9\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 2 12:55:08.673680 containerd[1890]: time="2026-03-02T12:55:08.673596908Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4459.2.101-f64d7d9b2a,Uid:f6be864e25f17e0423e7d96ea3571b69,Namespace:kube-system,Attempt:0,} returns sandbox id \"2d20e9e61bc6a449baf449d004d4d67949717f9e6a015921739b7b223740d24c\"" Mar 2 12:55:08.677284 containerd[1890]: time="2026-03-02T12:55:08.677259366Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4459.2.101-f64d7d9b2a,Uid:4ac4fd9df0728a504f752485c4a74e8c,Namespace:kube-system,Attempt:0,} returns sandbox id \"d05720edf9d8978aef4b5e74a2e0532d7750fe223706ed5c8a4f4188d36c996c\"" Mar 2 12:55:08.680325 containerd[1890]: time="2026-03-02T12:55:08.679838234Z" level=info msg="CreateContainer within sandbox \"2d20e9e61bc6a449baf449d004d4d67949717f9e6a015921739b7b223740d24c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 2 12:55:08.688401 containerd[1890]: time="2026-03-02T12:55:08.688373121Z" level=info msg="CreateContainer within sandbox \"d05720edf9d8978aef4b5e74a2e0532d7750fe223706ed5c8a4f4188d36c996c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 2 12:55:08.696229 containerd[1890]: time="2026-03-02T12:55:08.696203431Z" level=info msg="Container 8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:08.749222 containerd[1890]: time="2026-03-02T12:55:08.749178154Z" level=info msg="CreateContainer within sandbox \"de60cd9ec6566576d54024cc9c20a9bad3de352785a41a36bb3c606e9b243fe9\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c\"" Mar 2 12:55:08.749885 containerd[1890]: time="2026-03-02T12:55:08.749857898Z" level=info msg="StartContainer for \"8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c\"" Mar 2 12:55:08.755283 containerd[1890]: time="2026-03-02T12:55:08.755259226Z" level=info msg="connecting to shim 8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c" address="unix:///run/containerd/s/1f9440a98799fea3e38987c624c42f7f9794691b8c6eb06541c00845253a9be0" protocol=ttrpc version=3 Mar 2 12:55:08.760930 containerd[1890]: time="2026-03-02T12:55:08.760907899Z" level=info msg="Container d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:08.769356 containerd[1890]: time="2026-03-02T12:55:08.768998227Z" level=info msg="Container ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:08.772531 systemd[1]: Started cri-containerd-8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c.scope - libcontainer container 8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c. Mar 2 12:55:08.790976 containerd[1890]: time="2026-03-02T12:55:08.790906117Z" level=info msg="CreateContainer within sandbox \"2d20e9e61bc6a449baf449d004d4d67949717f9e6a015921739b7b223740d24c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa\"" Mar 2 12:55:08.791497 containerd[1890]: time="2026-03-02T12:55:08.791395863Z" level=info msg="StartContainer for \"ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa\"" Mar 2 12:55:08.794640 kubelet[3067]: I0302 12:55:08.794614 3067 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.795214 kubelet[3067]: E0302 12:55:08.795177 3067 kubelet_node_status.go:107] "Unable to register node with API server" err="Post \"https://10.200.20.30:6443/api/v1/nodes\": dial tcp 10.200.20.30:6443: connect: connection refused" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:08.799064 containerd[1890]: time="2026-03-02T12:55:08.798913826Z" level=info msg="CreateContainer within sandbox \"d05720edf9d8978aef4b5e74a2e0532d7750fe223706ed5c8a4f4188d36c996c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174\"" Mar 2 12:55:08.800013 containerd[1890]: time="2026-03-02T12:55:08.799992368Z" level=info msg="connecting to shim ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa" address="unix:///run/containerd/s/470272d77aed5fccd8f395896cbd4655c209857f7a8b411d3532aa160478c0b0" protocol=ttrpc version=3 Mar 2 12:55:08.803885 containerd[1890]: time="2026-03-02T12:55:08.800030090Z" level=info msg="StartContainer for \"d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174\"" Mar 2 12:55:08.804925 containerd[1890]: time="2026-03-02T12:55:08.804893734Z" level=info msg="connecting to shim d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174" address="unix:///run/containerd/s/6fa989cea3d3706e69b0613ebdadf4fa896d5197178b805d802d72eba63373d6" protocol=ttrpc version=3 Mar 2 12:55:08.819942 containerd[1890]: time="2026-03-02T12:55:08.819918956Z" level=info msg="StartContainer for \"8554c68988492281e920970151d596be719f3736f086af038d5c820850bba64c\" returns successfully" Mar 2 12:55:08.820617 systemd[1]: Started cri-containerd-ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa.scope - libcontainer container ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa. Mar 2 12:55:08.834707 systemd[1]: Started cri-containerd-d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174.scope - libcontainer container d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174. Mar 2 12:55:08.878585 containerd[1890]: time="2026-03-02T12:55:08.878526135Z" level=info msg="StartContainer for \"ba5d72c43794d471d6e5155de1755ffda66f667254f05f52543f44b4d3cdc3aa\" returns successfully" Mar 2 12:55:08.899620 containerd[1890]: time="2026-03-02T12:55:08.899585524Z" level=info msg="StartContainer for \"d42a30e043769a4fad3426223353939339d8b5fd89b100338249f98580696174\" returns successfully" Mar 2 12:55:09.056775 kubelet[3067]: E0302 12:55:09.056722 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:09.058988 kubelet[3067]: E0302 12:55:09.058963 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:09.061263 kubelet[3067]: E0302 12:55:09.061247 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:09.597586 kubelet[3067]: I0302 12:55:09.597191 3067 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.062464 kubelet[3067]: E0302 12:55:10.062327 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.062464 kubelet[3067]: E0302 12:55:10.062392 3067 kubelet.go:3305] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.208037 kubelet[3067]: E0302 12:55:10.208004 3067 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4459.2.101-f64d7d9b2a\" not found" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.229504 kubelet[3067]: I0302 12:55:10.229473 3067 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.267775 kubelet[3067]: I0302 12:55:10.267746 3067 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.273245 kubelet[3067]: E0302 12:55:10.273217 3067 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.101-f64d7d9b2a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.273245 kubelet[3067]: I0302 12:55:10.273242 3067 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.275661 kubelet[3067]: E0302 12:55:10.275638 3067 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.275730 kubelet[3067]: I0302 12:55:10.275685 3067 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.276818 kubelet[3067]: E0302 12:55:10.276794 3067 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:10.859116 kubelet[3067]: I0302 12:55:10.859073 3067 apiserver.go:52] "Watching apiserver" Mar 2 12:55:10.864079 kubelet[3067]: I0302 12:55:10.864059 3067 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 2 12:55:11.062580 kubelet[3067]: I0302 12:55:11.062554 3067 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:11.065530 kubelet[3067]: E0302 12:55:11.065379 3067 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:12.525376 systemd[1]: Reload requested from client PID 3352 ('systemctl') (unit session-9.scope)... Mar 2 12:55:12.525398 systemd[1]: Reloading... Mar 2 12:55:12.598003 zram_generator::config[3402]: No configuration found. Mar 2 12:55:12.703538 kubelet[3067]: I0302 12:55:12.703505 3067 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:12.711296 kubelet[3067]: I0302 12:55:12.711272 3067 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:55:12.762156 systemd[1]: Reloading finished in 236 ms. Mar 2 12:55:12.782865 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:55:12.800142 systemd[1]: kubelet.service: Deactivated successfully. Mar 2 12:55:12.800376 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:55:12.800444 systemd[1]: kubelet.service: Consumed 594ms CPU time, 127.8M memory peak. Mar 2 12:55:12.802364 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Mar 2 12:55:12.910568 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Mar 2 12:55:12.918659 (kubelet)[3463]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Mar 2 12:55:12.945476 kubelet[3463]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 12:55:12.945930 kubelet[3463]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Mar 2 12:55:12.945970 kubelet[3463]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 2 12:55:12.946078 kubelet[3463]: I0302 12:55:12.946046 3463 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 2 12:55:12.950908 kubelet[3463]: I0302 12:55:12.950876 3463 server.go:530] "Kubelet version" kubeletVersion="v1.33.8" Mar 2 12:55:12.950908 kubelet[3463]: I0302 12:55:12.950904 3463 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 2 12:55:12.951074 kubelet[3463]: I0302 12:55:12.951055 3463 server.go:956] "Client rotation is on, will bootstrap in background" Mar 2 12:55:12.951970 kubelet[3463]: I0302 12:55:12.951948 3463 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Mar 2 12:55:12.954031 kubelet[3463]: I0302 12:55:12.953992 3463 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 2 12:55:12.956803 kubelet[3463]: I0302 12:55:12.956784 3463 server.go:1446] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 2 12:55:12.959635 kubelet[3463]: I0302 12:55:12.959608 3463 server.go:782] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 2 12:55:12.961539 kubelet[3463]: I0302 12:55:12.961506 3463 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 2 12:55:12.961705 kubelet[3463]: I0302 12:55:12.961536 3463 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4459.2.101-f64d7d9b2a","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 2 12:55:12.961772 kubelet[3463]: I0302 12:55:12.961706 3463 topology_manager.go:138] "Creating topology manager with none policy" Mar 2 12:55:12.961772 kubelet[3463]: I0302 12:55:12.961713 3463 container_manager_linux.go:303] "Creating device plugin manager" Mar 2 12:55:12.961772 kubelet[3463]: I0302 12:55:12.961758 3463 state_mem.go:36] "Initialized new in-memory state store" Mar 2 12:55:12.961881 kubelet[3463]: I0302 12:55:12.961866 3463 kubelet.go:480] "Attempting to sync node with API server" Mar 2 12:55:12.961901 kubelet[3463]: I0302 12:55:12.961884 3463 kubelet.go:375] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 2 12:55:12.961924 kubelet[3463]: I0302 12:55:12.961918 3463 kubelet.go:386] "Adding apiserver pod source" Mar 2 12:55:12.961941 kubelet[3463]: I0302 12:55:12.961934 3463 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 2 12:55:12.967216 kubelet[3463]: I0302 12:55:12.965764 3463 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="containerd" version="v2.0.7" apiVersion="v1" Mar 2 12:55:12.967216 kubelet[3463]: I0302 12:55:12.966121 3463 kubelet.go:935] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Mar 2 12:55:12.969758 kubelet[3463]: I0302 12:55:12.969739 3463 watchdog_linux.go:99] "Systemd watchdog is not enabled" Mar 2 12:55:12.969818 kubelet[3463]: I0302 12:55:12.969775 3463 server.go:1289] "Started kubelet" Mar 2 12:55:12.972727 kubelet[3463]: I0302 12:55:12.972697 3463 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 2 12:55:12.973240 kubelet[3463]: E0302 12:55:12.973223 3463 kubelet.go:1600] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 2 12:55:12.979625 kubelet[3463]: I0302 12:55:12.979570 3463 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Mar 2 12:55:12.981644 kubelet[3463]: I0302 12:55:12.981626 3463 server.go:317] "Adding debug handlers to kubelet server" Mar 2 12:55:12.985001 kubelet[3463]: I0302 12:55:12.984962 3463 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 2 12:55:12.985311 kubelet[3463]: I0302 12:55:12.985295 3463 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 2 12:55:12.985714 kubelet[3463]: I0302 12:55:12.985697 3463 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Mar 2 12:55:12.986388 kubelet[3463]: I0302 12:55:12.986363 3463 volume_manager.go:297] "Starting Kubelet Volume Manager" Mar 2 12:55:12.986671 kubelet[3463]: E0302 12:55:12.986652 3463 kubelet_node_status.go:466] "Error getting the current node from lister" err="node \"ci-4459.2.101-f64d7d9b2a\" not found" Mar 2 12:55:12.988881 kubelet[3463]: I0302 12:55:12.988860 3463 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Mar 2 12:55:12.988972 kubelet[3463]: I0302 12:55:12.988959 3463 reconciler.go:26] "Reconciler: start to sync state" Mar 2 12:55:12.991348 kubelet[3463]: I0302 12:55:12.991310 3463 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Mar 2 12:55:12.992583 kubelet[3463]: I0302 12:55:12.992568 3463 factory.go:223] Registration of the containerd container factory successfully Mar 2 12:55:12.993319 kubelet[3463]: I0302 12:55:12.993297 3463 factory.go:223] Registration of the systemd container factory successfully Mar 2 12:55:12.993503 kubelet[3463]: I0302 12:55:12.993485 3463 factory.go:221] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 2 12:55:12.994522 kubelet[3463]: I0302 12:55:12.994500 3463 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Mar 2 12:55:12.994522 kubelet[3463]: I0302 12:55:12.994526 3463 status_manager.go:230] "Starting to sync pod status with apiserver" Mar 2 12:55:12.994596 kubelet[3463]: I0302 12:55:12.994542 3463 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Mar 2 12:55:12.994596 kubelet[3463]: I0302 12:55:12.994548 3463 kubelet.go:2436] "Starting kubelet main sync loop" Mar 2 12:55:12.994596 kubelet[3463]: E0302 12:55:12.994579 3463 kubelet.go:2460] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033462 3463 cpu_manager.go:221] "Starting CPU manager" policy="none" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033554 3463 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033598 3463 state_mem.go:36] "Initialized new in-memory state store" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033725 3463 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033733 3463 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033760 3463 policy_none.go:49] "None policy: Start" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033769 3463 memory_manager.go:186] "Starting memorymanager" policy="None" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033777 3463 state_mem.go:35] "Initializing new in-memory state store" Mar 2 12:55:13.034325 kubelet[3463]: I0302 12:55:13.033879 3463 state_mem.go:75] "Updated machine memory state" Mar 2 12:55:13.037755 kubelet[3463]: E0302 12:55:13.037679 3463 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Mar 2 12:55:13.095070 kubelet[3463]: E0302 12:55:13.095024 3463 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 2 12:55:13.295753 kubelet[3463]: E0302 12:55:13.295642 3463 kubelet.go:2460] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Mar 2 12:55:13.324951 kubelet[3463]: I0302 12:55:13.324684 3463 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 2 12:55:13.324951 kubelet[3463]: I0302 12:55:13.324714 3463 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 2 12:55:13.325741 kubelet[3463]: I0302 12:55:13.325558 3463 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 2 12:55:13.326746 kubelet[3463]: E0302 12:55:13.326683 3463 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Mar 2 12:55:13.430331 kubelet[3463]: I0302 12:55:13.430295 3463 kubelet_node_status.go:75] "Attempting to register node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.442599 kubelet[3463]: I0302 12:55:13.442548 3463 kubelet_node_status.go:124] "Node was previously registered" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.442791 kubelet[3463]: I0302 12:55:13.442639 3463 kubelet_node_status.go:78] "Successfully registered node" node="ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.588293 sudo[3501]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 2 12:55:13.588629 sudo[3501]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Mar 2 12:55:13.697777 kubelet[3463]: I0302 12:55:13.697746 3463 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.699080 kubelet[3463]: I0302 12:55:13.698333 3463 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.699080 kubelet[3463]: I0302 12:55:13.698557 3463 kubelet.go:3309] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.711752 kubelet[3463]: I0302 12:55:13.711727 3463 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:55:13.711911 kubelet[3463]: I0302 12:55:13.711888 3463 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:55:13.711992 kubelet[3463]: I0302 12:55:13.711978 3463 warnings.go:110] "Warning: metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots]" Mar 2 12:55:13.712024 kubelet[3463]: E0302 12:55:13.712011 3463 kubelet.go:3311] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4459.2.101-f64d7d9b2a\" already exists" pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793021 kubelet[3463]: I0302 12:55:13.792982 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/ba983be96769664dba45422d6e3f75ec-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" (UID: \"ba983be96769664dba45422d6e3f75ec\") " pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793747 kubelet[3463]: I0302 12:55:13.793654 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-ca-certs\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793747 kubelet[3463]: I0302 12:55:13.793697 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-flexvolume-dir\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793747 kubelet[3463]: I0302 12:55:13.793715 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/ba983be96769664dba45422d6e3f75ec-k8s-certs\") pod \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" (UID: \"ba983be96769664dba45422d6e3f75ec\") " pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793747 kubelet[3463]: I0302 12:55:13.793730 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-k8s-certs\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793747 kubelet[3463]: I0302 12:55:13.793744 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-kubeconfig\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793879 kubelet[3463]: I0302 12:55:13.793764 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4ac4fd9df0728a504f752485c4a74e8c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4459.2.101-f64d7d9b2a\" (UID: \"4ac4fd9df0728a504f752485c4a74e8c\") " pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793879 kubelet[3463]: I0302 12:55:13.793775 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f6be864e25f17e0423e7d96ea3571b69-kubeconfig\") pod \"kube-scheduler-ci-4459.2.101-f64d7d9b2a\" (UID: \"f6be864e25f17e0423e7d96ea3571b69\") " pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.793879 kubelet[3463]: I0302 12:55:13.793794 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/ba983be96769664dba45422d6e3f75ec-ca-certs\") pod \"kube-apiserver-ci-4459.2.101-f64d7d9b2a\" (UID: \"ba983be96769664dba45422d6e3f75ec\") " pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" Mar 2 12:55:13.831551 sudo[3501]: pam_unix(sudo:session): session closed for user root Mar 2 12:55:13.964866 kubelet[3463]: I0302 12:55:13.964598 3463 apiserver.go:52] "Watching apiserver" Mar 2 12:55:13.989390 kubelet[3463]: I0302 12:55:13.989354 3463 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Mar 2 12:55:14.049770 kubelet[3463]: I0302 12:55:14.049485 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4459.2.101-f64d7d9b2a" podStartSLOduration=1.049468807 podStartE2EDuration="1.049468807s" podCreationTimestamp="2026-03-02 12:55:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:55:14.039265534 +0000 UTC m=+1.117002834" watchObservedRunningTime="2026-03-02 12:55:14.049468807 +0000 UTC m=+1.127206091" Mar 2 12:55:14.059634 kubelet[3463]: I0302 12:55:14.059543 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4459.2.101-f64d7d9b2a" podStartSLOduration=1.059526325 podStartE2EDuration="1.059526325s" podCreationTimestamp="2026-03-02 12:55:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:55:14.050336433 +0000 UTC m=+1.128073733" watchObservedRunningTime="2026-03-02 12:55:14.059526325 +0000 UTC m=+1.137263609" Mar 2 12:55:14.073109 kubelet[3463]: I0302 12:55:14.072602 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4459.2.101-f64d7d9b2a" podStartSLOduration=2.072589299 podStartE2EDuration="2.072589299s" podCreationTimestamp="2026-03-02 12:55:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:55:14.060168808 +0000 UTC m=+1.137906124" watchObservedRunningTime="2026-03-02 12:55:14.072589299 +0000 UTC m=+1.150326583" Mar 2 12:55:14.970586 sudo[2365]: pam_unix(sudo:session): session closed for user root Mar 2 12:55:15.047262 sshd[2364]: Connection closed by 10.200.16.10 port 47050 Mar 2 12:55:15.048700 sshd-session[2361]: pam_unix(sshd:session): session closed for user core Mar 2 12:55:15.051697 systemd-logind[1868]: Session 9 logged out. Waiting for processes to exit. Mar 2 12:55:15.052208 systemd[1]: sshd@6-10.200.20.30:22-10.200.16.10:47050.service: Deactivated successfully. Mar 2 12:55:15.054059 systemd[1]: session-9.scope: Deactivated successfully. Mar 2 12:55:15.054291 systemd[1]: session-9.scope: Consumed 3.638s CPU time, 262.1M memory peak. Mar 2 12:55:15.056353 systemd-logind[1868]: Removed session 9. Mar 2 12:55:18.281662 kubelet[3463]: I0302 12:55:18.281540 3463 kuberuntime_manager.go:1746] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 2 12:55:18.282637 containerd[1890]: time="2026-03-02T12:55:18.282225409Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 2 12:55:18.282855 kubelet[3463]: I0302 12:55:18.282393 3463 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 2 12:55:19.332713 systemd[1]: Created slice kubepods-besteffort-pod90173a87_27c6_4614_8b97_f51f042078d7.slice - libcontainer container kubepods-besteffort-pod90173a87_27c6_4614_8b97_f51f042078d7.slice. Mar 2 12:55:19.347087 systemd[1]: Created slice kubepods-burstable-pod8defe60e_7a55_4ade_b1cc_01988615f84e.slice - libcontainer container kubepods-burstable-pod8defe60e_7a55_4ade_b1cc_01988615f84e.slice. Mar 2 12:55:19.426808 kubelet[3463]: I0302 12:55:19.426771 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46vmb\" (UniqueName: \"kubernetes.io/projected/90173a87-27c6-4614-8b97-f51f042078d7-kube-api-access-46vmb\") pod \"kube-proxy-p24m4\" (UID: \"90173a87-27c6-4614-8b97-f51f042078d7\") " pod="kube-system/kube-proxy-p24m4" Mar 2 12:55:19.427286 kubelet[3463]: I0302 12:55:19.426963 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-lib-modules\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427286 kubelet[3463]: I0302 12:55:19.426982 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-config-path\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427286 kubelet[3463]: I0302 12:55:19.426993 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-kernel\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427286 kubelet[3463]: I0302 12:55:19.427002 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nkb4\" (UniqueName: \"kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-kube-api-access-2nkb4\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427662 kubelet[3463]: I0302 12:55:19.427455 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/90173a87-27c6-4614-8b97-f51f042078d7-kube-proxy\") pod \"kube-proxy-p24m4\" (UID: \"90173a87-27c6-4614-8b97-f51f042078d7\") " pod="kube-system/kube-proxy-p24m4" Mar 2 12:55:19.427662 kubelet[3463]: I0302 12:55:19.427482 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-hostproc\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427662 kubelet[3463]: I0302 12:55:19.427499 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-etc-cni-netd\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427662 kubelet[3463]: I0302 12:55:19.427534 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-xtables-lock\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427662 kubelet[3463]: I0302 12:55:19.427546 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/90173a87-27c6-4614-8b97-f51f042078d7-lib-modules\") pod \"kube-proxy-p24m4\" (UID: \"90173a87-27c6-4614-8b97-f51f042078d7\") " pod="kube-system/kube-proxy-p24m4" Mar 2 12:55:19.427662 kubelet[3463]: I0302 12:55:19.427556 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-run\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427795 kubelet[3463]: I0302 12:55:19.427568 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cni-path\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427795 kubelet[3463]: I0302 12:55:19.427577 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8defe60e-7a55-4ade-b1cc-01988615f84e-clustermesh-secrets\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427795 kubelet[3463]: I0302 12:55:19.427610 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-net\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427795 kubelet[3463]: I0302 12:55:19.427621 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/90173a87-27c6-4614-8b97-f51f042078d7-xtables-lock\") pod \"kube-proxy-p24m4\" (UID: \"90173a87-27c6-4614-8b97-f51f042078d7\") " pod="kube-system/kube-proxy-p24m4" Mar 2 12:55:19.427795 kubelet[3463]: I0302 12:55:19.427630 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-bpf-maps\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427795 kubelet[3463]: I0302 12:55:19.427638 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-cgroup\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.427882 kubelet[3463]: I0302 12:55:19.427646 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-hubble-tls\") pod \"cilium-6pf4l\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " pod="kube-system/cilium-6pf4l" Mar 2 12:55:19.496488 systemd[1]: Created slice kubepods-besteffort-pod4df0e96b_7b35_42fc_b5d0_ec0b8cd65ba8.slice - libcontainer container kubepods-besteffort-pod4df0e96b_7b35_42fc_b5d0_ec0b8cd65ba8.slice. Mar 2 12:55:19.529570 kubelet[3463]: I0302 12:55:19.528523 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-znz4j\" (UID: \"4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8\") " pod="kube-system/cilium-operator-6c4d7847fc-znz4j" Mar 2 12:55:19.529570 kubelet[3463]: I0302 12:55:19.528596 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzc6z\" (UniqueName: \"kubernetes.io/projected/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-kube-api-access-kzc6z\") pod \"cilium-operator-6c4d7847fc-znz4j\" (UID: \"4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8\") " pod="kube-system/cilium-operator-6c4d7847fc-znz4j" Mar 2 12:55:19.642061 containerd[1890]: time="2026-03-02T12:55:19.641976962Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-p24m4,Uid:90173a87-27c6-4614-8b97-f51f042078d7,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:19.650729 containerd[1890]: time="2026-03-02T12:55:19.650699344Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6pf4l,Uid:8defe60e-7a55-4ade-b1cc-01988615f84e,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:19.726458 containerd[1890]: time="2026-03-02T12:55:19.726417860Z" level=info msg="connecting to shim 3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf" address="unix:///run/containerd/s/6ff8dcc12b6784907ab387cd008f4a7cbe79ef07dbd7e04e064767f6c012e506" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:19.728061 containerd[1890]: time="2026-03-02T12:55:19.728013668Z" level=info msg="connecting to shim c302b4d1116fca529e798bd294abdde1197b1e78fc5ec52923575e79d6b5db18" address="unix:///run/containerd/s/a68e3beed16b1932c673b4195e24c6665f0b040268487b7e33aa00e893bf6226" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:19.749544 systemd[1]: Started cri-containerd-3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf.scope - libcontainer container 3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf. Mar 2 12:55:19.752383 systemd[1]: Started cri-containerd-c302b4d1116fca529e798bd294abdde1197b1e78fc5ec52923575e79d6b5db18.scope - libcontainer container c302b4d1116fca529e798bd294abdde1197b1e78fc5ec52923575e79d6b5db18. Mar 2 12:55:19.780261 containerd[1890]: time="2026-03-02T12:55:19.780225910Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-p24m4,Uid:90173a87-27c6-4614-8b97-f51f042078d7,Namespace:kube-system,Attempt:0,} returns sandbox id \"c302b4d1116fca529e798bd294abdde1197b1e78fc5ec52923575e79d6b5db18\"" Mar 2 12:55:19.783763 containerd[1890]: time="2026-03-02T12:55:19.783729799Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6pf4l,Uid:8defe60e-7a55-4ade-b1cc-01988615f84e,Namespace:kube-system,Attempt:0,} returns sandbox id \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\"" Mar 2 12:55:19.784832 containerd[1890]: time="2026-03-02T12:55:19.784804271Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 2 12:55:19.789841 containerd[1890]: time="2026-03-02T12:55:19.789814030Z" level=info msg="CreateContainer within sandbox \"c302b4d1116fca529e798bd294abdde1197b1e78fc5ec52923575e79d6b5db18\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 2 12:55:19.799304 containerd[1890]: time="2026-03-02T12:55:19.799088277Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-znz4j,Uid:4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:19.821309 containerd[1890]: time="2026-03-02T12:55:19.821283120Z" level=info msg="Container dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:19.850998 containerd[1890]: time="2026-03-02T12:55:19.850964044Z" level=info msg="CreateContainer within sandbox \"c302b4d1116fca529e798bd294abdde1197b1e78fc5ec52923575e79d6b5db18\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba\"" Mar 2 12:55:19.852453 containerd[1890]: time="2026-03-02T12:55:19.851458147Z" level=info msg="StartContainer for \"dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba\"" Mar 2 12:55:19.852603 containerd[1890]: time="2026-03-02T12:55:19.852578732Z" level=info msg="connecting to shim dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba" address="unix:///run/containerd/s/a68e3beed16b1932c673b4195e24c6665f0b040268487b7e33aa00e893bf6226" protocol=ttrpc version=3 Mar 2 12:55:19.859095 containerd[1890]: time="2026-03-02T12:55:19.859015598Z" level=info msg="connecting to shim c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a" address="unix:///run/containerd/s/f11e89ea2f0d699230948686878c20aa462ebc7150a873c7d91d29d276f7d297" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:19.867833 systemd[1]: Started cri-containerd-dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba.scope - libcontainer container dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba. Mar 2 12:55:19.882657 systemd[1]: Started cri-containerd-c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a.scope - libcontainer container c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a. Mar 2 12:55:19.921489 containerd[1890]: time="2026-03-02T12:55:19.921300246Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-znz4j,Uid:4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8,Namespace:kube-system,Attempt:0,} returns sandbox id \"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\"" Mar 2 12:55:19.924572 containerd[1890]: time="2026-03-02T12:55:19.924500118Z" level=info msg="StartContainer for \"dcbe6b25384f7da14ef401671403a5653ed5feac4810661f0be9ebcd8330cdba\" returns successfully" Mar 2 12:55:20.059087 kubelet[3463]: I0302 12:55:20.059025 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-p24m4" podStartSLOduration=1.058935952 podStartE2EDuration="1.058935952s" podCreationTimestamp="2026-03-02 12:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:55:20.046523835 +0000 UTC m=+7.124261119" watchObservedRunningTime="2026-03-02 12:55:20.058935952 +0000 UTC m=+7.136673236" Mar 2 12:55:24.657247 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3428227573.mount: Deactivated successfully. Mar 2 12:55:27.597353 containerd[1890]: time="2026-03-02T12:55:27.597286271Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:27.602047 containerd[1890]: time="2026-03-02T12:55:27.601477565Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Mar 2 12:55:27.625040 containerd[1890]: time="2026-03-02T12:55:27.624793478Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:27.626014 containerd[1890]: time="2026-03-02T12:55:27.625986194Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 7.841153338s" Mar 2 12:55:27.626014 containerd[1890]: time="2026-03-02T12:55:27.626017323Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Mar 2 12:55:27.628645 containerd[1890]: time="2026-03-02T12:55:27.628298800Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 2 12:55:27.636156 containerd[1890]: time="2026-03-02T12:55:27.636120269Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 2 12:55:27.894739 containerd[1890]: time="2026-03-02T12:55:27.894144441Z" level=info msg="Container 854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:27.896197 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount813430959.mount: Deactivated successfully. Mar 2 12:55:27.939824 containerd[1890]: time="2026-03-02T12:55:27.939784973Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\"" Mar 2 12:55:27.940593 containerd[1890]: time="2026-03-02T12:55:27.940549925Z" level=info msg="StartContainer for \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\"" Mar 2 12:55:27.941459 containerd[1890]: time="2026-03-02T12:55:27.941379702Z" level=info msg="connecting to shim 854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180" address="unix:///run/containerd/s/6ff8dcc12b6784907ab387cd008f4a7cbe79ef07dbd7e04e064767f6c012e506" protocol=ttrpc version=3 Mar 2 12:55:27.960536 systemd[1]: Started cri-containerd-854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180.scope - libcontainer container 854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180. Mar 2 12:55:27.986465 containerd[1890]: time="2026-03-02T12:55:27.986431704Z" level=info msg="StartContainer for \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\" returns successfully" Mar 2 12:55:27.991528 systemd[1]: cri-containerd-854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180.scope: Deactivated successfully. Mar 2 12:55:27.995025 containerd[1890]: time="2026-03-02T12:55:27.994934681Z" level=info msg="received container exit event container_id:\"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\" id:\"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\" pid:3879 exited_at:{seconds:1772456127 nanos:993748990}" Mar 2 12:55:28.009528 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180-rootfs.mount: Deactivated successfully. Mar 2 12:55:30.065234 containerd[1890]: time="2026-03-02T12:55:30.064639406Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 2 12:55:30.119980 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3563828917.mount: Deactivated successfully. Mar 2 12:55:30.122889 containerd[1890]: time="2026-03-02T12:55:30.119950223Z" level=info msg="Container 59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:30.136107 containerd[1890]: time="2026-03-02T12:55:30.136068599Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\"" Mar 2 12:55:30.137427 containerd[1890]: time="2026-03-02T12:55:30.137339845Z" level=info msg="StartContainer for \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\"" Mar 2 12:55:30.138078 containerd[1890]: time="2026-03-02T12:55:30.138055371Z" level=info msg="connecting to shim 59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef" address="unix:///run/containerd/s/6ff8dcc12b6784907ab387cd008f4a7cbe79ef07dbd7e04e064767f6c012e506" protocol=ttrpc version=3 Mar 2 12:55:30.154536 systemd[1]: Started cri-containerd-59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef.scope - libcontainer container 59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef. Mar 2 12:55:30.185280 containerd[1890]: time="2026-03-02T12:55:30.185220525Z" level=info msg="StartContainer for \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\" returns successfully" Mar 2 12:55:30.193728 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 2 12:55:30.193887 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:55:30.194564 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Mar 2 12:55:30.196164 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Mar 2 12:55:30.199460 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 2 12:55:30.202107 systemd[1]: cri-containerd-59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef.scope: Deactivated successfully. Mar 2 12:55:30.209864 containerd[1890]: time="2026-03-02T12:55:30.209826774Z" level=info msg="received container exit event container_id:\"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\" id:\"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\" pid:3922 exited_at:{seconds:1772456130 nanos:209647264}" Mar 2 12:55:30.213555 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Mar 2 12:55:31.048371 containerd[1890]: time="2026-03-02T12:55:31.047877874Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:31.052147 containerd[1890]: time="2026-03-02T12:55:31.052120349Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Mar 2 12:55:31.056227 containerd[1890]: time="2026-03-02T12:55:31.056204355Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Mar 2 12:55:31.057743 containerd[1890]: time="2026-03-02T12:55:31.057439689Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 3.429107176s" Mar 2 12:55:31.057743 containerd[1890]: time="2026-03-02T12:55:31.057465746Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Mar 2 12:55:31.065316 containerd[1890]: time="2026-03-02T12:55:31.065288851Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 2 12:55:31.071905 containerd[1890]: time="2026-03-02T12:55:31.071720089Z" level=info msg="CreateContainer within sandbox \"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 2 12:55:31.087283 containerd[1890]: time="2026-03-02T12:55:31.087246440Z" level=info msg="Container 2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:31.106430 containerd[1890]: time="2026-03-02T12:55:31.106365309Z" level=info msg="Container 39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:31.120646 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef-rootfs.mount: Deactivated successfully. Mar 2 12:55:31.122857 containerd[1890]: time="2026-03-02T12:55:31.122817016Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\"" Mar 2 12:55:31.123576 containerd[1890]: time="2026-03-02T12:55:31.123502781Z" level=info msg="StartContainer for \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\"" Mar 2 12:55:31.125906 containerd[1890]: time="2026-03-02T12:55:31.125883295Z" level=info msg="connecting to shim 2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5" address="unix:///run/containerd/s/6ff8dcc12b6784907ab387cd008f4a7cbe79ef07dbd7e04e064767f6c012e506" protocol=ttrpc version=3 Mar 2 12:55:31.129506 containerd[1890]: time="2026-03-02T12:55:31.129399275Z" level=info msg="CreateContainer within sandbox \"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\"" Mar 2 12:55:31.131950 containerd[1890]: time="2026-03-02T12:55:31.131928249Z" level=info msg="StartContainer for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\"" Mar 2 12:55:31.133263 containerd[1890]: time="2026-03-02T12:55:31.133214153Z" level=info msg="connecting to shim 39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b" address="unix:///run/containerd/s/f11e89ea2f0d699230948686878c20aa462ebc7150a873c7d91d29d276f7d297" protocol=ttrpc version=3 Mar 2 12:55:31.152533 systemd[1]: Started cri-containerd-2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5.scope - libcontainer container 2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5. Mar 2 12:55:31.153242 systemd[1]: Started cri-containerd-39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b.scope - libcontainer container 39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b. Mar 2 12:55:31.192865 containerd[1890]: time="2026-03-02T12:55:31.192834310Z" level=info msg="StartContainer for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" returns successfully" Mar 2 12:55:31.203946 systemd[1]: cri-containerd-2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5.scope: Deactivated successfully. Mar 2 12:55:31.208938 containerd[1890]: time="2026-03-02T12:55:31.208020634Z" level=info msg="received container exit event container_id:\"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\" id:\"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\" pid:4002 exited_at:{seconds:1772456131 nanos:207657695}" Mar 2 12:55:31.208938 containerd[1890]: time="2026-03-02T12:55:31.208733840Z" level=info msg="StartContainer for \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\" returns successfully" Mar 2 12:55:31.232818 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5-rootfs.mount: Deactivated successfully. Mar 2 12:55:32.075042 containerd[1890]: time="2026-03-02T12:55:32.074974882Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 2 12:55:32.087905 kubelet[3463]: I0302 12:55:32.087858 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-znz4j" podStartSLOduration=1.952007997 podStartE2EDuration="13.087845359s" podCreationTimestamp="2026-03-02 12:55:19 +0000 UTC" firstStartedPulling="2026-03-02 12:55:19.922317949 +0000 UTC m=+7.000055233" lastFinishedPulling="2026-03-02 12:55:31.058155311 +0000 UTC m=+18.135892595" observedRunningTime="2026-03-02 12:55:32.072346401 +0000 UTC m=+19.150083685" watchObservedRunningTime="2026-03-02 12:55:32.087845359 +0000 UTC m=+19.165582643" Mar 2 12:55:32.094991 containerd[1890]: time="2026-03-02T12:55:32.094905320Z" level=info msg="Container c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:32.116418 containerd[1890]: time="2026-03-02T12:55:32.116370782Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\"" Mar 2 12:55:32.118330 containerd[1890]: time="2026-03-02T12:55:32.116785939Z" level=info msg="StartContainer for \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\"" Mar 2 12:55:32.118070 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount868429398.mount: Deactivated successfully. Mar 2 12:55:32.118666 containerd[1890]: time="2026-03-02T12:55:32.118627532Z" level=info msg="connecting to shim c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc" address="unix:///run/containerd/s/6ff8dcc12b6784907ab387cd008f4a7cbe79ef07dbd7e04e064767f6c012e506" protocol=ttrpc version=3 Mar 2 12:55:32.142920 systemd[1]: Started cri-containerd-c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc.scope - libcontainer container c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc. Mar 2 12:55:32.170859 systemd[1]: cri-containerd-c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc.scope: Deactivated successfully. Mar 2 12:55:32.176577 containerd[1890]: time="2026-03-02T12:55:32.176538068Z" level=info msg="received container exit event container_id:\"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\" id:\"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\" pid:4060 exited_at:{seconds:1772456132 nanos:172116956}" Mar 2 12:55:32.182090 containerd[1890]: time="2026-03-02T12:55:32.182013061Z" level=info msg="StartContainer for \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\" returns successfully" Mar 2 12:55:32.190290 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc-rootfs.mount: Deactivated successfully. Mar 2 12:55:33.080622 containerd[1890]: time="2026-03-02T12:55:33.080584916Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 2 12:55:33.104671 containerd[1890]: time="2026-03-02T12:55:33.104630545Z" level=info msg="Container 1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:33.121631 containerd[1890]: time="2026-03-02T12:55:33.121593636Z" level=info msg="CreateContainer within sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\"" Mar 2 12:55:33.122191 containerd[1890]: time="2026-03-02T12:55:33.122159045Z" level=info msg="StartContainer for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\"" Mar 2 12:55:33.123228 containerd[1890]: time="2026-03-02T12:55:33.123175476Z" level=info msg="connecting to shim 1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47" address="unix:///run/containerd/s/6ff8dcc12b6784907ab387cd008f4a7cbe79ef07dbd7e04e064767f6c012e506" protocol=ttrpc version=3 Mar 2 12:55:33.139626 systemd[1]: Started cri-containerd-1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47.scope - libcontainer container 1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47. Mar 2 12:55:33.172053 containerd[1890]: time="2026-03-02T12:55:33.172015494Z" level=info msg="StartContainer for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" returns successfully" Mar 2 12:55:33.244554 kubelet[3463]: I0302 12:55:33.243725 3463 kubelet_node_status.go:501] "Fast updating node status as it just became ready" Mar 2 12:55:33.296126 systemd[1]: Created slice kubepods-burstable-podd1290227_ba7f_4b79_a979_feb9fb900b9a.slice - libcontainer container kubepods-burstable-podd1290227_ba7f_4b79_a979_feb9fb900b9a.slice. Mar 2 12:55:33.303084 systemd[1]: Created slice kubepods-burstable-pod621f1962_edc2_42e9_9580_81ca50a36bb4.slice - libcontainer container kubepods-burstable-pod621f1962_edc2_42e9_9580_81ca50a36bb4.slice. Mar 2 12:55:33.317670 kubelet[3463]: I0302 12:55:33.317631 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b7x6\" (UniqueName: \"kubernetes.io/projected/621f1962-edc2-42e9-9580-81ca50a36bb4-kube-api-access-4b7x6\") pod \"coredns-674b8bbfcf-s257b\" (UID: \"621f1962-edc2-42e9-9580-81ca50a36bb4\") " pod="kube-system/coredns-674b8bbfcf-s257b" Mar 2 12:55:33.317797 kubelet[3463]: I0302 12:55:33.317776 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/621f1962-edc2-42e9-9580-81ca50a36bb4-config-volume\") pod \"coredns-674b8bbfcf-s257b\" (UID: \"621f1962-edc2-42e9-9580-81ca50a36bb4\") " pod="kube-system/coredns-674b8bbfcf-s257b" Mar 2 12:55:33.317833 kubelet[3463]: I0302 12:55:33.317800 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d1290227-ba7f-4b79-a979-feb9fb900b9a-config-volume\") pod \"coredns-674b8bbfcf-t249m\" (UID: \"d1290227-ba7f-4b79-a979-feb9fb900b9a\") " pod="kube-system/coredns-674b8bbfcf-t249m" Mar 2 12:55:33.317877 kubelet[3463]: I0302 12:55:33.317814 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw578\" (UniqueName: \"kubernetes.io/projected/d1290227-ba7f-4b79-a979-feb9fb900b9a-kube-api-access-fw578\") pod \"coredns-674b8bbfcf-t249m\" (UID: \"d1290227-ba7f-4b79-a979-feb9fb900b9a\") " pod="kube-system/coredns-674b8bbfcf-t249m" Mar 2 12:55:33.600009 containerd[1890]: time="2026-03-02T12:55:33.599975940Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-t249m,Uid:d1290227-ba7f-4b79-a979-feb9fb900b9a,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:33.608335 containerd[1890]: time="2026-03-02T12:55:33.608286980Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-s257b,Uid:621f1962-edc2-42e9-9580-81ca50a36bb4,Namespace:kube-system,Attempt:0,}" Mar 2 12:55:35.119340 systemd-networkd[1485]: cilium_host: Link UP Mar 2 12:55:35.120055 systemd-networkd[1485]: cilium_net: Link UP Mar 2 12:55:35.120738 systemd-networkd[1485]: cilium_net: Gained carrier Mar 2 12:55:35.121105 systemd-networkd[1485]: cilium_host: Gained carrier Mar 2 12:55:35.272663 systemd-networkd[1485]: cilium_vxlan: Link UP Mar 2 12:55:35.273015 systemd-networkd[1485]: cilium_vxlan: Gained carrier Mar 2 12:55:35.477536 kernel: NET: Registered PF_ALG protocol family Mar 2 12:55:35.859518 systemd-networkd[1485]: cilium_net: Gained IPv6LL Mar 2 12:55:35.993242 systemd-networkd[1485]: lxc_health: Link UP Mar 2 12:55:35.994139 systemd-networkd[1485]: lxc_health: Gained carrier Mar 2 12:55:36.050577 systemd-networkd[1485]: cilium_host: Gained IPv6LL Mar 2 12:55:36.132676 systemd-networkd[1485]: lxc741126488c4e: Link UP Mar 2 12:55:36.138640 kernel: eth0: renamed from tmpc171c Mar 2 12:55:36.137801 systemd-networkd[1485]: lxc741126488c4e: Gained carrier Mar 2 12:55:36.158329 systemd-networkd[1485]: lxc2a459a13b9cd: Link UP Mar 2 12:55:36.159525 kernel: eth0: renamed from tmp2bfbc Mar 2 12:55:36.160039 systemd-networkd[1485]: lxc2a459a13b9cd: Gained carrier Mar 2 12:55:36.627498 systemd-networkd[1485]: cilium_vxlan: Gained IPv6LL Mar 2 12:55:37.203542 systemd-networkd[1485]: lxc_health: Gained IPv6LL Mar 2 12:55:37.332525 systemd-networkd[1485]: lxc2a459a13b9cd: Gained IPv6LL Mar 2 12:55:37.668963 kubelet[3463]: I0302 12:55:37.668822 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-6pf4l" podStartSLOduration=10.826668527 podStartE2EDuration="18.668806647s" podCreationTimestamp="2026-03-02 12:55:19 +0000 UTC" firstStartedPulling="2026-03-02 12:55:19.784568864 +0000 UTC m=+6.862306148" lastFinishedPulling="2026-03-02 12:55:27.626706976 +0000 UTC m=+14.704444268" observedRunningTime="2026-03-02 12:55:34.090297724 +0000 UTC m=+21.168035008" watchObservedRunningTime="2026-03-02 12:55:37.668806647 +0000 UTC m=+24.746543931" Mar 2 12:55:37.842623 systemd-networkd[1485]: lxc741126488c4e: Gained IPv6LL Mar 2 12:55:38.677577 containerd[1890]: time="2026-03-02T12:55:38.677536497Z" level=info msg="connecting to shim c171cadb821e88143ae50c2b894183b1dd9b6e68ff89e570a3d89dedb6599204" address="unix:///run/containerd/s/5966f11d23bd516a7b2af799d9bae6a22a9a1117449a9c8f2cd1be33f2db2c8f" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:38.686636 containerd[1890]: time="2026-03-02T12:55:38.686602719Z" level=info msg="connecting to shim 2bfbca7c902215048d3d39f05e8223fd8dc4f3e2b82ccb4cf50e92edee62fa5f" address="unix:///run/containerd/s/8be87d409044d42ff3291ef0166162f2a9106ba72a44fd2c196140d7437c16eb" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:55:38.712543 systemd[1]: Started cri-containerd-c171cadb821e88143ae50c2b894183b1dd9b6e68ff89e570a3d89dedb6599204.scope - libcontainer container c171cadb821e88143ae50c2b894183b1dd9b6e68ff89e570a3d89dedb6599204. Mar 2 12:55:38.716706 systemd[1]: Started cri-containerd-2bfbca7c902215048d3d39f05e8223fd8dc4f3e2b82ccb4cf50e92edee62fa5f.scope - libcontainer container 2bfbca7c902215048d3d39f05e8223fd8dc4f3e2b82ccb4cf50e92edee62fa5f. Mar 2 12:55:38.753257 containerd[1890]: time="2026-03-02T12:55:38.753219600Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-s257b,Uid:621f1962-edc2-42e9-9580-81ca50a36bb4,Namespace:kube-system,Attempt:0,} returns sandbox id \"2bfbca7c902215048d3d39f05e8223fd8dc4f3e2b82ccb4cf50e92edee62fa5f\"" Mar 2 12:55:38.761014 containerd[1890]: time="2026-03-02T12:55:38.760979767Z" level=info msg="CreateContainer within sandbox \"2bfbca7c902215048d3d39f05e8223fd8dc4f3e2b82ccb4cf50e92edee62fa5f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 12:55:38.761923 containerd[1890]: time="2026-03-02T12:55:38.761895538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-674b8bbfcf-t249m,Uid:d1290227-ba7f-4b79-a979-feb9fb900b9a,Namespace:kube-system,Attempt:0,} returns sandbox id \"c171cadb821e88143ae50c2b894183b1dd9b6e68ff89e570a3d89dedb6599204\"" Mar 2 12:55:38.771282 containerd[1890]: time="2026-03-02T12:55:38.771251649Z" level=info msg="CreateContainer within sandbox \"c171cadb821e88143ae50c2b894183b1dd9b6e68ff89e570a3d89dedb6599204\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 2 12:55:38.792000 containerd[1890]: time="2026-03-02T12:55:38.791707530Z" level=info msg="Container 7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:38.797903 containerd[1890]: time="2026-03-02T12:55:38.797873178Z" level=info msg="Container 3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:55:38.833209 containerd[1890]: time="2026-03-02T12:55:38.833160045Z" level=info msg="CreateContainer within sandbox \"2bfbca7c902215048d3d39f05e8223fd8dc4f3e2b82ccb4cf50e92edee62fa5f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592\"" Mar 2 12:55:38.833798 containerd[1890]: time="2026-03-02T12:55:38.833763359Z" level=info msg="StartContainer for \"7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592\"" Mar 2 12:55:38.835633 containerd[1890]: time="2026-03-02T12:55:38.835601806Z" level=info msg="connecting to shim 7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592" address="unix:///run/containerd/s/8be87d409044d42ff3291ef0166162f2a9106ba72a44fd2c196140d7437c16eb" protocol=ttrpc version=3 Mar 2 12:55:38.846447 containerd[1890]: time="2026-03-02T12:55:38.846396864Z" level=info msg="CreateContainer within sandbox \"c171cadb821e88143ae50c2b894183b1dd9b6e68ff89e570a3d89dedb6599204\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b\"" Mar 2 12:55:38.848838 containerd[1890]: time="2026-03-02T12:55:38.848814424Z" level=info msg="StartContainer for \"3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b\"" Mar 2 12:55:38.849423 containerd[1890]: time="2026-03-02T12:55:38.849375265Z" level=info msg="connecting to shim 3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b" address="unix:///run/containerd/s/5966f11d23bd516a7b2af799d9bae6a22a9a1117449a9c8f2cd1be33f2db2c8f" protocol=ttrpc version=3 Mar 2 12:55:38.852564 systemd[1]: Started cri-containerd-7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592.scope - libcontainer container 7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592. Mar 2 12:55:38.870528 systemd[1]: Started cri-containerd-3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b.scope - libcontainer container 3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b. Mar 2 12:55:38.903840 containerd[1890]: time="2026-03-02T12:55:38.903801958Z" level=info msg="StartContainer for \"7593c151a2a7faad484bd4731684292b2fe7be7cf95719ee1c80d5bee06df592\" returns successfully" Mar 2 12:55:38.904321 containerd[1890]: time="2026-03-02T12:55:38.904219947Z" level=info msg="StartContainer for \"3d4523107b7f8975d3ed8ddc6ce3542d6be7c6e843abf94b947725fa015ae93b\" returns successfully" Mar 2 12:55:39.100742 kubelet[3463]: I0302 12:55:39.100254 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-s257b" podStartSLOduration=20.100240331 podStartE2EDuration="20.100240331s" podCreationTimestamp="2026-03-02 12:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:55:39.098570633 +0000 UTC m=+26.176307917" watchObservedRunningTime="2026-03-02 12:55:39.100240331 +0000 UTC m=+26.177977615" Mar 2 12:55:39.113274 kubelet[3463]: I0302 12:55:39.112771 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-674b8bbfcf-t249m" podStartSLOduration=20.11275636 podStartE2EDuration="20.11275636s" podCreationTimestamp="2026-03-02 12:55:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:55:39.11187899 +0000 UTC m=+26.189616402" watchObservedRunningTime="2026-03-02 12:55:39.11275636 +0000 UTC m=+26.190493644" Mar 2 12:55:39.663913 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3626447521.mount: Deactivated successfully. Mar 2 12:56:43.113646 systemd[1]: Started sshd@7-10.200.20.30:22-10.200.16.10:40460.service - OpenSSH per-connection server daemon (10.200.16.10:40460). Mar 2 12:56:43.531745 sshd[4782]: Accepted publickey for core from 10.200.16.10 port 40460 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:56:43.532853 sshd-session[4782]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:56:43.536307 systemd-logind[1868]: New session 10 of user core. Mar 2 12:56:43.541527 systemd[1]: Started session-10.scope - Session 10 of User core. Mar 2 12:56:43.815728 sshd[4785]: Connection closed by 10.200.16.10 port 40460 Mar 2 12:56:43.816381 sshd-session[4782]: pam_unix(sshd:session): session closed for user core Mar 2 12:56:43.819518 systemd[1]: sshd@7-10.200.20.30:22-10.200.16.10:40460.service: Deactivated successfully. Mar 2 12:56:43.820877 systemd[1]: session-10.scope: Deactivated successfully. Mar 2 12:56:43.821537 systemd-logind[1868]: Session 10 logged out. Waiting for processes to exit. Mar 2 12:56:43.822580 systemd-logind[1868]: Removed session 10. Mar 2 12:56:48.910105 systemd[1]: Started sshd@8-10.200.20.30:22-10.200.16.10:40472.service - OpenSSH per-connection server daemon (10.200.16.10:40472). Mar 2 12:56:49.328020 sshd[4797]: Accepted publickey for core from 10.200.16.10 port 40472 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:56:49.329183 sshd-session[4797]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:56:49.332531 systemd-logind[1868]: New session 11 of user core. Mar 2 12:56:49.343699 systemd[1]: Started session-11.scope - Session 11 of User core. Mar 2 12:56:49.601826 sshd[4801]: Connection closed by 10.200.16.10 port 40472 Mar 2 12:56:49.602479 sshd-session[4797]: pam_unix(sshd:session): session closed for user core Mar 2 12:56:49.605834 systemd[1]: sshd@8-10.200.20.30:22-10.200.16.10:40472.service: Deactivated successfully. Mar 2 12:56:49.607573 systemd[1]: session-11.scope: Deactivated successfully. Mar 2 12:56:49.608413 systemd-logind[1868]: Session 11 logged out. Waiting for processes to exit. Mar 2 12:56:49.609962 systemd-logind[1868]: Removed session 11. Mar 2 12:56:54.690690 systemd[1]: Started sshd@9-10.200.20.30:22-10.200.16.10:53694.service - OpenSSH per-connection server daemon (10.200.16.10:53694). Mar 2 12:56:55.108607 sshd[4816]: Accepted publickey for core from 10.200.16.10 port 53694 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:56:55.109777 sshd-session[4816]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:56:55.113244 systemd-logind[1868]: New session 12 of user core. Mar 2 12:56:55.120619 systemd[1]: Started session-12.scope - Session 12 of User core. Mar 2 12:56:55.378506 sshd[4819]: Connection closed by 10.200.16.10 port 53694 Mar 2 12:56:55.378502 sshd-session[4816]: pam_unix(sshd:session): session closed for user core Mar 2 12:56:55.382376 systemd[1]: sshd@9-10.200.20.30:22-10.200.16.10:53694.service: Deactivated successfully. Mar 2 12:56:55.384252 systemd[1]: session-12.scope: Deactivated successfully. Mar 2 12:56:55.385165 systemd-logind[1868]: Session 12 logged out. Waiting for processes to exit. Mar 2 12:56:55.386305 systemd-logind[1868]: Removed session 12. Mar 2 12:57:00.470731 systemd[1]: Started sshd@10-10.200.20.30:22-10.200.16.10:40762.service - OpenSSH per-connection server daemon (10.200.16.10:40762). Mar 2 12:57:00.893471 sshd[4831]: Accepted publickey for core from 10.200.16.10 port 40762 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:00.894333 sshd-session[4831]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:00.897855 systemd-logind[1868]: New session 13 of user core. Mar 2 12:57:00.902515 systemd[1]: Started session-13.scope - Session 13 of User core. Mar 2 12:57:01.163766 sshd[4834]: Connection closed by 10.200.16.10 port 40762 Mar 2 12:57:01.163600 sshd-session[4831]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:01.168034 systemd-logind[1868]: Session 13 logged out. Waiting for processes to exit. Mar 2 12:57:01.168239 systemd[1]: sshd@10-10.200.20.30:22-10.200.16.10:40762.service: Deactivated successfully. Mar 2 12:57:01.171048 systemd[1]: session-13.scope: Deactivated successfully. Mar 2 12:57:01.173266 systemd-logind[1868]: Removed session 13. Mar 2 12:57:01.252211 systemd[1]: Started sshd@11-10.200.20.30:22-10.200.16.10:40776.service - OpenSSH per-connection server daemon (10.200.16.10:40776). Mar 2 12:57:01.676807 sshd[4847]: Accepted publickey for core from 10.200.16.10 port 40776 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:01.677219 sshd-session[4847]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:01.680657 systemd-logind[1868]: New session 14 of user core. Mar 2 12:57:01.686525 systemd[1]: Started session-14.scope - Session 14 of User core. Mar 2 12:57:01.981337 sshd[4850]: Connection closed by 10.200.16.10 port 40776 Mar 2 12:57:01.981247 sshd-session[4847]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:01.984990 systemd[1]: sshd@11-10.200.20.30:22-10.200.16.10:40776.service: Deactivated successfully. Mar 2 12:57:01.986598 systemd[1]: session-14.scope: Deactivated successfully. Mar 2 12:57:01.987253 systemd-logind[1868]: Session 14 logged out. Waiting for processes to exit. Mar 2 12:57:01.988601 systemd-logind[1868]: Removed session 14. Mar 2 12:57:02.070272 systemd[1]: Started sshd@12-10.200.20.30:22-10.200.16.10:40778.service - OpenSSH per-connection server daemon (10.200.16.10:40778). Mar 2 12:57:02.492508 sshd[4860]: Accepted publickey for core from 10.200.16.10 port 40778 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:02.493559 sshd-session[4860]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:02.497268 systemd-logind[1868]: New session 15 of user core. Mar 2 12:57:02.505512 systemd[1]: Started session-15.scope - Session 15 of User core. Mar 2 12:57:02.766491 sshd[4863]: Connection closed by 10.200.16.10 port 40778 Mar 2 12:57:02.767109 sshd-session[4860]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:02.770311 systemd-logind[1868]: Session 15 logged out. Waiting for processes to exit. Mar 2 12:57:02.770987 systemd[1]: sshd@12-10.200.20.30:22-10.200.16.10:40778.service: Deactivated successfully. Mar 2 12:57:02.773094 systemd[1]: session-15.scope: Deactivated successfully. Mar 2 12:57:02.775202 systemd-logind[1868]: Removed session 15. Mar 2 12:57:07.862954 systemd[1]: Started sshd@13-10.200.20.30:22-10.200.16.10:40794.service - OpenSSH per-connection server daemon (10.200.16.10:40794). Mar 2 12:57:08.277386 sshd[4874]: Accepted publickey for core from 10.200.16.10 port 40794 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:08.278382 sshd-session[4874]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:08.281813 systemd-logind[1868]: New session 16 of user core. Mar 2 12:57:08.293524 systemd[1]: Started session-16.scope - Session 16 of User core. Mar 2 12:57:08.549283 sshd[4877]: Connection closed by 10.200.16.10 port 40794 Mar 2 12:57:08.549116 sshd-session[4874]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:08.552987 systemd[1]: sshd@13-10.200.20.30:22-10.200.16.10:40794.service: Deactivated successfully. Mar 2 12:57:08.555882 systemd[1]: session-16.scope: Deactivated successfully. Mar 2 12:57:08.556872 systemd-logind[1868]: Session 16 logged out. Waiting for processes to exit. Mar 2 12:57:08.558295 systemd-logind[1868]: Removed session 16. Mar 2 12:57:08.638858 systemd[1]: Started sshd@14-10.200.20.30:22-10.200.16.10:40804.service - OpenSSH per-connection server daemon (10.200.16.10:40804). Mar 2 12:57:09.057433 sshd[4888]: Accepted publickey for core from 10.200.16.10 port 40804 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:09.058339 sshd-session[4888]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:09.061627 systemd-logind[1868]: New session 17 of user core. Mar 2 12:57:09.066614 systemd[1]: Started session-17.scope - Session 17 of User core. Mar 2 12:57:09.379528 sshd[4891]: Connection closed by 10.200.16.10 port 40804 Mar 2 12:57:09.380467 sshd-session[4888]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:09.383303 systemd[1]: sshd@14-10.200.20.30:22-10.200.16.10:40804.service: Deactivated successfully. Mar 2 12:57:09.385060 systemd[1]: session-17.scope: Deactivated successfully. Mar 2 12:57:09.386026 systemd-logind[1868]: Session 17 logged out. Waiting for processes to exit. Mar 2 12:57:09.387234 systemd-logind[1868]: Removed session 17. Mar 2 12:57:09.467308 systemd[1]: Started sshd@15-10.200.20.30:22-10.200.16.10:40816.service - OpenSSH per-connection server daemon (10.200.16.10:40816). Mar 2 12:57:09.883538 sshd[4901]: Accepted publickey for core from 10.200.16.10 port 40816 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:09.884591 sshd-session[4901]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:09.888145 systemd-logind[1868]: New session 18 of user core. Mar 2 12:57:09.897712 systemd[1]: Started session-18.scope - Session 18 of User core. Mar 2 12:57:13.581073 sshd[4904]: Connection closed by 10.200.16.10 port 40816 Mar 2 12:57:13.582586 sshd-session[4901]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:13.585630 systemd[1]: sshd@15-10.200.20.30:22-10.200.16.10:40816.service: Deactivated successfully. Mar 2 12:57:13.587729 systemd[1]: session-18.scope: Deactivated successfully. Mar 2 12:57:13.589128 systemd-logind[1868]: Session 18 logged out. Waiting for processes to exit. Mar 2 12:57:13.590500 systemd-logind[1868]: Removed session 18. Mar 2 12:57:13.670060 systemd[1]: Started sshd@16-10.200.20.30:22-10.200.16.10:43368.service - OpenSSH per-connection server daemon (10.200.16.10:43368). Mar 2 12:57:14.093717 sshd[4923]: Accepted publickey for core from 10.200.16.10 port 43368 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:14.094896 sshd-session[4923]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:14.099656 systemd-logind[1868]: New session 19 of user core. Mar 2 12:57:14.105546 systemd[1]: Started session-19.scope - Session 19 of User core. Mar 2 12:57:14.440688 sshd[4926]: Connection closed by 10.200.16.10 port 43368 Mar 2 12:57:14.441224 sshd-session[4923]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:14.444327 systemd[1]: sshd@16-10.200.20.30:22-10.200.16.10:43368.service: Deactivated successfully. Mar 2 12:57:14.446137 systemd[1]: session-19.scope: Deactivated successfully. Mar 2 12:57:14.447048 systemd-logind[1868]: Session 19 logged out. Waiting for processes to exit. Mar 2 12:57:14.448799 systemd-logind[1868]: Removed session 19. Mar 2 12:57:14.556759 systemd[1]: Started sshd@17-10.200.20.30:22-10.200.16.10:43376.service - OpenSSH per-connection server daemon (10.200.16.10:43376). Mar 2 12:57:14.975610 sshd[4936]: Accepted publickey for core from 10.200.16.10 port 43376 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:14.976699 sshd-session[4936]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:14.980511 systemd-logind[1868]: New session 20 of user core. Mar 2 12:57:14.985544 systemd[1]: Started session-20.scope - Session 20 of User core. Mar 2 12:57:15.247151 sshd[4939]: Connection closed by 10.200.16.10 port 43376 Mar 2 12:57:15.249492 sshd-session[4936]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:15.252432 systemd-logind[1868]: Session 20 logged out. Waiting for processes to exit. Mar 2 12:57:15.252675 systemd[1]: sshd@17-10.200.20.30:22-10.200.16.10:43376.service: Deactivated successfully. Mar 2 12:57:15.256048 systemd[1]: session-20.scope: Deactivated successfully. Mar 2 12:57:15.258067 systemd-logind[1868]: Removed session 20. Mar 2 12:57:20.346778 systemd[1]: Started sshd@18-10.200.20.30:22-10.200.16.10:52738.service - OpenSSH per-connection server daemon (10.200.16.10:52738). Mar 2 12:57:20.768834 sshd[4955]: Accepted publickey for core from 10.200.16.10 port 52738 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:20.769945 sshd-session[4955]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:20.773370 systemd-logind[1868]: New session 21 of user core. Mar 2 12:57:20.780533 systemd[1]: Started session-21.scope - Session 21 of User core. Mar 2 12:57:21.039021 sshd[4958]: Connection closed by 10.200.16.10 port 52738 Mar 2 12:57:21.039509 sshd-session[4955]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:21.042934 systemd[1]: sshd@18-10.200.20.30:22-10.200.16.10:52738.service: Deactivated successfully. Mar 2 12:57:21.045362 systemd[1]: session-21.scope: Deactivated successfully. Mar 2 12:57:21.046272 systemd-logind[1868]: Session 21 logged out. Waiting for processes to exit. Mar 2 12:57:21.047822 systemd-logind[1868]: Removed session 21. Mar 2 12:57:26.131363 systemd[1]: Started sshd@19-10.200.20.30:22-10.200.16.10:52740.service - OpenSSH per-connection server daemon (10.200.16.10:52740). Mar 2 12:57:26.553438 sshd[4970]: Accepted publickey for core from 10.200.16.10 port 52740 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:26.554437 sshd-session[4970]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:26.557695 systemd-logind[1868]: New session 22 of user core. Mar 2 12:57:26.568531 systemd[1]: Started session-22.scope - Session 22 of User core. Mar 2 12:57:26.826231 sshd[4973]: Connection closed by 10.200.16.10 port 52740 Mar 2 12:57:26.826717 sshd-session[4970]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:26.830317 systemd[1]: sshd@19-10.200.20.30:22-10.200.16.10:52740.service: Deactivated successfully. Mar 2 12:57:26.832133 systemd[1]: session-22.scope: Deactivated successfully. Mar 2 12:57:26.833155 systemd-logind[1868]: Session 22 logged out. Waiting for processes to exit. Mar 2 12:57:26.834606 systemd-logind[1868]: Removed session 22. Mar 2 12:57:26.915521 systemd[1]: Started sshd@20-10.200.20.30:22-10.200.16.10:52746.service - OpenSSH per-connection server daemon (10.200.16.10:52746). Mar 2 12:57:27.338551 sshd[4985]: Accepted publickey for core from 10.200.16.10 port 52746 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:27.339644 sshd-session[4985]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:27.343042 systemd-logind[1868]: New session 23 of user core. Mar 2 12:57:27.351520 systemd[1]: Started session-23.scope - Session 23 of User core. Mar 2 12:57:28.777645 containerd[1890]: time="2026-03-02T12:57:28.777605909Z" level=info msg="StopContainer for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" with timeout 30 (s)" Mar 2 12:57:28.778415 containerd[1890]: time="2026-03-02T12:57:28.778355891Z" level=info msg="Stop container \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" with signal terminated" Mar 2 12:57:28.780811 containerd[1890]: time="2026-03-02T12:57:28.780785548Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 2 12:57:28.787367 containerd[1890]: time="2026-03-02T12:57:28.787348530Z" level=info msg="StopContainer for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" with timeout 2 (s)" Mar 2 12:57:28.788090 containerd[1890]: time="2026-03-02T12:57:28.788075336Z" level=info msg="Stop container \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" with signal terminated" Mar 2 12:57:28.788385 systemd[1]: cri-containerd-39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b.scope: Deactivated successfully. Mar 2 12:57:28.792258 containerd[1890]: time="2026-03-02T12:57:28.792192396Z" level=info msg="received container exit event container_id:\"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" id:\"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" pid:4001 exited_at:{seconds:1772456248 nanos:791908276}" Mar 2 12:57:28.798708 systemd-networkd[1485]: lxc_health: Link DOWN Mar 2 12:57:28.799012 systemd-networkd[1485]: lxc_health: Lost carrier Mar 2 12:57:28.813666 containerd[1890]: time="2026-03-02T12:57:28.813526768Z" level=info msg="received container exit event container_id:\"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" id:\"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" pid:4097 exited_at:{seconds:1772456248 nanos:813317634}" Mar 2 12:57:28.815057 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b-rootfs.mount: Deactivated successfully. Mar 2 12:57:28.816751 systemd[1]: cri-containerd-1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47.scope: Deactivated successfully. Mar 2 12:57:28.818732 systemd[1]: cri-containerd-1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47.scope: Consumed 4.262s CPU time, 125.3M memory peak, 112K read from disk, 12.9M written to disk. Mar 2 12:57:28.831941 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47-rootfs.mount: Deactivated successfully. Mar 2 12:57:28.896376 containerd[1890]: time="2026-03-02T12:57:28.896318145Z" level=info msg="StopContainer for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" returns successfully" Mar 2 12:57:28.897251 containerd[1890]: time="2026-03-02T12:57:28.897225684Z" level=info msg="StopPodSandbox for \"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\"" Mar 2 12:57:28.897397 containerd[1890]: time="2026-03-02T12:57:28.897380881Z" level=info msg="Container to stop \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:57:28.899766 containerd[1890]: time="2026-03-02T12:57:28.899730624Z" level=info msg="StopContainer for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" returns successfully" Mar 2 12:57:28.900422 containerd[1890]: time="2026-03-02T12:57:28.900232927Z" level=info msg="StopPodSandbox for \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\"" Mar 2 12:57:28.900422 containerd[1890]: time="2026-03-02T12:57:28.900271016Z" level=info msg="Container to stop \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:57:28.900422 containerd[1890]: time="2026-03-02T12:57:28.900278928Z" level=info msg="Container to stop \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:57:28.900422 containerd[1890]: time="2026-03-02T12:57:28.900284544Z" level=info msg="Container to stop \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:57:28.900422 containerd[1890]: time="2026-03-02T12:57:28.900289648Z" level=info msg="Container to stop \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:57:28.900422 containerd[1890]: time="2026-03-02T12:57:28.900294961Z" level=info msg="Container to stop \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 2 12:57:28.904296 systemd[1]: cri-containerd-c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a.scope: Deactivated successfully. Mar 2 12:57:28.906687 systemd[1]: cri-containerd-3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf.scope: Deactivated successfully. Mar 2 12:57:28.907991 containerd[1890]: time="2026-03-02T12:57:28.907966536Z" level=info msg="received sandbox exit event container_id:\"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" id:\"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" exit_status:137 exited_at:{seconds:1772456248 nanos:907572380}" monitor_name=podsandbox Mar 2 12:57:28.910812 containerd[1890]: time="2026-03-02T12:57:28.910791109Z" level=info msg="received sandbox exit event container_id:\"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" id:\"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" exit_status:137 exited_at:{seconds:1772456248 nanos:910651089}" monitor_name=podsandbox Mar 2 12:57:28.934437 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a-rootfs.mount: Deactivated successfully. Mar 2 12:57:28.939080 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf-rootfs.mount: Deactivated successfully. Mar 2 12:57:28.970676 containerd[1890]: time="2026-03-02T12:57:28.970088369Z" level=info msg="shim disconnected" id=c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a namespace=k8s.io Mar 2 12:57:28.970676 containerd[1890]: time="2026-03-02T12:57:28.970561560Z" level=warning msg="cleaning up after shim disconnected" id=c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a namespace=k8s.io Mar 2 12:57:28.970676 containerd[1890]: time="2026-03-02T12:57:28.970596217Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 12:57:28.970676 containerd[1890]: time="2026-03-02T12:57:28.970394515Z" level=info msg="shim disconnected" id=3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf namespace=k8s.io Mar 2 12:57:28.970676 containerd[1890]: time="2026-03-02T12:57:28.970638066Z" level=warning msg="cleaning up after shim disconnected" id=3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf namespace=k8s.io Mar 2 12:57:28.970676 containerd[1890]: time="2026-03-02T12:57:28.970656787Z" level=info msg="cleaning up dead shim" namespace=k8s.io Mar 2 12:57:28.980396 containerd[1890]: time="2026-03-02T12:57:28.980332374Z" level=info msg="received sandbox container exit event sandbox_id:\"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" exit_status:137 exited_at:{seconds:1772456248 nanos:910651089}" monitor_name=criService Mar 2 12:57:28.982351 containerd[1890]: time="2026-03-02T12:57:28.982232608Z" level=info msg="TearDown network for sandbox \"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" successfully" Mar 2 12:57:28.982351 containerd[1890]: time="2026-03-02T12:57:28.982254128Z" level=info msg="StopPodSandbox for \"c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a\" returns successfully" Mar 2 12:57:28.982460 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c4a8e77cabc906a868ca0300767d45b43c2698914d2421271aa3ff736857906a-shm.mount: Deactivated successfully. Mar 2 12:57:28.985337 containerd[1890]: time="2026-03-02T12:57:28.985215570Z" level=info msg="received sandbox container exit event sandbox_id:\"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" exit_status:137 exited_at:{seconds:1772456248 nanos:907572380}" monitor_name=criService Mar 2 12:57:28.985872 containerd[1890]: time="2026-03-02T12:57:28.985823636Z" level=info msg="TearDown network for sandbox \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" successfully" Mar 2 12:57:28.985872 containerd[1890]: time="2026-03-02T12:57:28.985844668Z" level=info msg="StopPodSandbox for \"3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf\" returns successfully" Mar 2 12:57:29.107933 kubelet[3463]: I0302 12:57:29.106126 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.107933 kubelet[3463]: I0302 12:57:29.105842 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-xtables-lock\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.107933 kubelet[3463]: I0302 12:57:29.106280 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-net\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.107933 kubelet[3463]: I0302 12:57:29.106309 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-config-path\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.107933 kubelet[3463]: I0302 12:57:29.106325 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-kernel\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.107933 kubelet[3463]: I0302 12:57:29.106336 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-lib-modules\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108339 kubelet[3463]: I0302 12:57:29.106349 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-run\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108339 kubelet[3463]: I0302 12:57:29.106363 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8defe60e-7a55-4ade-b1cc-01988615f84e-clustermesh-secrets\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108339 kubelet[3463]: I0302 12:57:29.106451 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-bpf-maps\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108339 kubelet[3463]: I0302 12:57:29.106470 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nkb4\" (UniqueName: \"kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-kube-api-access-2nkb4\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108339 kubelet[3463]: I0302 12:57:29.106584 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-hubble-tls\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108339 kubelet[3463]: I0302 12:57:29.106600 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-hostproc\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108448 kubelet[3463]: I0302 12:57:29.106696 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-etc-cni-netd\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108448 kubelet[3463]: I0302 12:57:29.106711 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-cgroup\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108448 kubelet[3463]: I0302 12:57:29.106725 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-cilium-config-path\") pod \"4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8\" (UID: \"4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8\") " Mar 2 12:57:29.108448 kubelet[3463]: I0302 12:57:29.106778 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cni-path\") pod \"8defe60e-7a55-4ade-b1cc-01988615f84e\" (UID: \"8defe60e-7a55-4ade-b1cc-01988615f84e\") " Mar 2 12:57:29.108448 kubelet[3463]: I0302 12:57:29.106789 3463 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzc6z\" (UniqueName: \"kubernetes.io/projected/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-kube-api-access-kzc6z\") pod \"4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8\" (UID: \"4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8\") " Mar 2 12:57:29.108448 kubelet[3463]: I0302 12:57:29.106855 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.108537 kubelet[3463]: I0302 12:57:29.106876 3463 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-xtables-lock\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.108537 kubelet[3463]: I0302 12:57:29.107196 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.108537 kubelet[3463]: I0302 12:57:29.108463 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 12:57:29.109138 kubelet[3463]: I0302 12:57:29.109107 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.109195 kubelet[3463]: I0302 12:57:29.109142 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.109195 kubelet[3463]: I0302 12:57:29.109155 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.110535 kubelet[3463]: I0302 12:57:29.110512 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.110643 kubelet[3463]: I0302 12:57:29.110631 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-hostproc" (OuterVolumeSpecName: "hostproc") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.110721 kubelet[3463]: I0302 12:57:29.110710 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.111198 kubelet[3463]: I0302 12:57:29.111167 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cni-path" (OuterVolumeSpecName: "cni-path") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Mar 2 12:57:29.115021 kubelet[3463]: I0302 12:57:29.114989 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-kube-api-access-2nkb4" (OuterVolumeSpecName: "kube-api-access-2nkb4") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "kube-api-access-2nkb4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 12:57:29.116959 kubelet[3463]: I0302 12:57:29.115566 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8" (UID: "4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Mar 2 12:57:29.116959 kubelet[3463]: I0302 12:57:29.115632 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8defe60e-7a55-4ade-b1cc-01988615f84e-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Mar 2 12:57:29.116959 kubelet[3463]: I0302 12:57:29.116876 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-kube-api-access-kzc6z" (OuterVolumeSpecName: "kube-api-access-kzc6z") pod "4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8" (UID: "4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8"). InnerVolumeSpecName "kube-api-access-kzc6z". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 12:57:29.116959 kubelet[3463]: I0302 12:57:29.116914 3463 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "8defe60e-7a55-4ade-b1cc-01988615f84e" (UID: "8defe60e-7a55-4ade-b1cc-01988615f84e"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Mar 2 12:57:29.207559 kubelet[3463]: I0302 12:57:29.207520 3463 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-config-path\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207559 kubelet[3463]: I0302 12:57:29.207551 3463 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-kernel\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207559 kubelet[3463]: I0302 12:57:29.207560 3463 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-lib-modules\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207559 kubelet[3463]: I0302 12:57:29.207569 3463 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-run\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207559 kubelet[3463]: I0302 12:57:29.207577 3463 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/8defe60e-7a55-4ade-b1cc-01988615f84e-clustermesh-secrets\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207583 3463 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-bpf-maps\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207589 3463 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-2nkb4\" (UniqueName: \"kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-kube-api-access-2nkb4\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207596 3463 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/8defe60e-7a55-4ade-b1cc-01988615f84e-hubble-tls\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207601 3463 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-hostproc\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207605 3463 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-etc-cni-netd\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207610 3463 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cilium-cgroup\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207615 3463 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-cilium-config-path\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207785 kubelet[3463]: I0302 12:57:29.207620 3463 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-cni-path\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207899 kubelet[3463]: I0302 12:57:29.207625 3463 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kzc6z\" (UniqueName: \"kubernetes.io/projected/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8-kube-api-access-kzc6z\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.207899 kubelet[3463]: I0302 12:57:29.207630 3463 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/8defe60e-7a55-4ade-b1cc-01988615f84e-host-proc-sys-net\") on node \"ci-4459.2.101-f64d7d9b2a\" DevicePath \"\"" Mar 2 12:57:29.278155 kubelet[3463]: I0302 12:57:29.277978 3463 scope.go:117] "RemoveContainer" containerID="39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b" Mar 2 12:57:29.282773 containerd[1890]: time="2026-03-02T12:57:29.282049081Z" level=info msg="RemoveContainer for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\"" Mar 2 12:57:29.283087 systemd[1]: Removed slice kubepods-besteffort-pod4df0e96b_7b35_42fc_b5d0_ec0b8cd65ba8.slice - libcontainer container kubepods-besteffort-pod4df0e96b_7b35_42fc_b5d0_ec0b8cd65ba8.slice. Mar 2 12:57:29.292301 systemd[1]: Removed slice kubepods-burstable-pod8defe60e_7a55_4ade_b1cc_01988615f84e.slice - libcontainer container kubepods-burstable-pod8defe60e_7a55_4ade_b1cc_01988615f84e.slice. Mar 2 12:57:29.292384 systemd[1]: kubepods-burstable-pod8defe60e_7a55_4ade_b1cc_01988615f84e.slice: Consumed 4.324s CPU time, 125.7M memory peak, 112K read from disk, 12.9M written to disk. Mar 2 12:57:29.299199 containerd[1890]: time="2026-03-02T12:57:29.299088395Z" level=info msg="RemoveContainer for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" returns successfully" Mar 2 12:57:29.299445 kubelet[3463]: I0302 12:57:29.299385 3463 scope.go:117] "RemoveContainer" containerID="39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b" Mar 2 12:57:29.301062 containerd[1890]: time="2026-03-02T12:57:29.301004933Z" level=error msg="ContainerStatus for \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\": not found" Mar 2 12:57:29.301298 kubelet[3463]: E0302 12:57:29.301224 3463 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\": not found" containerID="39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b" Mar 2 12:57:29.301298 kubelet[3463]: I0302 12:57:29.301249 3463 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b"} err="failed to get container status \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\": rpc error: code = NotFound desc = an error occurred when try to find container \"39d875bd9cb68b9a1f496ed2f99fb2dee0a044233ad4543fea4ecc442c50435b\": not found" Mar 2 12:57:29.301474 kubelet[3463]: I0302 12:57:29.301287 3463 scope.go:117] "RemoveContainer" containerID="1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47" Mar 2 12:57:29.303446 containerd[1890]: time="2026-03-02T12:57:29.303367196Z" level=info msg="RemoveContainer for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\"" Mar 2 12:57:29.315574 containerd[1890]: time="2026-03-02T12:57:29.315548876Z" level=info msg="RemoveContainer for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" returns successfully" Mar 2 12:57:29.316425 kubelet[3463]: I0302 12:57:29.316237 3463 scope.go:117] "RemoveContainer" containerID="c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc" Mar 2 12:57:29.318151 containerd[1890]: time="2026-03-02T12:57:29.318122257Z" level=info msg="RemoveContainer for \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\"" Mar 2 12:57:29.327641 containerd[1890]: time="2026-03-02T12:57:29.327608327Z" level=info msg="RemoveContainer for \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\" returns successfully" Mar 2 12:57:29.327942 kubelet[3463]: I0302 12:57:29.327925 3463 scope.go:117] "RemoveContainer" containerID="2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5" Mar 2 12:57:29.330009 containerd[1890]: time="2026-03-02T12:57:29.329942438Z" level=info msg="RemoveContainer for \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\"" Mar 2 12:57:29.343683 containerd[1890]: time="2026-03-02T12:57:29.343647163Z" level=info msg="RemoveContainer for \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\" returns successfully" Mar 2 12:57:29.343893 kubelet[3463]: I0302 12:57:29.343871 3463 scope.go:117] "RemoveContainer" containerID="59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef" Mar 2 12:57:29.345213 containerd[1890]: time="2026-03-02T12:57:29.345185289Z" level=info msg="RemoveContainer for \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\"" Mar 2 12:57:29.356772 containerd[1890]: time="2026-03-02T12:57:29.356738470Z" level=info msg="RemoveContainer for \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\" returns successfully" Mar 2 12:57:29.356988 kubelet[3463]: I0302 12:57:29.356963 3463 scope.go:117] "RemoveContainer" containerID="854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180" Mar 2 12:57:29.359020 containerd[1890]: time="2026-03-02T12:57:29.358572853Z" level=info msg="RemoveContainer for \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\"" Mar 2 12:57:29.366982 containerd[1890]: time="2026-03-02T12:57:29.366951130Z" level=info msg="RemoveContainer for \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\" returns successfully" Mar 2 12:57:29.367312 kubelet[3463]: I0302 12:57:29.367291 3463 scope.go:117] "RemoveContainer" containerID="1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47" Mar 2 12:57:29.367524 containerd[1890]: time="2026-03-02T12:57:29.367492394Z" level=error msg="ContainerStatus for \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\": not found" Mar 2 12:57:29.367658 kubelet[3463]: E0302 12:57:29.367636 3463 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\": not found" containerID="1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47" Mar 2 12:57:29.367740 kubelet[3463]: I0302 12:57:29.367720 3463 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47"} err="failed to get container status \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\": rpc error: code = NotFound desc = an error occurred when try to find container \"1600a9cb8667b13c0abaf0560825329c502ec930e06971d5b29fc02889041f47\": not found" Mar 2 12:57:29.367851 kubelet[3463]: I0302 12:57:29.367782 3463 scope.go:117] "RemoveContainer" containerID="c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc" Mar 2 12:57:29.368030 containerd[1890]: time="2026-03-02T12:57:29.367932679Z" level=error msg="ContainerStatus for \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\": not found" Mar 2 12:57:29.368117 kubelet[3463]: E0302 12:57:29.368097 3463 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\": not found" containerID="c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc" Mar 2 12:57:29.368155 kubelet[3463]: I0302 12:57:29.368118 3463 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc"} err="failed to get container status \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\": rpc error: code = NotFound desc = an error occurred when try to find container \"c67cf5d59c39347934fad5cd3eef20762847617e910ff5ebe5ae912ba42ce2cc\": not found" Mar 2 12:57:29.368177 kubelet[3463]: I0302 12:57:29.368157 3463 scope.go:117] "RemoveContainer" containerID="2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5" Mar 2 12:57:29.368318 containerd[1890]: time="2026-03-02T12:57:29.368290026Z" level=error msg="ContainerStatus for \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\": not found" Mar 2 12:57:29.368506 kubelet[3463]: E0302 12:57:29.368416 3463 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\": not found" containerID="2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5" Mar 2 12:57:29.368506 kubelet[3463]: I0302 12:57:29.368437 3463 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5"} err="failed to get container status \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\": rpc error: code = NotFound desc = an error occurred when try to find container \"2a3a34a039da553334b65846602d880ba954ba53a1e66b374da5a8bbe9a1c1f5\": not found" Mar 2 12:57:29.368506 kubelet[3463]: I0302 12:57:29.368448 3463 scope.go:117] "RemoveContainer" containerID="59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef" Mar 2 12:57:29.368789 containerd[1890]: time="2026-03-02T12:57:29.368760424Z" level=error msg="ContainerStatus for \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\": not found" Mar 2 12:57:29.369025 kubelet[3463]: E0302 12:57:29.369007 3463 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\": not found" containerID="59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef" Mar 2 12:57:29.369112 kubelet[3463]: I0302 12:57:29.369095 3463 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef"} err="failed to get container status \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\": rpc error: code = NotFound desc = an error occurred when try to find container \"59be241e920455c0e88b80c8ffd782514589b70ad90ca4e7f89055155ef145ef\": not found" Mar 2 12:57:29.369239 kubelet[3463]: I0302 12:57:29.369162 3463 scope.go:117] "RemoveContainer" containerID="854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180" Mar 2 12:57:29.369397 containerd[1890]: time="2026-03-02T12:57:29.369371995Z" level=error msg="ContainerStatus for \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\": not found" Mar 2 12:57:29.369622 kubelet[3463]: E0302 12:57:29.369574 3463 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\": not found" containerID="854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180" Mar 2 12:57:29.369622 kubelet[3463]: I0302 12:57:29.369617 3463 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180"} err="failed to get container status \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\": rpc error: code = NotFound desc = an error occurred when try to find container \"854c60fe5ee96fee5886906d8d597caf4cf19abfc9e99898bcefa1b41f23e180\": not found" Mar 2 12:57:29.813907 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3af1fa26cd213fbe23958d71de3435eed340debe6cbab7acbefed347980b0edf-shm.mount: Deactivated successfully. Mar 2 12:57:29.814001 systemd[1]: var-lib-kubelet-pods-4df0e96b\x2d7b35\x2d42fc\x2db5d0\x2dec0b8cd65ba8-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkzc6z.mount: Deactivated successfully. Mar 2 12:57:29.814050 systemd[1]: var-lib-kubelet-pods-8defe60e\x2d7a55\x2d4ade\x2db1cc\x2d01988615f84e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d2nkb4.mount: Deactivated successfully. Mar 2 12:57:29.814086 systemd[1]: var-lib-kubelet-pods-8defe60e\x2d7a55\x2d4ade\x2db1cc\x2d01988615f84e-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 2 12:57:29.814122 systemd[1]: var-lib-kubelet-pods-8defe60e\x2d7a55\x2d4ade\x2db1cc\x2d01988615f84e-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 2 12:57:30.801510 sshd[4988]: Connection closed by 10.200.16.10 port 52746 Mar 2 12:57:30.802074 sshd-session[4985]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:30.805517 systemd-logind[1868]: Session 23 logged out. Waiting for processes to exit. Mar 2 12:57:30.806459 systemd[1]: sshd@20-10.200.20.30:22-10.200.16.10:52746.service: Deactivated successfully. Mar 2 12:57:30.808271 systemd[1]: session-23.scope: Deactivated successfully. Mar 2 12:57:30.809452 systemd-logind[1868]: Removed session 23. Mar 2 12:57:30.894838 systemd[1]: Started sshd@21-10.200.20.30:22-10.200.16.10:45878.service - OpenSSH per-connection server daemon (10.200.16.10:45878). Mar 2 12:57:30.997017 kubelet[3463]: I0302 12:57:30.996979 3463 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8" path="/var/lib/kubelet/pods/4df0e96b-7b35-42fc-b5d0-ec0b8cd65ba8/volumes" Mar 2 12:57:30.997341 kubelet[3463]: I0302 12:57:30.997264 3463 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8defe60e-7a55-4ade-b1cc-01988615f84e" path="/var/lib/kubelet/pods/8defe60e-7a55-4ade-b1cc-01988615f84e/volumes" Mar 2 12:57:31.309956 sshd[5132]: Accepted publickey for core from 10.200.16.10 port 45878 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:31.311011 sshd-session[5132]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:31.314571 systemd-logind[1868]: New session 24 of user core. Mar 2 12:57:31.323528 systemd[1]: Started session-24.scope - Session 24 of User core. Mar 2 12:57:31.975708 systemd[1]: Created slice kubepods-burstable-pod6c3c269d_4fdf_44de_8d92_962fe4f88e97.slice - libcontainer container kubepods-burstable-pod6c3c269d_4fdf_44de_8d92_962fe4f88e97.slice. Mar 2 12:57:32.015807 sshd[5135]: Connection closed by 10.200.16.10 port 45878 Mar 2 12:57:32.016687 sshd-session[5132]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:32.021072 kubelet[3463]: I0302 12:57:32.021044 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-cilium-cgroup\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022561 kubelet[3463]: I0302 12:57:32.021076 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6c3c269d-4fdf-44de-8d92-962fe4f88e97-hubble-tls\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022561 kubelet[3463]: I0302 12:57:32.021095 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6c3c269d-4fdf-44de-8d92-962fe4f88e97-clustermesh-secrets\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022561 kubelet[3463]: I0302 12:57:32.021109 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-cilium-run\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022561 kubelet[3463]: I0302 12:57:32.021118 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-host-proc-sys-net\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022561 kubelet[3463]: I0302 12:57:32.021131 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-hostproc\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022561 kubelet[3463]: I0302 12:57:32.021141 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-lib-modules\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022275 systemd[1]: sshd@21-10.200.20.30:22-10.200.16.10:45878.service: Deactivated successfully. Mar 2 12:57:32.022765 kubelet[3463]: I0302 12:57:32.021152 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6c3c269d-4fdf-44de-8d92-962fe4f88e97-cilium-ipsec-secrets\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022765 kubelet[3463]: I0302 12:57:32.021166 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-host-proc-sys-kernel\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022765 kubelet[3463]: I0302 12:57:32.021181 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-cni-path\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022765 kubelet[3463]: I0302 12:57:32.021190 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-etc-cni-netd\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022765 kubelet[3463]: I0302 12:57:32.021200 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-xtables-lock\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022765 kubelet[3463]: I0302 12:57:32.021216 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6c3c269d-4fdf-44de-8d92-962fe4f88e97-bpf-maps\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022858 kubelet[3463]: I0302 12:57:32.021224 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv6b6\" (UniqueName: \"kubernetes.io/projected/6c3c269d-4fdf-44de-8d92-962fe4f88e97-kube-api-access-pv6b6\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.022858 kubelet[3463]: I0302 12:57:32.021234 3463 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6c3c269d-4fdf-44de-8d92-962fe4f88e97-cilium-config-path\") pod \"cilium-wcx4h\" (UID: \"6c3c269d-4fdf-44de-8d92-962fe4f88e97\") " pod="kube-system/cilium-wcx4h" Mar 2 12:57:32.025247 systemd[1]: session-24.scope: Deactivated successfully. Mar 2 12:57:32.025998 systemd-logind[1868]: Session 24 logged out. Waiting for processes to exit. Mar 2 12:57:32.027383 systemd-logind[1868]: Removed session 24. Mar 2 12:57:32.104437 systemd[1]: Started sshd@22-10.200.20.30:22-10.200.16.10:45892.service - OpenSSH per-connection server daemon (10.200.16.10:45892). Mar 2 12:57:32.281719 containerd[1890]: time="2026-03-02T12:57:32.281262538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wcx4h,Uid:6c3c269d-4fdf-44de-8d92-962fe4f88e97,Namespace:kube-system,Attempt:0,}" Mar 2 12:57:32.317912 containerd[1890]: time="2026-03-02T12:57:32.317840158Z" level=info msg="connecting to shim 9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92" address="unix:///run/containerd/s/357e8f80a9c3fa94ee79fe9f051fa55eb18d6e1c63ec7866cf3ecd4ec9c54531" namespace=k8s.io protocol=ttrpc version=3 Mar 2 12:57:32.337552 systemd[1]: Started cri-containerd-9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92.scope - libcontainer container 9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92. Mar 2 12:57:32.359990 containerd[1890]: time="2026-03-02T12:57:32.359953178Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wcx4h,Uid:6c3c269d-4fdf-44de-8d92-962fe4f88e97,Namespace:kube-system,Attempt:0,} returns sandbox id \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\"" Mar 2 12:57:32.377772 containerd[1890]: time="2026-03-02T12:57:32.377734732Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 2 12:57:32.395844 containerd[1890]: time="2026-03-02T12:57:32.395809248Z" level=info msg="Container 9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:57:32.410237 containerd[1890]: time="2026-03-02T12:57:32.410201380Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3\"" Mar 2 12:57:32.411561 containerd[1890]: time="2026-03-02T12:57:32.411522436Z" level=info msg="StartContainer for \"9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3\"" Mar 2 12:57:32.412469 containerd[1890]: time="2026-03-02T12:57:32.412443144Z" level=info msg="connecting to shim 9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3" address="unix:///run/containerd/s/357e8f80a9c3fa94ee79fe9f051fa55eb18d6e1c63ec7866cf3ecd4ec9c54531" protocol=ttrpc version=3 Mar 2 12:57:32.427531 systemd[1]: Started cri-containerd-9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3.scope - libcontainer container 9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3. Mar 2 12:57:32.454738 containerd[1890]: time="2026-03-02T12:57:32.454643262Z" level=info msg="StartContainer for \"9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3\" returns successfully" Mar 2 12:57:32.459022 systemd[1]: cri-containerd-9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3.scope: Deactivated successfully. Mar 2 12:57:32.462624 containerd[1890]: time="2026-03-02T12:57:32.462337616Z" level=info msg="received container exit event container_id:\"9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3\" id:\"9ae7e85e4a1273d62f01128f8474adf61f7aa123a7dcd3bda47e8376e126d1f3\" pid:5211 exited_at:{seconds:1772456252 nanos:462133409}" Mar 2 12:57:32.554146 sshd[5145]: Accepted publickey for core from 10.200.16.10 port 45892 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:32.555639 sshd-session[5145]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:32.559865 systemd-logind[1868]: New session 25 of user core. Mar 2 12:57:32.569527 systemd[1]: Started session-25.scope - Session 25 of User core. Mar 2 12:57:32.785125 sshd[5241]: Connection closed by 10.200.16.10 port 45892 Mar 2 12:57:32.785744 sshd-session[5145]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:32.788955 systemd[1]: sshd@22-10.200.20.30:22-10.200.16.10:45892.service: Deactivated successfully. Mar 2 12:57:32.790397 systemd[1]: session-25.scope: Deactivated successfully. Mar 2 12:57:32.791075 systemd-logind[1868]: Session 25 logged out. Waiting for processes to exit. Mar 2 12:57:32.792367 systemd-logind[1868]: Removed session 25. Mar 2 12:57:32.877078 systemd[1]: Started sshd@23-10.200.20.30:22-10.200.16.10:45898.service - OpenSSH per-connection server daemon (10.200.16.10:45898). Mar 2 12:57:33.297519 sshd[5248]: Accepted publickey for core from 10.200.16.10 port 45898 ssh2: RSA SHA256:7ukVy6tXsczvRkKnjXS5ykZo8M2KdxhCNukcDYzlKCM Mar 2 12:57:33.300267 sshd-session[5248]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Mar 2 12:57:33.305965 containerd[1890]: time="2026-03-02T12:57:33.305889228Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 2 12:57:33.306044 systemd-logind[1868]: New session 26 of user core. Mar 2 12:57:33.312527 systemd[1]: Started session-26.scope - Session 26 of User core. Mar 2 12:57:33.329426 containerd[1890]: time="2026-03-02T12:57:33.329330706Z" level=info msg="Container 3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:57:33.344453 containerd[1890]: time="2026-03-02T12:57:33.344391907Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216\"" Mar 2 12:57:33.345056 containerd[1890]: time="2026-03-02T12:57:33.345028678Z" level=info msg="StartContainer for \"3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216\"" Mar 2 12:57:33.345767 containerd[1890]: time="2026-03-02T12:57:33.345744644Z" level=info msg="connecting to shim 3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216" address="unix:///run/containerd/s/357e8f80a9c3fa94ee79fe9f051fa55eb18d6e1c63ec7866cf3ecd4ec9c54531" protocol=ttrpc version=3 Mar 2 12:57:33.355519 kubelet[3463]: E0302 12:57:33.355486 3463 kubelet.go:3117] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 2 12:57:33.364532 systemd[1]: Started cri-containerd-3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216.scope - libcontainer container 3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216. Mar 2 12:57:33.390633 systemd[1]: cri-containerd-3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216.scope: Deactivated successfully. Mar 2 12:57:33.392977 containerd[1890]: time="2026-03-02T12:57:33.392941545Z" level=info msg="received container exit event container_id:\"3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216\" id:\"3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216\" pid:5265 exited_at:{seconds:1772456253 nanos:391148499}" Mar 2 12:57:33.393966 containerd[1890]: time="2026-03-02T12:57:33.393947528Z" level=info msg="StartContainer for \"3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216\" returns successfully" Mar 2 12:57:34.125023 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3771798bc2e858314dd5fbecd0f224f4af1c3e6423089bc25ceaa5c3e2fe6216-rootfs.mount: Deactivated successfully. Mar 2 12:57:34.308806 containerd[1890]: time="2026-03-02T12:57:34.308762483Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 2 12:57:34.333585 containerd[1890]: time="2026-03-02T12:57:34.333540522Z" level=info msg="Container 8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:57:34.350436 containerd[1890]: time="2026-03-02T12:57:34.350336311Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4\"" Mar 2 12:57:34.351278 containerd[1890]: time="2026-03-02T12:57:34.351263251Z" level=info msg="StartContainer for \"8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4\"" Mar 2 12:57:34.352611 containerd[1890]: time="2026-03-02T12:57:34.352590683Z" level=info msg="connecting to shim 8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4" address="unix:///run/containerd/s/357e8f80a9c3fa94ee79fe9f051fa55eb18d6e1c63ec7866cf3ecd4ec9c54531" protocol=ttrpc version=3 Mar 2 12:57:34.370536 systemd[1]: Started cri-containerd-8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4.scope - libcontainer container 8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4. Mar 2 12:57:34.428220 systemd[1]: cri-containerd-8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4.scope: Deactivated successfully. Mar 2 12:57:34.431582 containerd[1890]: time="2026-03-02T12:57:34.431533371Z" level=info msg="received container exit event container_id:\"8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4\" id:\"8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4\" pid:5313 exited_at:{seconds:1772456254 nanos:430218867}" Mar 2 12:57:34.440658 containerd[1890]: time="2026-03-02T12:57:34.440620278Z" level=info msg="StartContainer for \"8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4\" returns successfully" Mar 2 12:57:34.451810 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8cdd5cc70d82c041fa0cd2ec12c38b71d45a0c2fdf34f463b70120c8d9d091c4-rootfs.mount: Deactivated successfully. Mar 2 12:57:35.312608 containerd[1890]: time="2026-03-02T12:57:35.312561902Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 2 12:57:35.334084 containerd[1890]: time="2026-03-02T12:57:35.333637221Z" level=info msg="Container eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:57:35.347931 containerd[1890]: time="2026-03-02T12:57:35.347890709Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a\"" Mar 2 12:57:35.348621 containerd[1890]: time="2026-03-02T12:57:35.348596138Z" level=info msg="StartContainer for \"eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a\"" Mar 2 12:57:35.351261 containerd[1890]: time="2026-03-02T12:57:35.351225042Z" level=info msg="connecting to shim eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a" address="unix:///run/containerd/s/357e8f80a9c3fa94ee79fe9f051fa55eb18d6e1c63ec7866cf3ecd4ec9c54531" protocol=ttrpc version=3 Mar 2 12:57:35.370541 systemd[1]: Started cri-containerd-eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a.scope - libcontainer container eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a. Mar 2 12:57:35.389175 systemd[1]: cri-containerd-eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a.scope: Deactivated successfully. Mar 2 12:57:35.396642 containerd[1890]: time="2026-03-02T12:57:35.396532606Z" level=info msg="received container exit event container_id:\"eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a\" id:\"eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a\" pid:5354 exited_at:{seconds:1772456255 nanos:390901268}" Mar 2 12:57:35.402035 containerd[1890]: time="2026-03-02T12:57:35.401888257Z" level=info msg="StartContainer for \"eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a\" returns successfully" Mar 2 12:57:35.411954 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-eddd4fe73873ae7be50e2052bb4b2b248758c47af3a9bb9fe485d58431411b6a-rootfs.mount: Deactivated successfully. Mar 2 12:57:36.316835 containerd[1890]: time="2026-03-02T12:57:36.316739492Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 2 12:57:36.336972 containerd[1890]: time="2026-03-02T12:57:36.336941624Z" level=info msg="Container 0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d: CDI devices from CRI Config.CDIDevices: []" Mar 2 12:57:36.357553 containerd[1890]: time="2026-03-02T12:57:36.357479022Z" level=info msg="CreateContainer within sandbox \"9f2cb9594277034ffc884b256c97f0cba808e34d0f4f5d9fa6e2034c0ef73a92\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d\"" Mar 2 12:57:36.358651 containerd[1890]: time="2026-03-02T12:57:36.358582576Z" level=info msg="StartContainer for \"0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d\"" Mar 2 12:57:36.360140 containerd[1890]: time="2026-03-02T12:57:36.360115518Z" level=info msg="connecting to shim 0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d" address="unix:///run/containerd/s/357e8f80a9c3fa94ee79fe9f051fa55eb18d6e1c63ec7866cf3ecd4ec9c54531" protocol=ttrpc version=3 Mar 2 12:57:36.379524 systemd[1]: Started cri-containerd-0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d.scope - libcontainer container 0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d. Mar 2 12:57:36.416492 containerd[1890]: time="2026-03-02T12:57:36.416459169Z" level=info msg="StartContainer for \"0ae7dd3810f52f827a7e18180c3f2c9ca75874816d281164526d8fb31c0c3b7d\" returns successfully" Mar 2 12:57:36.423603 kubelet[3463]: I0302 12:57:36.422655 3463 setters.go:618] "Node became not ready" node="ci-4459.2.101-f64d7d9b2a" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-02T12:57:36Z","lastTransitionTime":"2026-03-02T12:57:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Mar 2 12:57:36.720439 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Mar 2 12:57:39.065870 systemd-networkd[1485]: lxc_health: Link UP Mar 2 12:57:39.074863 systemd-networkd[1485]: lxc_health: Gained carrier Mar 2 12:57:39.720196 kubelet[3463]: E0302 12:57:39.720127 3463 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 127.0.0.1:42838->127.0.0.1:34963: write tcp 127.0.0.1:42838->127.0.0.1:34963: write: broken pipe Mar 2 12:57:40.306990 kubelet[3463]: I0302 12:57:40.306916 3463 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-wcx4h" podStartSLOduration=9.306811406 podStartE2EDuration="9.306811406s" podCreationTimestamp="2026-03-02 12:57:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-02 12:57:37.329651644 +0000 UTC m=+144.407388928" watchObservedRunningTime="2026-03-02 12:57:40.306811406 +0000 UTC m=+147.384548690" Mar 2 12:57:40.978541 systemd-networkd[1485]: lxc_health: Gained IPv6LL Mar 2 12:57:46.037230 sshd[5251]: Connection closed by 10.200.16.10 port 45898 Mar 2 12:57:46.038586 sshd-session[5248]: pam_unix(sshd:session): session closed for user core Mar 2 12:57:46.041793 systemd-logind[1868]: Session 26 logged out. Waiting for processes to exit. Mar 2 12:57:46.042435 systemd[1]: sshd@23-10.200.20.30:22-10.200.16.10:45898.service: Deactivated successfully. Mar 2 12:57:46.046385 systemd[1]: session-26.scope: Deactivated successfully. Mar 2 12:57:46.049038 systemd-logind[1868]: Removed session 26.