Mar 17 18:23:20.743545 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Mar 17 18:23:20.743568 kernel: Linux version 5.15.179-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Mon Mar 17 17:11:44 -00 2025 Mar 17 18:23:20.743575 kernel: efi: EFI v2.70 by EDK II Mar 17 18:23:20.743581 kernel: efi: SMBIOS 3.0=0xd9260000 ACPI 2.0=0xd9240000 MEMATTR=0xda32b018 RNG=0xd9220018 MEMRESERVE=0xd9521c18 Mar 17 18:23:20.743585 kernel: random: crng init done Mar 17 18:23:20.743591 kernel: ACPI: Early table checksum verification disabled Mar 17 18:23:20.743597 kernel: ACPI: RSDP 0x00000000D9240000 000024 (v02 BOCHS ) Mar 17 18:23:20.743604 kernel: ACPI: XSDT 0x00000000D9230000 000064 (v01 BOCHS BXPC 00000001 01000013) Mar 17 18:23:20.743609 kernel: ACPI: FACP 0x00000000D91E0000 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743615 kernel: ACPI: DSDT 0x00000000D91F0000 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743620 kernel: ACPI: APIC 0x00000000D91D0000 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743626 kernel: ACPI: PPTT 0x00000000D91C0000 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743663 kernel: ACPI: GTDT 0x00000000D91B0000 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743669 kernel: ACPI: MCFG 0x00000000D91A0000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743679 kernel: ACPI: SPCR 0x00000000D9190000 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743685 kernel: ACPI: DBG2 0x00000000D9180000 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743691 kernel: ACPI: IORT 0x00000000D9170000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Mar 17 18:23:20.743697 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Mar 17 18:23:20.743703 kernel: NUMA: Failed to initialise from firmware Mar 17 18:23:20.743709 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Mar 17 18:23:20.743715 kernel: NUMA: NODE_DATA [mem 0xdcb0c900-0xdcb11fff] Mar 17 18:23:20.743720 kernel: Zone ranges: Mar 17 18:23:20.743726 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Mar 17 18:23:20.743733 kernel: DMA32 empty Mar 17 18:23:20.743738 kernel: Normal empty Mar 17 18:23:20.743744 kernel: Movable zone start for each node Mar 17 18:23:20.743750 kernel: Early memory node ranges Mar 17 18:23:20.743756 kernel: node 0: [mem 0x0000000040000000-0x00000000d924ffff] Mar 17 18:23:20.743761 kernel: node 0: [mem 0x00000000d9250000-0x00000000d951ffff] Mar 17 18:23:20.743767 kernel: node 0: [mem 0x00000000d9520000-0x00000000dc7fffff] Mar 17 18:23:20.743773 kernel: node 0: [mem 0x00000000dc800000-0x00000000dc88ffff] Mar 17 18:23:20.743779 kernel: node 0: [mem 0x00000000dc890000-0x00000000dc89ffff] Mar 17 18:23:20.743785 kernel: node 0: [mem 0x00000000dc8a0000-0x00000000dc9bffff] Mar 17 18:23:20.743790 kernel: node 0: [mem 0x00000000dc9c0000-0x00000000dcffffff] Mar 17 18:23:20.743796 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Mar 17 18:23:20.743803 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Mar 17 18:23:20.743809 kernel: psci: probing for conduit method from ACPI. Mar 17 18:23:20.743814 kernel: psci: PSCIv1.1 detected in firmware. Mar 17 18:23:20.743820 kernel: psci: Using standard PSCI v0.2 function IDs Mar 17 18:23:20.743826 kernel: psci: Trusted OS migration not required Mar 17 18:23:20.743834 kernel: psci: SMC Calling Convention v1.1 Mar 17 18:23:20.743841 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Mar 17 18:23:20.743848 kernel: ACPI: SRAT not present Mar 17 18:23:20.743854 kernel: percpu: Embedded 30 pages/cpu s83032 r8192 d31656 u122880 Mar 17 18:23:20.743860 kernel: pcpu-alloc: s83032 r8192 d31656 u122880 alloc=30*4096 Mar 17 18:23:20.743867 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Mar 17 18:23:20.743873 kernel: Detected PIPT I-cache on CPU0 Mar 17 18:23:20.743879 kernel: CPU features: detected: GIC system register CPU interface Mar 17 18:23:20.743885 kernel: CPU features: detected: Hardware dirty bit management Mar 17 18:23:20.743891 kernel: CPU features: detected: Spectre-v4 Mar 17 18:23:20.743897 kernel: CPU features: detected: Spectre-BHB Mar 17 18:23:20.743904 kernel: CPU features: kernel page table isolation forced ON by KASLR Mar 17 18:23:20.743911 kernel: CPU features: detected: Kernel page table isolation (KPTI) Mar 17 18:23:20.743917 kernel: CPU features: detected: ARM erratum 1418040 Mar 17 18:23:20.743923 kernel: CPU features: detected: SSBS not fully self-synchronizing Mar 17 18:23:20.743929 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Mar 17 18:23:20.743935 kernel: Policy zone: DMA Mar 17 18:23:20.744105 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=e034db32d58fe7496a3db6ba3879dd9052cea2cf1597d65edfc7b26afc92530d Mar 17 18:23:20.744114 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Mar 17 18:23:20.744120 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Mar 17 18:23:20.744127 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Mar 17 18:23:20.744133 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Mar 17 18:23:20.744144 kernel: Memory: 2457408K/2572288K available (9792K kernel code, 2094K rwdata, 7584K rodata, 36416K init, 777K bss, 114880K reserved, 0K cma-reserved) Mar 17 18:23:20.744151 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Mar 17 18:23:20.744157 kernel: trace event string verifier disabled Mar 17 18:23:20.744163 kernel: rcu: Preemptible hierarchical RCU implementation. Mar 17 18:23:20.744170 kernel: rcu: RCU event tracing is enabled. Mar 17 18:23:20.744176 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Mar 17 18:23:20.744183 kernel: Trampoline variant of Tasks RCU enabled. Mar 17 18:23:20.744189 kernel: Tracing variant of Tasks RCU enabled. Mar 17 18:23:20.744196 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Mar 17 18:23:20.744202 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Mar 17 18:23:20.744208 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Mar 17 18:23:20.744215 kernel: GICv3: 256 SPIs implemented Mar 17 18:23:20.744221 kernel: GICv3: 0 Extended SPIs implemented Mar 17 18:23:20.744228 kernel: GICv3: Distributor has no Range Selector support Mar 17 18:23:20.744234 kernel: Root IRQ handler: gic_handle_irq Mar 17 18:23:20.744240 kernel: GICv3: 16 PPIs implemented Mar 17 18:23:20.744246 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Mar 17 18:23:20.744252 kernel: ACPI: SRAT not present Mar 17 18:23:20.744258 kernel: ITS [mem 0x08080000-0x0809ffff] Mar 17 18:23:20.744264 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400b0000 (indirect, esz 8, psz 64K, shr 1) Mar 17 18:23:20.744271 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400c0000 (flat, esz 8, psz 64K, shr 1) Mar 17 18:23:20.744277 kernel: GICv3: using LPI property table @0x00000000400d0000 Mar 17 18:23:20.744283 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000000400e0000 Mar 17 18:23:20.744291 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 17 18:23:20.744297 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Mar 17 18:23:20.744303 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Mar 17 18:23:20.744310 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Mar 17 18:23:20.744316 kernel: arm-pv: using stolen time PV Mar 17 18:23:20.744322 kernel: Console: colour dummy device 80x25 Mar 17 18:23:20.744352 kernel: ACPI: Core revision 20210730 Mar 17 18:23:20.744360 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Mar 17 18:23:20.744366 kernel: pid_max: default: 32768 minimum: 301 Mar 17 18:23:20.744372 kernel: LSM: Security Framework initializing Mar 17 18:23:20.744381 kernel: SELinux: Initializing. Mar 17 18:23:20.744387 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 17 18:23:20.744394 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Mar 17 18:23:20.744400 kernel: rcu: Hierarchical SRCU implementation. Mar 17 18:23:20.744409 kernel: Platform MSI: ITS@0x8080000 domain created Mar 17 18:23:20.744416 kernel: PCI/MSI: ITS@0x8080000 domain created Mar 17 18:23:20.744422 kernel: Remapping and enabling EFI services. Mar 17 18:23:20.744428 kernel: smp: Bringing up secondary CPUs ... Mar 17 18:23:20.744434 kernel: Detected PIPT I-cache on CPU1 Mar 17 18:23:20.744442 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Mar 17 18:23:20.744448 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000000400f0000 Mar 17 18:23:20.744455 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 17 18:23:20.744461 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Mar 17 18:23:20.744467 kernel: Detected PIPT I-cache on CPU2 Mar 17 18:23:20.744474 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Mar 17 18:23:20.744480 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040100000 Mar 17 18:23:20.744487 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 17 18:23:20.744493 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Mar 17 18:23:20.744499 kernel: Detected PIPT I-cache on CPU3 Mar 17 18:23:20.744594 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Mar 17 18:23:20.744603 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040110000 Mar 17 18:23:20.744610 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Mar 17 18:23:20.744616 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Mar 17 18:23:20.744656 kernel: smp: Brought up 1 node, 4 CPUs Mar 17 18:23:20.744666 kernel: SMP: Total of 4 processors activated. Mar 17 18:23:20.744673 kernel: CPU features: detected: 32-bit EL0 Support Mar 17 18:23:20.744679 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Mar 17 18:23:20.744686 kernel: CPU features: detected: Common not Private translations Mar 17 18:23:20.744693 kernel: CPU features: detected: CRC32 instructions Mar 17 18:23:20.744699 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Mar 17 18:23:20.744706 kernel: CPU features: detected: LSE atomic instructions Mar 17 18:23:20.744714 kernel: CPU features: detected: Privileged Access Never Mar 17 18:23:20.744721 kernel: CPU features: detected: RAS Extension Support Mar 17 18:23:20.744727 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Mar 17 18:23:20.744734 kernel: CPU: All CPU(s) started at EL1 Mar 17 18:23:20.744740 kernel: alternatives: patching kernel code Mar 17 18:23:20.744749 kernel: devtmpfs: initialized Mar 17 18:23:20.744755 kernel: KASLR enabled Mar 17 18:23:20.744762 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Mar 17 18:23:20.744769 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Mar 17 18:23:20.744775 kernel: pinctrl core: initialized pinctrl subsystem Mar 17 18:23:20.744782 kernel: SMBIOS 3.0.0 present. Mar 17 18:23:20.744788 kernel: DMI: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 Mar 17 18:23:20.744795 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Mar 17 18:23:20.744801 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Mar 17 18:23:20.744810 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Mar 17 18:23:20.744816 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Mar 17 18:23:20.744823 kernel: audit: initializing netlink subsys (disabled) Mar 17 18:23:20.744830 kernel: audit: type=2000 audit(0.032:1): state=initialized audit_enabled=0 res=1 Mar 17 18:23:20.744836 kernel: thermal_sys: Registered thermal governor 'step_wise' Mar 17 18:23:20.744843 kernel: cpuidle: using governor menu Mar 17 18:23:20.744849 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Mar 17 18:23:20.744856 kernel: ASID allocator initialised with 32768 entries Mar 17 18:23:20.744862 kernel: ACPI: bus type PCI registered Mar 17 18:23:20.744870 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Mar 17 18:23:20.744877 kernel: Serial: AMBA PL011 UART driver Mar 17 18:23:20.744884 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Mar 17 18:23:20.744890 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages Mar 17 18:23:20.744897 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Mar 17 18:23:20.744903 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages Mar 17 18:23:20.744910 kernel: cryptd: max_cpu_qlen set to 1000 Mar 17 18:23:20.744916 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Mar 17 18:23:20.744923 kernel: ACPI: Added _OSI(Module Device) Mar 17 18:23:20.744931 kernel: ACPI: Added _OSI(Processor Device) Mar 17 18:23:20.744938 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Mar 17 18:23:20.744944 kernel: ACPI: Added _OSI(Processor Aggregator Device) Mar 17 18:23:20.744951 kernel: ACPI: Added _OSI(Linux-Dell-Video) Mar 17 18:23:20.744958 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Mar 17 18:23:20.744965 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Mar 17 18:23:20.744971 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Mar 17 18:23:20.744978 kernel: ACPI: Interpreter enabled Mar 17 18:23:20.744984 kernel: ACPI: Using GIC for interrupt routing Mar 17 18:23:20.744992 kernel: ACPI: MCFG table detected, 1 entries Mar 17 18:23:20.744999 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Mar 17 18:23:20.745010 kernel: printk: console [ttyAMA0] enabled Mar 17 18:23:20.745017 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Mar 17 18:23:20.745305 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Mar 17 18:23:20.745399 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Mar 17 18:23:20.745469 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Mar 17 18:23:20.745532 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Mar 17 18:23:20.745589 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Mar 17 18:23:20.745598 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Mar 17 18:23:20.745605 kernel: PCI host bridge to bus 0000:00 Mar 17 18:23:20.745864 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Mar 17 18:23:20.745931 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Mar 17 18:23:20.745984 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Mar 17 18:23:20.746048 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Mar 17 18:23:20.746131 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Mar 17 18:23:20.746207 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Mar 17 18:23:20.749031 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Mar 17 18:23:20.749103 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Mar 17 18:23:20.749166 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Mar 17 18:23:20.749228 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Mar 17 18:23:20.749296 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Mar 17 18:23:20.749634 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Mar 17 18:23:20.749776 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Mar 17 18:23:20.750114 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Mar 17 18:23:20.750233 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Mar 17 18:23:20.750245 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Mar 17 18:23:20.750253 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Mar 17 18:23:20.750260 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Mar 17 18:23:20.750274 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Mar 17 18:23:20.750281 kernel: iommu: Default domain type: Translated Mar 17 18:23:20.750288 kernel: iommu: DMA domain TLB invalidation policy: strict mode Mar 17 18:23:20.750295 kernel: vgaarb: loaded Mar 17 18:23:20.750302 kernel: pps_core: LinuxPPS API ver. 1 registered Mar 17 18:23:20.750309 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Mar 17 18:23:20.750316 kernel: PTP clock support registered Mar 17 18:23:20.750323 kernel: Registered efivars operations Mar 17 18:23:20.750371 kernel: clocksource: Switched to clocksource arch_sys_counter Mar 17 18:23:20.750382 kernel: VFS: Disk quotas dquot_6.6.0 Mar 17 18:23:20.750389 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Mar 17 18:23:20.750396 kernel: pnp: PnP ACPI init Mar 17 18:23:20.750653 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Mar 17 18:23:20.750671 kernel: pnp: PnP ACPI: found 1 devices Mar 17 18:23:20.750678 kernel: NET: Registered PF_INET protocol family Mar 17 18:23:20.750685 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Mar 17 18:23:20.750693 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Mar 17 18:23:20.750705 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Mar 17 18:23:20.750712 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Mar 17 18:23:20.750719 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Mar 17 18:23:20.750726 kernel: TCP: Hash tables configured (established 32768 bind 32768) Mar 17 18:23:20.750733 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 17 18:23:20.750740 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Mar 17 18:23:20.750747 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Mar 17 18:23:20.750754 kernel: PCI: CLS 0 bytes, default 64 Mar 17 18:23:20.750761 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Mar 17 18:23:20.750769 kernel: kvm [1]: HYP mode not available Mar 17 18:23:20.750776 kernel: Initialise system trusted keyrings Mar 17 18:23:20.750783 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Mar 17 18:23:20.750790 kernel: Key type asymmetric registered Mar 17 18:23:20.750797 kernel: Asymmetric key parser 'x509' registered Mar 17 18:23:20.750804 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Mar 17 18:23:20.750810 kernel: io scheduler mq-deadline registered Mar 17 18:23:20.750817 kernel: io scheduler kyber registered Mar 17 18:23:20.750824 kernel: io scheduler bfq registered Mar 17 18:23:20.750832 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Mar 17 18:23:20.750839 kernel: ACPI: button: Power Button [PWRB] Mar 17 18:23:20.750846 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Mar 17 18:23:20.750924 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Mar 17 18:23:20.750941 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Mar 17 18:23:20.750948 kernel: thunder_xcv, ver 1.0 Mar 17 18:23:20.750956 kernel: thunder_bgx, ver 1.0 Mar 17 18:23:20.750963 kernel: nicpf, ver 1.0 Mar 17 18:23:20.750970 kernel: nicvf, ver 1.0 Mar 17 18:23:20.751251 kernel: rtc-efi rtc-efi.0: registered as rtc0 Mar 17 18:23:20.751340 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-03-17T18:23:20 UTC (1742235800) Mar 17 18:23:20.751351 kernel: hid: raw HID events driver (C) Jiri Kosina Mar 17 18:23:20.751358 kernel: NET: Registered PF_INET6 protocol family Mar 17 18:23:20.751365 kernel: Segment Routing with IPv6 Mar 17 18:23:20.751371 kernel: In-situ OAM (IOAM) with IPv6 Mar 17 18:23:20.751378 kernel: NET: Registered PF_PACKET protocol family Mar 17 18:23:20.751385 kernel: Key type dns_resolver registered Mar 17 18:23:20.751397 kernel: registered taskstats version 1 Mar 17 18:23:20.751404 kernel: Loading compiled-in X.509 certificates Mar 17 18:23:20.751411 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.179-flatcar: c6f3fb83dc6bb7052b07ec5b1ef41d12f9b3f7e4' Mar 17 18:23:20.751418 kernel: Key type .fscrypt registered Mar 17 18:23:20.751424 kernel: Key type fscrypt-provisioning registered Mar 17 18:23:20.751431 kernel: ima: No TPM chip found, activating TPM-bypass! Mar 17 18:23:20.751438 kernel: ima: Allocated hash algorithm: sha1 Mar 17 18:23:20.751444 kernel: ima: No architecture policies found Mar 17 18:23:20.751451 kernel: clk: Disabling unused clocks Mar 17 18:23:20.751459 kernel: Freeing unused kernel memory: 36416K Mar 17 18:23:20.751466 kernel: Run /init as init process Mar 17 18:23:20.751472 kernel: with arguments: Mar 17 18:23:20.751479 kernel: /init Mar 17 18:23:20.751485 kernel: with environment: Mar 17 18:23:20.751492 kernel: HOME=/ Mar 17 18:23:20.751498 kernel: TERM=linux Mar 17 18:23:20.751505 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Mar 17 18:23:20.751514 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Mar 17 18:23:20.752207 systemd[1]: Detected virtualization kvm. Mar 17 18:23:20.752400 systemd[1]: Detected architecture arm64. Mar 17 18:23:20.752408 systemd[1]: Running in initrd. Mar 17 18:23:20.752415 systemd[1]: No hostname configured, using default hostname. Mar 17 18:23:20.752423 systemd[1]: Hostname set to . Mar 17 18:23:20.752431 systemd[1]: Initializing machine ID from VM UUID. Mar 17 18:23:20.752438 systemd[1]: Queued start job for default target initrd.target. Mar 17 18:23:20.752451 systemd[1]: Started systemd-ask-password-console.path. Mar 17 18:23:20.752458 systemd[1]: Reached target cryptsetup.target. Mar 17 18:23:20.752465 systemd[1]: Reached target paths.target. Mar 17 18:23:20.752472 systemd[1]: Reached target slices.target. Mar 17 18:23:20.752479 systemd[1]: Reached target swap.target. Mar 17 18:23:20.752485 systemd[1]: Reached target timers.target. Mar 17 18:23:20.752493 systemd[1]: Listening on iscsid.socket. Mar 17 18:23:20.752502 systemd[1]: Listening on iscsiuio.socket. Mar 17 18:23:20.752509 systemd[1]: Listening on systemd-journald-audit.socket. Mar 17 18:23:20.752516 systemd[1]: Listening on systemd-journald-dev-log.socket. Mar 17 18:23:20.752523 systemd[1]: Listening on systemd-journald.socket. Mar 17 18:23:20.752530 systemd[1]: Listening on systemd-networkd.socket. Mar 17 18:23:20.752538 systemd[1]: Listening on systemd-udevd-control.socket. Mar 17 18:23:20.752545 systemd[1]: Listening on systemd-udevd-kernel.socket. Mar 17 18:23:20.752552 systemd[1]: Reached target sockets.target. Mar 17 18:23:20.752559 systemd[1]: Starting kmod-static-nodes.service... Mar 17 18:23:20.752568 systemd[1]: Finished network-cleanup.service. Mar 17 18:23:20.752609 systemd[1]: Starting systemd-fsck-usr.service... Mar 17 18:23:20.752616 systemd[1]: Starting systemd-journald.service... Mar 17 18:23:20.752623 systemd[1]: Starting systemd-modules-load.service... Mar 17 18:23:20.752630 systemd[1]: Starting systemd-resolved.service... Mar 17 18:23:20.752638 systemd[1]: Starting systemd-vconsole-setup.service... Mar 17 18:23:20.752645 systemd[1]: Finished kmod-static-nodes.service. Mar 17 18:23:20.752653 systemd[1]: Finished systemd-fsck-usr.service. Mar 17 18:23:20.752660 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Mar 17 18:23:20.752669 systemd[1]: Finished systemd-vconsole-setup.service. Mar 17 18:23:20.752676 systemd[1]: Starting dracut-cmdline-ask.service... Mar 17 18:23:20.752684 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Mar 17 18:23:20.752692 kernel: audit: type=1130 audit(1742235800.746:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.752700 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Mar 17 18:23:20.752710 systemd-journald[291]: Journal started Mar 17 18:23:20.752915 systemd-journald[291]: Runtime Journal (/run/log/journal/a025cd63b78d4273a9d395da5abdbdbc) is 6.0M, max 48.7M, 42.6M free. Mar 17 18:23:20.746000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.731395 systemd-modules-load[292]: Inserted module 'overlay' Mar 17 18:23:20.755202 systemd[1]: Started systemd-journald.service. Mar 17 18:23:20.757000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.757972 systemd-resolved[293]: Positive Trust Anchors: Mar 17 18:23:20.760772 kernel: audit: type=1130 audit(1742235800.757:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.760791 kernel: Bridge firewalling registered Mar 17 18:23:20.757979 systemd-resolved[293]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 18:23:20.758015 systemd-resolved[293]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Mar 17 18:23:20.762000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.760032 systemd-modules-load[292]: Inserted module 'br_netfilter' Mar 17 18:23:20.776205 kernel: audit: type=1130 audit(1742235800.762:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.776226 kernel: audit: type=1130 audit(1742235800.772:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.776236 kernel: SCSI subsystem initialized Mar 17 18:23:20.772000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.761061 systemd[1]: Finished dracut-cmdline-ask.service. Mar 17 18:23:20.764855 systemd-resolved[293]: Defaulting to hostname 'linux'. Mar 17 18:23:20.772111 systemd[1]: Started systemd-resolved.service. Mar 17 18:23:20.773029 systemd[1]: Reached target nss-lookup.target. Mar 17 18:23:20.777717 systemd[1]: Starting dracut-cmdline.service... Mar 17 18:23:20.784891 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Mar 17 18:23:20.784928 kernel: device-mapper: uevent: version 1.0.3 Mar 17 18:23:20.784938 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Mar 17 18:23:20.787165 systemd-modules-load[292]: Inserted module 'dm_multipath' Mar 17 18:23:20.787935 systemd[1]: Finished systemd-modules-load.service. Mar 17 18:23:20.788000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.791966 dracut-cmdline[308]: dracut-dracut-053 Mar 17 18:23:20.793565 kernel: audit: type=1130 audit(1742235800.788:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.789701 systemd[1]: Starting systemd-sysctl.service... Mar 17 18:23:20.794413 dracut-cmdline[308]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=e034db32d58fe7496a3db6ba3879dd9052cea2cf1597d65edfc7b26afc92530d Mar 17 18:23:20.797783 systemd[1]: Finished systemd-sysctl.service. Mar 17 18:23:20.799000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.802352 kernel: audit: type=1130 audit(1742235800.799:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.854352 kernel: Loading iSCSI transport class v2.0-870. Mar 17 18:23:20.870357 kernel: iscsi: registered transport (tcp) Mar 17 18:23:20.882547 kernel: iscsi: registered transport (qla4xxx) Mar 17 18:23:20.882580 kernel: QLogic iSCSI HBA Driver Mar 17 18:23:20.916720 systemd[1]: Finished dracut-cmdline.service. Mar 17 18:23:20.917000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.918370 systemd[1]: Starting dracut-pre-udev.service... Mar 17 18:23:20.921511 kernel: audit: type=1130 audit(1742235800.917:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:20.962350 kernel: raid6: neonx8 gen() 13722 MB/s Mar 17 18:23:20.979342 kernel: raid6: neonx8 xor() 10749 MB/s Mar 17 18:23:20.996343 kernel: raid6: neonx4 gen() 13521 MB/s Mar 17 18:23:21.013343 kernel: raid6: neonx4 xor() 11271 MB/s Mar 17 18:23:21.030344 kernel: raid6: neonx2 gen() 12984 MB/s Mar 17 18:23:21.047341 kernel: raid6: neonx2 xor() 10360 MB/s Mar 17 18:23:21.064346 kernel: raid6: neonx1 gen() 10523 MB/s Mar 17 18:23:21.081339 kernel: raid6: neonx1 xor() 8784 MB/s Mar 17 18:23:21.098339 kernel: raid6: int64x8 gen() 6268 MB/s Mar 17 18:23:21.115341 kernel: raid6: int64x8 xor() 3544 MB/s Mar 17 18:23:21.132340 kernel: raid6: int64x4 gen() 7224 MB/s Mar 17 18:23:21.149342 kernel: raid6: int64x4 xor() 3849 MB/s Mar 17 18:23:21.166342 kernel: raid6: int64x2 gen() 6149 MB/s Mar 17 18:23:21.183339 kernel: raid6: int64x2 xor() 3321 MB/s Mar 17 18:23:21.200362 kernel: raid6: int64x1 gen() 5034 MB/s Mar 17 18:23:21.217546 kernel: raid6: int64x1 xor() 2645 MB/s Mar 17 18:23:21.217561 kernel: raid6: using algorithm neonx8 gen() 13722 MB/s Mar 17 18:23:21.217570 kernel: raid6: .... xor() 10749 MB/s, rmw enabled Mar 17 18:23:21.217578 kernel: raid6: using neon recovery algorithm Mar 17 18:23:21.228697 kernel: xor: measuring software checksum speed Mar 17 18:23:21.228712 kernel: 8regs : 17213 MB/sec Mar 17 18:23:21.228721 kernel: 32regs : 20728 MB/sec Mar 17 18:23:21.229619 kernel: arm64_neon : 27700 MB/sec Mar 17 18:23:21.229632 kernel: xor: using function: arm64_neon (27700 MB/sec) Mar 17 18:23:21.293349 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no Mar 17 18:23:21.305089 systemd[1]: Finished dracut-pre-udev.service. Mar 17 18:23:21.308417 kernel: audit: type=1130 audit(1742235801.305:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:21.308446 kernel: audit: type=1334 audit(1742235801.307:10): prog-id=7 op=LOAD Mar 17 18:23:21.305000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:21.307000 audit: BPF prog-id=7 op=LOAD Mar 17 18:23:21.308000 audit: BPF prog-id=8 op=LOAD Mar 17 18:23:21.309415 systemd[1]: Starting systemd-udevd.service... Mar 17 18:23:21.321495 systemd-udevd[490]: Using default interface naming scheme 'v252'. Mar 17 18:23:21.324779 systemd[1]: Started systemd-udevd.service. Mar 17 18:23:21.325000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:21.330226 systemd[1]: Starting dracut-pre-trigger.service... Mar 17 18:23:21.341695 dracut-pre-trigger[506]: rd.md=0: removing MD RAID activation Mar 17 18:23:21.368142 systemd[1]: Finished dracut-pre-trigger.service. Mar 17 18:23:21.368000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:21.369632 systemd[1]: Starting systemd-udev-trigger.service... Mar 17 18:23:21.402761 systemd[1]: Finished systemd-udev-trigger.service. Mar 17 18:23:21.403000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:21.427350 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Mar 17 18:23:21.434427 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Mar 17 18:23:21.434442 kernel: GPT:9289727 != 19775487 Mar 17 18:23:21.434455 kernel: GPT:Alternate GPT header not at the end of the disk. Mar 17 18:23:21.434464 kernel: GPT:9289727 != 19775487 Mar 17 18:23:21.434472 kernel: GPT: Use GNU Parted to correct GPT errors. Mar 17 18:23:21.434482 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 18:23:21.446346 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (537) Mar 17 18:23:21.449714 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Mar 17 18:23:21.454458 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Mar 17 18:23:21.455405 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Mar 17 18:23:21.459546 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Mar 17 18:23:21.467210 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Mar 17 18:23:21.469074 systemd[1]: Starting disk-uuid.service... Mar 17 18:23:21.475351 disk-uuid[562]: Primary Header is updated. Mar 17 18:23:21.475351 disk-uuid[562]: Secondary Entries is updated. Mar 17 18:23:21.475351 disk-uuid[562]: Secondary Header is updated. Mar 17 18:23:21.478359 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 18:23:21.490365 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 18:23:22.490353 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Mar 17 18:23:22.490422 disk-uuid[563]: The operation has completed successfully. Mar 17 18:23:22.513533 systemd[1]: disk-uuid.service: Deactivated successfully. Mar 17 18:23:22.513000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.514000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.513622 systemd[1]: Finished disk-uuid.service. Mar 17 18:23:22.517604 systemd[1]: Starting verity-setup.service... Mar 17 18:23:22.533349 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Mar 17 18:23:22.553538 systemd[1]: Found device dev-mapper-usr.device. Mar 17 18:23:22.555638 systemd[1]: Mounting sysusr-usr.mount... Mar 17 18:23:22.557585 systemd[1]: Finished verity-setup.service. Mar 17 18:23:22.558000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.601356 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Mar 17 18:23:22.601889 systemd[1]: Mounted sysusr-usr.mount. Mar 17 18:23:22.603435 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Mar 17 18:23:22.605415 systemd[1]: Starting ignition-setup.service... Mar 17 18:23:22.607275 systemd[1]: Starting parse-ip-for-networkd.service... Mar 17 18:23:22.613481 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Mar 17 18:23:22.613517 kernel: BTRFS info (device vda6): using free space tree Mar 17 18:23:22.613527 kernel: BTRFS info (device vda6): has skinny extents Mar 17 18:23:22.622386 systemd[1]: mnt-oem.mount: Deactivated successfully. Mar 17 18:23:22.628460 systemd[1]: Finished ignition-setup.service. Mar 17 18:23:22.628000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.629940 systemd[1]: Starting ignition-fetch-offline.service... Mar 17 18:23:22.689873 systemd[1]: Finished parse-ip-for-networkd.service. Mar 17 18:23:22.690000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.691000 audit: BPF prog-id=9 op=LOAD Mar 17 18:23:22.692066 systemd[1]: Starting systemd-networkd.service... Mar 17 18:23:22.714710 ignition[651]: Ignition 2.14.0 Mar 17 18:23:22.714719 ignition[651]: Stage: fetch-offline Mar 17 18:23:22.714752 ignition[651]: no configs at "/usr/lib/ignition/base.d" Mar 17 18:23:22.717000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.716076 systemd-networkd[739]: lo: Link UP Mar 17 18:23:22.714761 ignition[651]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 18:23:22.716080 systemd-networkd[739]: lo: Gained carrier Mar 17 18:23:22.714882 ignition[651]: parsed url from cmdline: "" Mar 17 18:23:22.716448 systemd-networkd[739]: Enumeration completed Mar 17 18:23:22.714885 ignition[651]: no config URL provided Mar 17 18:23:22.716620 systemd[1]: Started systemd-networkd.service. Mar 17 18:23:22.714889 ignition[651]: reading system config file "/usr/lib/ignition/user.ign" Mar 17 18:23:22.716627 systemd-networkd[739]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 18:23:22.714895 ignition[651]: no config at "/usr/lib/ignition/user.ign" Mar 17 18:23:22.717691 systemd-networkd[739]: eth0: Link UP Mar 17 18:23:22.714911 ignition[651]: op(1): [started] loading QEMU firmware config module Mar 17 18:23:22.730000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.717695 systemd-networkd[739]: eth0: Gained carrier Mar 17 18:23:22.714918 ignition[651]: op(1): executing: "modprobe" "qemu_fw_cfg" Mar 17 18:23:22.717938 systemd[1]: Reached target network.target. Mar 17 18:23:22.728395 ignition[651]: op(1): [finished] loading QEMU firmware config module Mar 17 18:23:22.720553 systemd[1]: Starting iscsiuio.service... Mar 17 18:23:22.735914 iscsid[745]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Mar 17 18:23:22.735914 iscsid[745]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Mar 17 18:23:22.735914 iscsid[745]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Mar 17 18:23:22.735914 iscsid[745]: If using hardware iscsi like qla4xxx this message can be ignored. Mar 17 18:23:22.735914 iscsid[745]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Mar 17 18:23:22.735914 iscsid[745]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Mar 17 18:23:22.743000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.729791 systemd[1]: Started iscsiuio.service. Mar 17 18:23:22.731502 systemd[1]: Starting iscsid.service... Mar 17 18:23:22.737573 systemd[1]: Started iscsid.service. Mar 17 18:23:22.737602 systemd-networkd[739]: eth0: DHCPv4 address 10.0.0.89/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 17 18:23:22.744714 systemd[1]: Starting dracut-initqueue.service... Mar 17 18:23:22.754467 systemd[1]: Finished dracut-initqueue.service. Mar 17 18:23:22.755000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.755427 systemd[1]: Reached target remote-fs-pre.target. Mar 17 18:23:22.756938 systemd[1]: Reached target remote-cryptsetup.target. Mar 17 18:23:22.758536 systemd[1]: Reached target remote-fs.target. Mar 17 18:23:22.760722 systemd[1]: Starting dracut-pre-mount.service... Mar 17 18:23:22.768057 systemd[1]: Finished dracut-pre-mount.service. Mar 17 18:23:22.768000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.790826 ignition[651]: parsing config with SHA512: 335c91e2c625c835ee8267d5fc43bef49b4e33728a72e195c77bcff62405e6d40a9723783e26eae6b95dbb690245fab3b81331062843eb6f22889e05be633934 Mar 17 18:23:22.797151 unknown[651]: fetched base config from "system" Mar 17 18:23:22.797697 ignition[651]: fetch-offline: fetch-offline passed Mar 17 18:23:22.797161 unknown[651]: fetched user config from "qemu" Mar 17 18:23:22.799000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.797753 ignition[651]: Ignition finished successfully Mar 17 18:23:22.798724 systemd[1]: Finished ignition-fetch-offline.service. Mar 17 18:23:22.800172 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Mar 17 18:23:22.800838 systemd[1]: Starting ignition-kargs.service... Mar 17 18:23:22.809072 ignition[760]: Ignition 2.14.0 Mar 17 18:23:22.809082 ignition[760]: Stage: kargs Mar 17 18:23:22.809170 ignition[760]: no configs at "/usr/lib/ignition/base.d" Mar 17 18:23:22.809179 ignition[760]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 18:23:22.811376 systemd[1]: Finished ignition-kargs.service. Mar 17 18:23:22.812000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.810010 ignition[760]: kargs: kargs passed Mar 17 18:23:22.810052 ignition[760]: Ignition finished successfully Mar 17 18:23:22.813666 systemd[1]: Starting ignition-disks.service... Mar 17 18:23:22.819711 ignition[766]: Ignition 2.14.0 Mar 17 18:23:22.819728 ignition[766]: Stage: disks Mar 17 18:23:22.819815 ignition[766]: no configs at "/usr/lib/ignition/base.d" Mar 17 18:23:22.819826 ignition[766]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 18:23:22.821909 systemd[1]: Finished ignition-disks.service. Mar 17 18:23:22.820679 ignition[766]: disks: disks passed Mar 17 18:23:22.822819 systemd[1]: Reached target initrd-root-device.target. Mar 17 18:23:22.822000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.820722 ignition[766]: Ignition finished successfully Mar 17 18:23:22.824400 systemd[1]: Reached target local-fs-pre.target. Mar 17 18:23:22.825657 systemd[1]: Reached target local-fs.target. Mar 17 18:23:22.826791 systemd[1]: Reached target sysinit.target. Mar 17 18:23:22.828076 systemd[1]: Reached target basic.target. Mar 17 18:23:22.830070 systemd[1]: Starting systemd-fsck-root.service... Mar 17 18:23:22.840336 systemd-fsck[774]: ROOT: clean, 623/553520 files, 56021/553472 blocks Mar 17 18:23:22.845905 systemd[1]: Finished systemd-fsck-root.service. Mar 17 18:23:22.846000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.847510 systemd[1]: Mounting sysroot.mount... Mar 17 18:23:22.854047 systemd[1]: Mounted sysroot.mount. Mar 17 18:23:22.855273 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Mar 17 18:23:22.854818 systemd[1]: Reached target initrd-root-fs.target. Mar 17 18:23:22.856954 systemd[1]: Mounting sysroot-usr.mount... Mar 17 18:23:22.857793 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Mar 17 18:23:22.857830 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Mar 17 18:23:22.857852 systemd[1]: Reached target ignition-diskful.target. Mar 17 18:23:22.859570 systemd[1]: Mounted sysroot-usr.mount. Mar 17 18:23:22.861385 systemd[1]: Starting initrd-setup-root.service... Mar 17 18:23:22.865440 initrd-setup-root[784]: cut: /sysroot/etc/passwd: No such file or directory Mar 17 18:23:22.868849 initrd-setup-root[792]: cut: /sysroot/etc/group: No such file or directory Mar 17 18:23:22.872839 initrd-setup-root[800]: cut: /sysroot/etc/shadow: No such file or directory Mar 17 18:23:22.876763 initrd-setup-root[808]: cut: /sysroot/etc/gshadow: No such file or directory Mar 17 18:23:22.901187 systemd[1]: Finished initrd-setup-root.service. Mar 17 18:23:22.901000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.902683 systemd[1]: Starting ignition-mount.service... Mar 17 18:23:22.903913 systemd[1]: Starting sysroot-boot.service... Mar 17 18:23:22.908682 bash[825]: umount: /sysroot/usr/share/oem: not mounted. Mar 17 18:23:22.918175 ignition[827]: INFO : Ignition 2.14.0 Mar 17 18:23:22.920003 ignition[827]: INFO : Stage: mount Mar 17 18:23:22.920003 ignition[827]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 18:23:22.920003 ignition[827]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 18:23:22.920003 ignition[827]: INFO : mount: mount passed Mar 17 18:23:22.920003 ignition[827]: INFO : Ignition finished successfully Mar 17 18:23:22.919000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.921000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:22.919179 systemd[1]: Finished sysroot-boot.service. Mar 17 18:23:22.920797 systemd[1]: Finished ignition-mount.service. Mar 17 18:23:23.564155 systemd[1]: Mounting sysroot-usr-share-oem.mount... Mar 17 18:23:23.570362 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (837) Mar 17 18:23:23.571648 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Mar 17 18:23:23.571664 kernel: BTRFS info (device vda6): using free space tree Mar 17 18:23:23.571674 kernel: BTRFS info (device vda6): has skinny extents Mar 17 18:23:23.575210 systemd[1]: Mounted sysroot-usr-share-oem.mount. Mar 17 18:23:23.576941 systemd[1]: Starting ignition-files.service... Mar 17 18:23:23.593120 ignition[857]: INFO : Ignition 2.14.0 Mar 17 18:23:23.593120 ignition[857]: INFO : Stage: files Mar 17 18:23:23.594698 ignition[857]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 18:23:23.594698 ignition[857]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 18:23:23.594698 ignition[857]: DEBUG : files: compiled without relabeling support, skipping Mar 17 18:23:23.601263 ignition[857]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Mar 17 18:23:23.601263 ignition[857]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Mar 17 18:23:23.604186 ignition[857]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Mar 17 18:23:23.604186 ignition[857]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Mar 17 18:23:23.606888 ignition[857]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Mar 17 18:23:23.606888 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Mar 17 18:23:23.606888 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Mar 17 18:23:23.605359 unknown[857]: wrote ssh authorized keys file for user: core Mar 17 18:23:23.680784 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Mar 17 18:23:23.870035 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Mar 17 18:23:23.872045 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 17 18:23:23.872045 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Mar 17 18:23:24.107509 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Mar 17 18:23:24.188187 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Mar 17 18:23:24.188187 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 18:23:24.191619 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Mar 17 18:23:24.461119 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Mar 17 18:23:24.576646 systemd-networkd[739]: eth0: Gained IPv6LL Mar 17 18:23:24.696290 ignition[857]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Mar 17 18:23:24.696290 ignition[857]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(e): op(f): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(11): [started] setting preset to disabled for "coreos-metadata.service" Mar 17 18:23:24.700375 ignition[857]: INFO : files: op(11): op(12): [started] removing enablement symlink(s) for "coreos-metadata.service" Mar 17 18:23:24.738465 ignition[857]: INFO : files: op(11): op(12): [finished] removing enablement symlink(s) for "coreos-metadata.service" Mar 17 18:23:24.740813 ignition[857]: INFO : files: op(11): [finished] setting preset to disabled for "coreos-metadata.service" Mar 17 18:23:24.740813 ignition[857]: INFO : files: createResultFile: createFiles: op(13): [started] writing file "/sysroot/etc/.ignition-result.json" Mar 17 18:23:24.740813 ignition[857]: INFO : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json" Mar 17 18:23:24.740813 ignition[857]: INFO : files: files passed Mar 17 18:23:24.740813 ignition[857]: INFO : Ignition finished successfully Mar 17 18:23:24.742000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.740872 systemd[1]: Finished ignition-files.service. Mar 17 18:23:24.744064 systemd[1]: Starting initrd-setup-root-after-ignition.service... Mar 17 18:23:24.750000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.750000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.751867 initrd-setup-root-after-ignition[882]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory Mar 17 18:23:24.745430 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Mar 17 18:23:24.754973 initrd-setup-root-after-ignition[884]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Mar 17 18:23:24.746132 systemd[1]: Starting ignition-quench.service... Mar 17 18:23:24.756000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.749908 systemd[1]: ignition-quench.service: Deactivated successfully. Mar 17 18:23:24.749994 systemd[1]: Finished ignition-quench.service. Mar 17 18:23:24.755130 systemd[1]: Finished initrd-setup-root-after-ignition.service. Mar 17 18:23:24.757369 systemd[1]: Reached target ignition-complete.target. Mar 17 18:23:24.759290 systemd[1]: Starting initrd-parse-etc.service... Mar 17 18:23:24.771452 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Mar 17 18:23:24.771547 systemd[1]: Finished initrd-parse-etc.service. Mar 17 18:23:24.772000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.772000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.773217 systemd[1]: Reached target initrd-fs.target. Mar 17 18:23:24.774414 systemd[1]: Reached target initrd.target. Mar 17 18:23:24.775687 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Mar 17 18:23:24.776458 systemd[1]: Starting dracut-pre-pivot.service... Mar 17 18:23:24.786569 systemd[1]: Finished dracut-pre-pivot.service. Mar 17 18:23:24.787000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.788114 systemd[1]: Starting initrd-cleanup.service... Mar 17 18:23:24.796304 systemd[1]: Stopped target nss-lookup.target. Mar 17 18:23:24.797180 systemd[1]: Stopped target remote-cryptsetup.target. Mar 17 18:23:24.798596 systemd[1]: Stopped target timers.target. Mar 17 18:23:24.799924 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Mar 17 18:23:24.800000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.800044 systemd[1]: Stopped dracut-pre-pivot.service. Mar 17 18:23:24.801297 systemd[1]: Stopped target initrd.target. Mar 17 18:23:24.802669 systemd[1]: Stopped target basic.target. Mar 17 18:23:24.803934 systemd[1]: Stopped target ignition-complete.target. Mar 17 18:23:24.805226 systemd[1]: Stopped target ignition-diskful.target. Mar 17 18:23:24.806537 systemd[1]: Stopped target initrd-root-device.target. Mar 17 18:23:24.808003 systemd[1]: Stopped target remote-fs.target. Mar 17 18:23:24.809342 systemd[1]: Stopped target remote-fs-pre.target. Mar 17 18:23:24.810764 systemd[1]: Stopped target sysinit.target. Mar 17 18:23:24.812059 systemd[1]: Stopped target local-fs.target. Mar 17 18:23:24.813364 systemd[1]: Stopped target local-fs-pre.target. Mar 17 18:23:24.814656 systemd[1]: Stopped target swap.target. Mar 17 18:23:24.816000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.815874 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Mar 17 18:23:24.815996 systemd[1]: Stopped dracut-pre-mount.service. Mar 17 18:23:24.819000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.817285 systemd[1]: Stopped target cryptsetup.target. Mar 17 18:23:24.820000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.818430 systemd[1]: dracut-initqueue.service: Deactivated successfully. Mar 17 18:23:24.818532 systemd[1]: Stopped dracut-initqueue.service. Mar 17 18:23:24.820030 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Mar 17 18:23:24.820129 systemd[1]: Stopped ignition-fetch-offline.service. Mar 17 18:23:24.821393 systemd[1]: Stopped target paths.target. Mar 17 18:23:24.822531 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Mar 17 18:23:24.826378 systemd[1]: Stopped systemd-ask-password-console.path. Mar 17 18:23:24.827314 systemd[1]: Stopped target slices.target. Mar 17 18:23:24.828871 systemd[1]: Stopped target sockets.target. Mar 17 18:23:24.831000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.830208 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Mar 17 18:23:24.832000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.830322 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Mar 17 18:23:24.835378 iscsid[745]: iscsid shutting down. Mar 17 18:23:24.831681 systemd[1]: ignition-files.service: Deactivated successfully. Mar 17 18:23:24.831772 systemd[1]: Stopped ignition-files.service. Mar 17 18:23:24.833837 systemd[1]: Stopping ignition-mount.service... Mar 17 18:23:24.838000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.835464 systemd[1]: Stopping iscsid.service... Mar 17 18:23:24.840000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.836761 systemd[1]: Stopping sysroot-boot.service... Mar 17 18:23:24.838063 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Mar 17 18:23:24.843106 ignition[897]: INFO : Ignition 2.14.0 Mar 17 18:23:24.843106 ignition[897]: INFO : Stage: umount Mar 17 18:23:24.843106 ignition[897]: INFO : no configs at "/usr/lib/ignition/base.d" Mar 17 18:23:24.843106 ignition[897]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Mar 17 18:23:24.843106 ignition[897]: INFO : umount: umount passed Mar 17 18:23:24.843106 ignition[897]: INFO : Ignition finished successfully Mar 17 18:23:24.843000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.844000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.847000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.849000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.850000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.838188 systemd[1]: Stopped systemd-udev-trigger.service. Mar 17 18:23:24.839540 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Mar 17 18:23:24.839626 systemd[1]: Stopped dracut-pre-trigger.service. Mar 17 18:23:24.854000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.842173 systemd[1]: iscsid.service: Deactivated successfully. Mar 17 18:23:24.855000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.855000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.842262 systemd[1]: Stopped iscsid.service. Mar 17 18:23:24.844196 systemd[1]: ignition-mount.service: Deactivated successfully. Mar 17 18:23:24.844273 systemd[1]: Stopped ignition-mount.service. Mar 17 18:23:24.845507 systemd[1]: iscsid.socket: Deactivated successfully. Mar 17 18:23:24.845571 systemd[1]: Closed iscsid.socket. Mar 17 18:23:24.846373 systemd[1]: ignition-disks.service: Deactivated successfully. Mar 17 18:23:24.846412 systemd[1]: Stopped ignition-disks.service. Mar 17 18:23:24.847921 systemd[1]: ignition-kargs.service: Deactivated successfully. Mar 17 18:23:24.847972 systemd[1]: Stopped ignition-kargs.service. Mar 17 18:23:24.866000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.849722 systemd[1]: ignition-setup.service: Deactivated successfully. Mar 17 18:23:24.849762 systemd[1]: Stopped ignition-setup.service. Mar 17 18:23:24.851301 systemd[1]: Stopping iscsiuio.service... Mar 17 18:23:24.870000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.853305 systemd[1]: sysroot-boot.mount: Deactivated successfully. Mar 17 18:23:24.873000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.853727 systemd[1]: iscsiuio.service: Deactivated successfully. Mar 17 18:23:24.875000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.853811 systemd[1]: Stopped iscsiuio.service. Mar 17 18:23:24.855253 systemd[1]: initrd-cleanup.service: Deactivated successfully. Mar 17 18:23:24.855338 systemd[1]: Finished initrd-cleanup.service. Mar 17 18:23:24.857190 systemd[1]: Stopped target network.target. Mar 17 18:23:24.881000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.858679 systemd[1]: iscsiuio.socket: Deactivated successfully. Mar 17 18:23:24.858714 systemd[1]: Closed iscsiuio.socket. Mar 17 18:23:24.860165 systemd[1]: Stopping systemd-networkd.service... Mar 17 18:23:24.885000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.861453 systemd[1]: Stopping systemd-resolved.service... Mar 17 18:23:24.887000 audit: BPF prog-id=6 op=UNLOAD Mar 17 18:23:24.863768 systemd-networkd[739]: eth0: DHCPv6 lease lost Mar 17 18:23:24.887000 audit: BPF prog-id=9 op=UNLOAD Mar 17 18:23:24.865271 systemd[1]: systemd-networkd.service: Deactivated successfully. Mar 17 18:23:24.888000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.865373 systemd[1]: Stopped systemd-networkd.service. Mar 17 18:23:24.867205 systemd[1]: systemd-networkd.socket: Deactivated successfully. Mar 17 18:23:24.893000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.867233 systemd[1]: Closed systemd-networkd.socket. Mar 17 18:23:24.894000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.868840 systemd[1]: Stopping network-cleanup.service... Mar 17 18:23:24.896000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.869705 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Mar 17 18:23:24.869758 systemd[1]: Stopped parse-ip-for-networkd.service. Mar 17 18:23:24.871287 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 17 18:23:24.899000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.871340 systemd[1]: Stopped systemd-sysctl.service. Mar 17 18:23:24.901000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.874596 systemd[1]: systemd-modules-load.service: Deactivated successfully. Mar 17 18:23:24.902000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.874636 systemd[1]: Stopped systemd-modules-load.service. Mar 17 18:23:24.878105 systemd[1]: Stopping systemd-udevd.service... Mar 17 18:23:24.905000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.880240 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Mar 17 18:23:24.906000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.906000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.880761 systemd[1]: systemd-resolved.service: Deactivated successfully. Mar 17 18:23:24.880851 systemd[1]: Stopped systemd-resolved.service. Mar 17 18:23:24.909000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:24.884818 systemd[1]: network-cleanup.service: Deactivated successfully. Mar 17 18:23:24.884908 systemd[1]: Stopped network-cleanup.service. Mar 17 18:23:24.888100 systemd[1]: systemd-udevd.service: Deactivated successfully. Mar 17 18:23:24.888209 systemd[1]: Stopped systemd-udevd.service. Mar 17 18:23:24.889518 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Mar 17 18:23:24.889554 systemd[1]: Closed systemd-udevd-control.socket. Mar 17 18:23:24.891002 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Mar 17 18:23:24.891033 systemd[1]: Closed systemd-udevd-kernel.socket. Mar 17 18:23:24.891841 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Mar 17 18:23:24.891883 systemd[1]: Stopped dracut-pre-udev.service. Mar 17 18:23:24.893455 systemd[1]: dracut-cmdline.service: Deactivated successfully. Mar 17 18:23:24.893496 systemd[1]: Stopped dracut-cmdline.service. Mar 17 18:23:24.895052 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Mar 17 18:23:24.895161 systemd[1]: Stopped dracut-cmdline-ask.service. Mar 17 18:23:24.897183 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Mar 17 18:23:24.898829 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Mar 17 18:23:24.898889 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Mar 17 18:23:24.901067 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Mar 17 18:23:24.901103 systemd[1]: Stopped kmod-static-nodes.service. Mar 17 18:23:24.901969 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Mar 17 18:23:24.902008 systemd[1]: Stopped systemd-vconsole-setup.service. Mar 17 18:23:24.903999 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Mar 17 18:23:24.904427 systemd[1]: sysroot-boot.service: Deactivated successfully. Mar 17 18:23:24.904510 systemd[1]: Stopped sysroot-boot.service. Mar 17 18:23:24.905820 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Mar 17 18:23:24.905894 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Mar 17 18:23:24.907202 systemd[1]: Reached target initrd-switch-root.target. Mar 17 18:23:24.936123 systemd-journald[291]: Received SIGTERM from PID 1 (n/a). Mar 17 18:23:24.908845 systemd[1]: initrd-setup-root.service: Deactivated successfully. Mar 17 18:23:24.908895 systemd[1]: Stopped initrd-setup-root.service. Mar 17 18:23:24.911003 systemd[1]: Starting initrd-switch-root.service... Mar 17 18:23:24.917260 systemd[1]: Switching root. Mar 17 18:23:24.939134 systemd-journald[291]: Journal stopped Mar 17 18:23:26.984372 kernel: SELinux: Class mctp_socket not defined in policy. Mar 17 18:23:26.984423 kernel: SELinux: Class anon_inode not defined in policy. Mar 17 18:23:26.984435 kernel: SELinux: the above unknown classes and permissions will be allowed Mar 17 18:23:26.984448 kernel: SELinux: policy capability network_peer_controls=1 Mar 17 18:23:26.984458 kernel: SELinux: policy capability open_perms=1 Mar 17 18:23:26.984468 kernel: SELinux: policy capability extended_socket_class=1 Mar 17 18:23:26.984478 kernel: SELinux: policy capability always_check_network=0 Mar 17 18:23:26.984488 kernel: SELinux: policy capability cgroup_seclabel=1 Mar 17 18:23:26.984500 kernel: SELinux: policy capability nnp_nosuid_transition=1 Mar 17 18:23:26.984510 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Mar 17 18:23:26.984522 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Mar 17 18:23:26.984533 systemd[1]: Successfully loaded SELinux policy in 34.100ms. Mar 17 18:23:26.984555 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 8.217ms. Mar 17 18:23:26.984567 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Mar 17 18:23:26.984578 systemd[1]: Detected virtualization kvm. Mar 17 18:23:26.984592 systemd[1]: Detected architecture arm64. Mar 17 18:23:26.984602 systemd[1]: Detected first boot. Mar 17 18:23:26.984613 systemd[1]: Initializing machine ID from VM UUID. Mar 17 18:23:26.984623 kernel: kauditd_printk_skb: 65 callbacks suppressed Mar 17 18:23:26.984633 kernel: audit: type=1400 audit(1742235805.109:76): avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Mar 17 18:23:26.984644 kernel: audit: type=1400 audit(1742235805.112:77): avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Mar 17 18:23:26.984655 kernel: audit: type=1334 audit(1742235805.113:78): prog-id=10 op=LOAD Mar 17 18:23:26.984665 kernel: audit: type=1334 audit(1742235805.113:79): prog-id=10 op=UNLOAD Mar 17 18:23:26.984675 kernel: audit: type=1334 audit(1742235805.114:80): prog-id=11 op=LOAD Mar 17 18:23:26.984684 kernel: audit: type=1334 audit(1742235805.114:81): prog-id=11 op=UNLOAD Mar 17 18:23:26.984694 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Mar 17 18:23:26.984705 kernel: audit: type=1400 audit(1742235805.159:82): avc: denied { associate } for pid=931 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Mar 17 18:23:26.984719 kernel: audit: type=1300 audit(1742235805.159:82): arch=c00000b7 syscall=5 success=yes exit=0 a0=40001c58b2 a1=40000c8de0 a2=40000cf0c0 a3=32 items=0 ppid=914 pid=931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 18:23:26.984731 kernel: audit: type=1327 audit(1742235805.159:82): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 18:23:26.984744 kernel: audit: type=1400 audit(1742235805.160:83): avc: denied { associate } for pid=931 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Mar 17 18:23:26.984756 systemd[1]: Populated /etc with preset unit settings. Mar 17 18:23:26.984768 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 18:23:26.984779 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 18:23:26.984791 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 18:23:26.984802 systemd[1]: initrd-switch-root.service: Deactivated successfully. Mar 17 18:23:26.984813 systemd[1]: Stopped initrd-switch-root.service. Mar 17 18:23:26.984824 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Mar 17 18:23:26.984835 systemd[1]: Created slice system-addon\x2dconfig.slice. Mar 17 18:23:26.984845 systemd[1]: Created slice system-addon\x2drun.slice. Mar 17 18:23:26.984856 systemd[1]: Created slice system-getty.slice. Mar 17 18:23:26.984867 systemd[1]: Created slice system-modprobe.slice. Mar 17 18:23:26.984878 systemd[1]: Created slice system-serial\x2dgetty.slice. Mar 17 18:23:26.984888 systemd[1]: Created slice system-system\x2dcloudinit.slice. Mar 17 18:23:26.984900 systemd[1]: Created slice system-systemd\x2dfsck.slice. Mar 17 18:23:26.984911 systemd[1]: Created slice user.slice. Mar 17 18:23:26.984929 systemd[1]: Started systemd-ask-password-console.path. Mar 17 18:23:26.984941 systemd[1]: Started systemd-ask-password-wall.path. Mar 17 18:23:26.984951 systemd[1]: Set up automount boot.automount. Mar 17 18:23:26.984962 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Mar 17 18:23:26.984973 systemd[1]: Stopped target initrd-switch-root.target. Mar 17 18:23:26.984983 systemd[1]: Stopped target initrd-fs.target. Mar 17 18:23:26.984993 systemd[1]: Stopped target initrd-root-fs.target. Mar 17 18:23:26.985012 systemd[1]: Reached target integritysetup.target. Mar 17 18:23:26.985023 systemd[1]: Reached target remote-cryptsetup.target. Mar 17 18:23:26.985034 systemd[1]: Reached target remote-fs.target. Mar 17 18:23:26.985044 systemd[1]: Reached target slices.target. Mar 17 18:23:26.985054 systemd[1]: Reached target swap.target. Mar 17 18:23:26.985065 systemd[1]: Reached target torcx.target. Mar 17 18:23:26.985076 systemd[1]: Reached target veritysetup.target. Mar 17 18:23:26.985087 systemd[1]: Listening on systemd-coredump.socket. Mar 17 18:23:26.985097 systemd[1]: Listening on systemd-initctl.socket. Mar 17 18:23:26.985108 systemd[1]: Listening on systemd-networkd.socket. Mar 17 18:23:26.985120 systemd[1]: Listening on systemd-udevd-control.socket. Mar 17 18:23:26.985131 systemd[1]: Listening on systemd-udevd-kernel.socket. Mar 17 18:23:26.985142 systemd[1]: Listening on systemd-userdbd.socket. Mar 17 18:23:26.985152 systemd[1]: Mounting dev-hugepages.mount... Mar 17 18:23:26.985163 systemd[1]: Mounting dev-mqueue.mount... Mar 17 18:23:26.985173 systemd[1]: Mounting media.mount... Mar 17 18:23:26.985183 systemd[1]: Mounting sys-kernel-debug.mount... Mar 17 18:23:26.985193 systemd[1]: Mounting sys-kernel-tracing.mount... Mar 17 18:23:26.985203 systemd[1]: Mounting tmp.mount... Mar 17 18:23:26.985215 systemd[1]: Starting flatcar-tmpfiles.service... Mar 17 18:23:26.985226 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 18:23:26.985237 systemd[1]: Starting kmod-static-nodes.service... Mar 17 18:23:26.985248 systemd[1]: Starting modprobe@configfs.service... Mar 17 18:23:26.985258 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 18:23:26.985268 systemd[1]: Starting modprobe@drm.service... Mar 17 18:23:26.985278 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 18:23:26.985288 systemd[1]: Starting modprobe@fuse.service... Mar 17 18:23:26.985298 systemd[1]: Starting modprobe@loop.service... Mar 17 18:23:26.985310 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Mar 17 18:23:26.985322 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Mar 17 18:23:26.985472 systemd[1]: Stopped systemd-fsck-root.service. Mar 17 18:23:26.985489 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Mar 17 18:23:26.985500 systemd[1]: Stopped systemd-fsck-usr.service. Mar 17 18:23:26.985510 systemd[1]: Stopped systemd-journald.service. Mar 17 18:23:26.985521 systemd[1]: Starting systemd-journald.service... Mar 17 18:23:26.985531 kernel: fuse: init (API version 7.34) Mar 17 18:23:26.985541 systemd[1]: Starting systemd-modules-load.service... Mar 17 18:23:26.985554 systemd[1]: Starting systemd-network-generator.service... Mar 17 18:23:26.985565 systemd[1]: Starting systemd-remount-fs.service... Mar 17 18:23:26.985575 systemd[1]: Starting systemd-udev-trigger.service... Mar 17 18:23:26.985585 systemd[1]: verity-setup.service: Deactivated successfully. Mar 17 18:23:26.985595 systemd[1]: Stopped verity-setup.service. Mar 17 18:23:26.985605 kernel: loop: module loaded Mar 17 18:23:26.985615 systemd[1]: Mounted dev-hugepages.mount. Mar 17 18:23:26.985625 systemd[1]: Mounted dev-mqueue.mount. Mar 17 18:23:26.985634 systemd[1]: Mounted media.mount. Mar 17 18:23:26.985645 systemd[1]: Mounted sys-kernel-debug.mount. Mar 17 18:23:26.985656 systemd[1]: Mounted sys-kernel-tracing.mount. Mar 17 18:23:26.985669 systemd[1]: Mounted tmp.mount. Mar 17 18:23:26.985679 systemd[1]: Finished kmod-static-nodes.service. Mar 17 18:23:26.985689 systemd[1]: modprobe@configfs.service: Deactivated successfully. Mar 17 18:23:26.985700 systemd[1]: Finished modprobe@configfs.service. Mar 17 18:23:26.985711 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 18:23:26.985721 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 18:23:26.985736 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 18:23:26.985749 systemd[1]: Finished modprobe@drm.service. Mar 17 18:23:26.985759 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 18:23:26.985770 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 18:23:26.985780 systemd[1]: modprobe@fuse.service: Deactivated successfully. Mar 17 18:23:26.985790 systemd[1]: Finished modprobe@fuse.service. Mar 17 18:23:26.985801 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 18:23:26.985813 systemd[1]: Finished modprobe@loop.service. Mar 17 18:23:26.985823 systemd[1]: Finished systemd-modules-load.service. Mar 17 18:23:26.985833 systemd[1]: Finished systemd-network-generator.service. Mar 17 18:23:26.985845 systemd-journald[993]: Journal started Mar 17 18:23:26.985886 systemd-journald[993]: Runtime Journal (/run/log/journal/a025cd63b78d4273a9d395da5abdbdbc) is 6.0M, max 48.7M, 42.6M free. Mar 17 18:23:24.999000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Mar 17 18:23:25.109000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Mar 17 18:23:25.112000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Mar 17 18:23:25.113000 audit: BPF prog-id=10 op=LOAD Mar 17 18:23:25.113000 audit: BPF prog-id=10 op=UNLOAD Mar 17 18:23:25.114000 audit: BPF prog-id=11 op=LOAD Mar 17 18:23:25.114000 audit: BPF prog-id=11 op=UNLOAD Mar 17 18:23:25.159000 audit[931]: AVC avc: denied { associate } for pid=931 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Mar 17 18:23:25.159000 audit[931]: SYSCALL arch=c00000b7 syscall=5 success=yes exit=0 a0=40001c58b2 a1=40000c8de0 a2=40000cf0c0 a3=32 items=0 ppid=914 pid=931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 18:23:25.159000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 18:23:25.160000 audit[931]: AVC avc: denied { associate } for pid=931 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Mar 17 18:23:25.160000 audit[931]: SYSCALL arch=c00000b7 syscall=34 success=yes exit=0 a0=ffffffffffffff9c a1=40001c5989 a2=1ed a3=0 items=2 ppid=914 pid=931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 18:23:25.160000 audit: CWD cwd="/" Mar 17 18:23:25.160000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 18:23:25.160000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Mar 17 18:23:25.160000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Mar 17 18:23:26.837000 audit: BPF prog-id=12 op=LOAD Mar 17 18:23:26.837000 audit: BPF prog-id=3 op=UNLOAD Mar 17 18:23:26.837000 audit: BPF prog-id=13 op=LOAD Mar 17 18:23:26.837000 audit: BPF prog-id=14 op=LOAD Mar 17 18:23:26.837000 audit: BPF prog-id=4 op=UNLOAD Mar 17 18:23:26.837000 audit: BPF prog-id=5 op=UNLOAD Mar 17 18:23:26.838000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.840000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.840000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.849000 audit: BPF prog-id=12 op=UNLOAD Mar 17 18:23:26.928000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.929000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.931000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.931000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.933000 audit: BPF prog-id=15 op=LOAD Mar 17 18:23:26.933000 audit: BPF prog-id=16 op=LOAD Mar 17 18:23:26.934000 audit: BPF prog-id=17 op=LOAD Mar 17 18:23:26.934000 audit: BPF prog-id=13 op=UNLOAD Mar 17 18:23:26.934000 audit: BPF prog-id=14 op=UNLOAD Mar 17 18:23:26.956000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.967000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.970000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.970000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.972000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.972000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.974000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.974000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.977000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.977000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.979000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.979000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.982000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.982000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.983000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Mar 17 18:23:26.983000 audit[993]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=6 a1=ffffd0183230 a2=4000 a3=1 items=0 ppid=1 pid=993 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 18:23:26.983000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Mar 17 18:23:26.983000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.985000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:25.158285 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 18:23:26.835801 systemd[1]: Queued start job for default target multi-user.target. Mar 17 18:23:25.158612 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Mar 17 18:23:26.835813 systemd[1]: Unnecessary job was removed for dev-vda6.device. Mar 17 18:23:25.158632 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Mar 17 18:23:26.838428 systemd[1]: systemd-journald.service: Deactivated successfully. Mar 17 18:23:25.158663 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Mar 17 18:23:25.158673 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="skipped missing lower profile" missing profile=oem Mar 17 18:23:25.158702 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Mar 17 18:23:25.158713 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Mar 17 18:23:25.158903 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Mar 17 18:23:25.158936 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Mar 17 18:23:25.158960 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Mar 17 18:23:25.159425 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Mar 17 18:23:25.159460 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Mar 17 18:23:25.159479 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.7: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.7 Mar 17 18:23:25.159493 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Mar 17 18:23:25.159510 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.7: no such file or directory" path=/var/lib/torcx/store/3510.3.7 Mar 17 18:23:25.159523 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:25Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Mar 17 18:23:26.597511 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:26Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 18:23:26.988681 systemd[1]: Finished systemd-remount-fs.service. Mar 17 18:23:26.597767 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:26Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 18:23:26.597858 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:26Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 18:23:26.988000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.990363 systemd[1]: Started systemd-journald.service. Mar 17 18:23:26.989000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:26.598039 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:26Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Mar 17 18:23:26.598087 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:26Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Mar 17 18:23:26.598142 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2025-03-17T18:23:26Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Mar 17 18:23:26.991015 systemd[1]: Reached target network-pre.target. Mar 17 18:23:26.992983 systemd[1]: Mounting sys-fs-fuse-connections.mount... Mar 17 18:23:26.994820 systemd[1]: Mounting sys-kernel-config.mount... Mar 17 18:23:26.995633 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Mar 17 18:23:26.997465 systemd[1]: Starting systemd-hwdb-update.service... Mar 17 18:23:26.999517 systemd[1]: Starting systemd-journal-flush.service... Mar 17 18:23:27.000528 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 18:23:27.001538 systemd[1]: Starting systemd-random-seed.service... Mar 17 18:23:27.002498 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 18:23:27.003624 systemd[1]: Starting systemd-sysctl.service... Mar 17 18:23:27.013048 systemd-journald[993]: Time spent on flushing to /var/log/journal/a025cd63b78d4273a9d395da5abdbdbc is 12.844ms for 992 entries. Mar 17 18:23:27.013048 systemd-journald[993]: System Journal (/var/log/journal/a025cd63b78d4273a9d395da5abdbdbc) is 8.0M, max 195.6M, 187.6M free. Mar 17 18:23:27.041268 systemd-journald[993]: Received client request to flush runtime journal. Mar 17 18:23:27.008000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.010000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.011000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.022000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.037000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.007160 systemd[1]: Finished flatcar-tmpfiles.service. Mar 17 18:23:27.008647 systemd[1]: Mounted sys-fs-fuse-connections.mount. Mar 17 18:23:27.046000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.009794 systemd[1]: Finished systemd-udev-trigger.service. Mar 17 18:23:27.048761 udevadm[1032]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Mar 17 18:23:27.010992 systemd[1]: Mounted sys-kernel-config.mount. Mar 17 18:23:27.011993 systemd[1]: Finished systemd-random-seed.service. Mar 17 18:23:27.013013 systemd[1]: Reached target first-boot-complete.target. Mar 17 18:23:27.015543 systemd[1]: Starting systemd-sysusers.service... Mar 17 18:23:27.017280 systemd[1]: Starting systemd-udev-settle.service... Mar 17 18:23:27.022045 systemd[1]: Finished systemd-sysctl.service. Mar 17 18:23:27.036615 systemd[1]: Finished systemd-sysusers.service. Mar 17 18:23:27.038508 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Mar 17 18:23:27.046150 systemd[1]: Finished systemd-journal-flush.service. Mar 17 18:23:27.055436 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Mar 17 18:23:27.056000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.360029 systemd[1]: Finished systemd-hwdb-update.service. Mar 17 18:23:27.360000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.361000 audit: BPF prog-id=18 op=LOAD Mar 17 18:23:27.361000 audit: BPF prog-id=19 op=LOAD Mar 17 18:23:27.361000 audit: BPF prog-id=7 op=UNLOAD Mar 17 18:23:27.361000 audit: BPF prog-id=8 op=UNLOAD Mar 17 18:23:27.362401 systemd[1]: Starting systemd-udevd.service... Mar 17 18:23:27.377596 systemd-udevd[1037]: Using default interface naming scheme 'v252'. Mar 17 18:23:27.389000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.390000 audit: BPF prog-id=20 op=LOAD Mar 17 18:23:27.394000 audit: BPF prog-id=21 op=LOAD Mar 17 18:23:27.394000 audit: BPF prog-id=22 op=LOAD Mar 17 18:23:27.394000 audit: BPF prog-id=23 op=LOAD Mar 17 18:23:27.389210 systemd[1]: Started systemd-udevd.service. Mar 17 18:23:27.391678 systemd[1]: Starting systemd-networkd.service... Mar 17 18:23:27.395407 systemd[1]: Starting systemd-userdbd.service... Mar 17 18:23:27.412143 systemd[1]: Condition check resulted in dev-ttyAMA0.device being skipped. Mar 17 18:23:27.432000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.431846 systemd[1]: Started systemd-userdbd.service. Mar 17 18:23:27.442950 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Mar 17 18:23:27.491289 systemd-networkd[1046]: lo: Link UP Mar 17 18:23:27.491299 systemd-networkd[1046]: lo: Gained carrier Mar 17 18:23:27.491641 systemd-networkd[1046]: Enumeration completed Mar 17 18:23:27.491723 systemd[1]: Started systemd-networkd.service. Mar 17 18:23:27.491741 systemd-networkd[1046]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Mar 17 18:23:27.492000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.494650 systemd-networkd[1046]: eth0: Link UP Mar 17 18:23:27.494661 systemd-networkd[1046]: eth0: Gained carrier Mar 17 18:23:27.505701 systemd[1]: Finished systemd-udev-settle.service. Mar 17 18:23:27.506000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.507641 systemd[1]: Starting lvm2-activation-early.service... Mar 17 18:23:27.519453 systemd-networkd[1046]: eth0: DHCPv4 address 10.0.0.89/16, gateway 10.0.0.1 acquired from 10.0.0.1 Mar 17 18:23:27.524813 lvm[1070]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 18:23:27.555120 systemd[1]: Finished lvm2-activation-early.service. Mar 17 18:23:27.555000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.556109 systemd[1]: Reached target cryptsetup.target. Mar 17 18:23:27.557891 systemd[1]: Starting lvm2-activation.service... Mar 17 18:23:27.561430 lvm[1071]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Mar 17 18:23:27.587051 systemd[1]: Finished lvm2-activation.service. Mar 17 18:23:27.587000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.587968 systemd[1]: Reached target local-fs-pre.target. Mar 17 18:23:27.588806 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Mar 17 18:23:27.588837 systemd[1]: Reached target local-fs.target. Mar 17 18:23:27.589600 systemd[1]: Reached target machines.target. Mar 17 18:23:27.591311 systemd[1]: Starting ldconfig.service... Mar 17 18:23:27.592277 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 18:23:27.592326 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:27.593257 systemd[1]: Starting systemd-boot-update.service... Mar 17 18:23:27.595398 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Mar 17 18:23:27.597444 systemd[1]: Starting systemd-machine-id-commit.service... Mar 17 18:23:27.600169 systemd[1]: Starting systemd-sysext.service... Mar 17 18:23:27.603934 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1073 (bootctl) Mar 17 18:23:27.605244 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Mar 17 18:23:27.606874 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Mar 17 18:23:27.608000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.614581 systemd[1]: Unmounting usr-share-oem.mount... Mar 17 18:23:27.619304 systemd[1]: usr-share-oem.mount: Deactivated successfully. Mar 17 18:23:27.619509 systemd[1]: Unmounted usr-share-oem.mount. Mar 17 18:23:27.669368 kernel: loop0: detected capacity change from 0 to 194096 Mar 17 18:23:27.677776 systemd[1]: Finished systemd-machine-id-commit.service. Mar 17 18:23:27.678000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.683502 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Mar 17 18:23:27.684575 systemd-fsck[1082]: fsck.fat 4.2 (2021-01-31) Mar 17 18:23:27.684575 systemd-fsck[1082]: /dev/vda1: 236 files, 117179/258078 clusters Mar 17 18:23:27.687429 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Mar 17 18:23:27.688000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.702402 kernel: loop1: detected capacity change from 0 to 194096 Mar 17 18:23:27.707886 (sd-sysext)[1087]: Using extensions 'kubernetes'. Mar 17 18:23:27.708218 (sd-sysext)[1087]: Merged extensions into '/usr'. Mar 17 18:23:27.723550 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 18:23:27.724717 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 18:23:27.726637 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 18:23:27.728647 systemd[1]: Starting modprobe@loop.service... Mar 17 18:23:27.729675 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 18:23:27.729807 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:27.730573 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 18:23:27.730690 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 18:23:27.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.731000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.732039 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 18:23:27.732144 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 18:23:27.733000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.733000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.733533 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 18:23:27.733631 systemd[1]: Finished modprobe@loop.service. Mar 17 18:23:27.734000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.734000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.735016 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 18:23:27.735111 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 18:23:27.778731 ldconfig[1072]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Mar 17 18:23:27.781888 systemd[1]: Finished ldconfig.service. Mar 17 18:23:27.782000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.958053 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Mar 17 18:23:27.959897 systemd[1]: Mounting boot.mount... Mar 17 18:23:27.961771 systemd[1]: Mounting usr-share-oem.mount... Mar 17 18:23:27.967714 systemd[1]: Mounted boot.mount. Mar 17 18:23:27.968612 systemd[1]: Mounted usr-share-oem.mount. Mar 17 18:23:27.970525 systemd[1]: Finished systemd-sysext.service. Mar 17 18:23:27.971000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.973290 systemd[1]: Starting ensure-sysext.service... Mar 17 18:23:27.975288 systemd[1]: Starting systemd-tmpfiles-setup.service... Mar 17 18:23:27.978086 systemd[1]: Finished systemd-boot-update.service. Mar 17 18:23:27.978000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:27.980471 systemd[1]: Reloading. Mar 17 18:23:27.984701 systemd-tmpfiles[1095]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Mar 17 18:23:27.985737 systemd-tmpfiles[1095]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Mar 17 18:23:27.987286 systemd-tmpfiles[1095]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Mar 17 18:23:28.007000 /usr/lib/systemd/system-generators/torcx-generator[1115]: time="2025-03-17T18:23:28Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 18:23:28.007026 /usr/lib/systemd/system-generators/torcx-generator[1115]: time="2025-03-17T18:23:28Z" level=info msg="torcx already run" Mar 17 18:23:28.067768 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 18:23:28.067785 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 18:23:28.083574 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 18:23:28.125000 audit: BPF prog-id=24 op=LOAD Mar 17 18:23:28.125000 audit: BPF prog-id=20 op=UNLOAD Mar 17 18:23:28.126000 audit: BPF prog-id=25 op=LOAD Mar 17 18:23:28.126000 audit: BPF prog-id=15 op=UNLOAD Mar 17 18:23:28.126000 audit: BPF prog-id=26 op=LOAD Mar 17 18:23:28.126000 audit: BPF prog-id=27 op=LOAD Mar 17 18:23:28.126000 audit: BPF prog-id=16 op=UNLOAD Mar 17 18:23:28.126000 audit: BPF prog-id=17 op=UNLOAD Mar 17 18:23:28.127000 audit: BPF prog-id=28 op=LOAD Mar 17 18:23:28.127000 audit: BPF prog-id=29 op=LOAD Mar 17 18:23:28.127000 audit: BPF prog-id=18 op=UNLOAD Mar 17 18:23:28.127000 audit: BPF prog-id=19 op=UNLOAD Mar 17 18:23:28.128000 audit: BPF prog-id=30 op=LOAD Mar 17 18:23:28.128000 audit: BPF prog-id=21 op=UNLOAD Mar 17 18:23:28.128000 audit: BPF prog-id=31 op=LOAD Mar 17 18:23:28.128000 audit: BPF prog-id=32 op=LOAD Mar 17 18:23:28.128000 audit: BPF prog-id=22 op=UNLOAD Mar 17 18:23:28.128000 audit: BPF prog-id=23 op=UNLOAD Mar 17 18:23:28.130549 systemd[1]: Finished systemd-tmpfiles-setup.service. Mar 17 18:23:28.131000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.135391 systemd[1]: Starting audit-rules.service... Mar 17 18:23:28.137276 systemd[1]: Starting clean-ca-certificates.service... Mar 17 18:23:28.139592 systemd[1]: Starting systemd-journal-catalog-update.service... Mar 17 18:23:28.143000 audit: BPF prog-id=33 op=LOAD Mar 17 18:23:28.144694 systemd[1]: Starting systemd-resolved.service... Mar 17 18:23:28.146000 audit: BPF prog-id=34 op=LOAD Mar 17 18:23:28.147536 systemd[1]: Starting systemd-timesyncd.service... Mar 17 18:23:28.149651 systemd[1]: Starting systemd-update-utmp.service... Mar 17 18:23:28.159247 systemd[1]: Finished clean-ca-certificates.service. Mar 17 18:23:28.160549 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.161738 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 18:23:28.160000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.161000 audit[1164]: SYSTEM_BOOT pid=1164 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.163809 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 18:23:28.165815 systemd[1]: Starting modprobe@loop.service... Mar 17 18:23:28.166786 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.166951 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:28.167074 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 17 18:23:28.169002 systemd[1]: Finished systemd-journal-catalog-update.service. Mar 17 18:23:28.169000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.170422 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 18:23:28.170535 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 18:23:28.171000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.171000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.171742 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 18:23:28.171854 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 18:23:28.172000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.172000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.173137 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 18:23:28.173248 systemd[1]: Finished modprobe@loop.service. Mar 17 18:23:28.174000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.174000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.176251 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 18:23:28.176416 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.177859 systemd[1]: Starting systemd-update-done.service... Mar 17 18:23:28.180696 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.182032 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 18:23:28.183981 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 18:23:28.185929 systemd[1]: Starting modprobe@loop.service... Mar 17 18:23:28.186727 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.186888 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:28.187026 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 17 18:23:28.187871 systemd[1]: Finished systemd-update-done.service. Mar 17 18:23:28.188000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.189310 systemd[1]: Finished systemd-update-utmp.service. Mar 17 18:23:28.189000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.190607 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 18:23:28.190713 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 18:23:28.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.191000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.191980 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 18:23:28.192086 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 18:23:28.192000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.192000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.193552 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 18:23:28.193674 systemd[1]: Finished modprobe@loop.service. Mar 17 18:23:28.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.194000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.195667 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 18:23:28.195756 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.197972 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.199125 systemd[1]: Starting modprobe@dm_mod.service... Mar 17 18:23:28.201066 systemd[1]: Starting modprobe@drm.service... Mar 17 18:23:28.203013 systemd[1]: Starting modprobe@efi_pstore.service... Mar 17 18:23:28.204911 systemd[1]: Starting modprobe@loop.service... Mar 17 18:23:28.205769 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.205917 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:28.207199 systemd[1]: Starting systemd-networkd-wait-online.service... Mar 17 18:23:28.208250 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Mar 17 18:23:28.209232 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Mar 17 18:23:28.209368 systemd[1]: Finished modprobe@dm_mod.service. Mar 17 18:23:28.210000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.210000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.211537 systemd[1]: modprobe@drm.service: Deactivated successfully. Mar 17 18:23:28.211653 systemd[1]: Finished modprobe@drm.service. Mar 17 18:23:28.211000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Mar 17 18:23:28.211000 audit[1180]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffc9d6ea20 a2=420 a3=0 items=0 ppid=1153 pid=1180 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Mar 17 18:23:28.211000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Mar 17 18:23:28.212000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.212000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Mar 17 18:23:28.212998 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Mar 17 18:23:28.213114 systemd[1]: Finished modprobe@efi_pstore.service. Mar 17 18:23:28.213573 augenrules[1180]: No rules Mar 17 18:23:28.214427 systemd[1]: modprobe@loop.service: Deactivated successfully. Mar 17 18:23:28.214547 systemd[1]: Finished modprobe@loop.service. Mar 17 18:23:28.216001 systemd[1]: Finished audit-rules.service. Mar 17 18:23:28.217240 systemd[1]: Finished ensure-sysext.service. Mar 17 18:23:28.219137 systemd[1]: Started systemd-timesyncd.service. Mar 17 18:23:28.219326 systemd-timesyncd[1163]: Contacted time server 10.0.0.1:123 (10.0.0.1). Mar 17 18:23:28.219408 systemd-timesyncd[1163]: Initial clock synchronization to Mon 2025-03-17 18:23:27.876816 UTC. Mar 17 18:23:28.220190 systemd-resolved[1160]: Positive Trust Anchors: Mar 17 18:23:28.220203 systemd-resolved[1160]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Mar 17 18:23:28.220229 systemd-resolved[1160]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Mar 17 18:23:28.220395 systemd[1]: Reached target time-set.target. Mar 17 18:23:28.221318 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Mar 17 18:23:28.221377 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.231750 systemd-resolved[1160]: Defaulting to hostname 'linux'. Mar 17 18:23:28.233139 systemd[1]: Started systemd-resolved.service. Mar 17 18:23:28.234035 systemd[1]: Reached target network.target. Mar 17 18:23:28.234817 systemd[1]: Reached target nss-lookup.target. Mar 17 18:23:28.235634 systemd[1]: Reached target sysinit.target. Mar 17 18:23:28.236469 systemd[1]: Started motdgen.path. Mar 17 18:23:28.237167 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Mar 17 18:23:28.238494 systemd[1]: Started logrotate.timer. Mar 17 18:23:28.239317 systemd[1]: Started mdadm.timer. Mar 17 18:23:28.239990 systemd[1]: Started systemd-tmpfiles-clean.timer. Mar 17 18:23:28.240879 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Mar 17 18:23:28.240920 systemd[1]: Reached target paths.target. Mar 17 18:23:28.241675 systemd[1]: Reached target timers.target. Mar 17 18:23:28.242710 systemd[1]: Listening on dbus.socket. Mar 17 18:23:28.244427 systemd[1]: Starting docker.socket... Mar 17 18:23:28.247446 systemd[1]: Listening on sshd.socket. Mar 17 18:23:28.248245 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:28.248665 systemd[1]: Listening on docker.socket. Mar 17 18:23:28.249480 systemd[1]: Reached target sockets.target. Mar 17 18:23:28.250209 systemd[1]: Reached target basic.target. Mar 17 18:23:28.251017 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.251048 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Mar 17 18:23:28.252108 systemd[1]: Starting containerd.service... Mar 17 18:23:28.253784 systemd[1]: Starting dbus.service... Mar 17 18:23:28.255404 systemd[1]: Starting enable-oem-cloudinit.service... Mar 17 18:23:28.257365 systemd[1]: Starting extend-filesystems.service... Mar 17 18:23:28.258232 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Mar 17 18:23:28.259626 systemd[1]: Starting motdgen.service... Mar 17 18:23:28.261536 systemd[1]: Starting prepare-helm.service... Mar 17 18:23:28.264073 systemd[1]: Starting ssh-key-proc-cmdline.service... Mar 17 18:23:28.267455 jq[1195]: false Mar 17 18:23:28.269143 systemd[1]: Starting sshd-keygen.service... Mar 17 18:23:28.272370 systemd[1]: Starting systemd-logind.service... Mar 17 18:23:28.273557 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Mar 17 18:23:28.273636 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Mar 17 18:23:28.276120 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Mar 17 18:23:28.276792 systemd[1]: Starting update-engine.service... Mar 17 18:23:28.278591 systemd[1]: Starting update-ssh-keys-after-ignition.service... Mar 17 18:23:28.281150 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Mar 17 18:23:28.282115 jq[1213]: true Mar 17 18:23:28.281342 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Mar 17 18:23:28.282417 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Mar 17 18:23:28.282590 systemd[1]: Finished ssh-key-proc-cmdline.service. Mar 17 18:23:28.288316 systemd[1]: motdgen.service: Deactivated successfully. Mar 17 18:23:28.288524 systemd[1]: Finished motdgen.service. Mar 17 18:23:28.291379 extend-filesystems[1196]: Found loop1 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda1 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda2 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda3 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found usr Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda4 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda6 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda7 Mar 17 18:23:28.291379 extend-filesystems[1196]: Found vda9 Mar 17 18:23:28.291379 extend-filesystems[1196]: Checking size of /dev/vda9 Mar 17 18:23:28.309312 tar[1216]: linux-arm64/helm Mar 17 18:23:28.296195 dbus-daemon[1194]: [system] SELinux support is enabled Mar 17 18:23:28.309671 jq[1218]: true Mar 17 18:23:28.296392 systemd[1]: Started dbus.service. Mar 17 18:23:28.298791 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Mar 17 18:23:28.298811 systemd[1]: Reached target system-config.target. Mar 17 18:23:28.299800 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Mar 17 18:23:28.299814 systemd[1]: Reached target user-config.target. Mar 17 18:23:28.319248 extend-filesystems[1196]: Resized partition /dev/vda9 Mar 17 18:23:28.336273 extend-filesystems[1235]: resize2fs 1.46.5 (30-Dec-2021) Mar 17 18:23:28.351368 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Mar 17 18:23:28.367718 systemd-logind[1210]: Watching system buttons on /dev/input/event0 (Power Button) Mar 17 18:23:28.368647 systemd-logind[1210]: New seat seat0. Mar 17 18:23:28.369355 update_engine[1211]: I0317 18:23:28.369123 1211 main.cc:92] Flatcar Update Engine starting Mar 17 18:23:28.372530 systemd[1]: Started update-engine.service. Mar 17 18:23:28.372679 update_engine[1211]: I0317 18:23:28.372659 1211 update_check_scheduler.cc:74] Next update check in 11m15s Mar 17 18:23:28.374631 systemd[1]: Started systemd-logind.service. Mar 17 18:23:28.376347 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Mar 17 18:23:28.377829 systemd[1]: Started locksmithd.service. Mar 17 18:23:28.391983 extend-filesystems[1235]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Mar 17 18:23:28.391983 extend-filesystems[1235]: old_desc_blocks = 1, new_desc_blocks = 1 Mar 17 18:23:28.391983 extend-filesystems[1235]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Mar 17 18:23:28.396454 extend-filesystems[1196]: Resized filesystem in /dev/vda9 Mar 17 18:23:28.392787 systemd[1]: extend-filesystems.service: Deactivated successfully. Mar 17 18:23:28.392968 systemd[1]: Finished extend-filesystems.service. Mar 17 18:23:28.400097 bash[1245]: Updated "/home/core/.ssh/authorized_keys" Mar 17 18:23:28.400803 systemd[1]: Finished update-ssh-keys-after-ignition.service. Mar 17 18:23:28.417662 env[1217]: time="2025-03-17T18:23:28.417593840Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Mar 17 18:23:28.434698 env[1217]: time="2025-03-17T18:23:28.434655720Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Mar 17 18:23:28.435019 env[1217]: time="2025-03-17T18:23:28.434997360Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Mar 17 18:23:28.436288 env[1217]: time="2025-03-17T18:23:28.436259560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.179-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Mar 17 18:23:28.436412 env[1217]: time="2025-03-17T18:23:28.436394920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Mar 17 18:23:28.436659 env[1217]: time="2025-03-17T18:23:28.436635120Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 18:23:28.436738 env[1217]: time="2025-03-17T18:23:28.436722400Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Mar 17 18:23:28.436796 env[1217]: time="2025-03-17T18:23:28.436781600Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Mar 17 18:23:28.436846 env[1217]: time="2025-03-17T18:23:28.436833560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Mar 17 18:23:28.436997 env[1217]: time="2025-03-17T18:23:28.436978320Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Mar 17 18:23:28.437368 env[1217]: time="2025-03-17T18:23:28.437320320Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Mar 17 18:23:28.437563 env[1217]: time="2025-03-17T18:23:28.437540920Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Mar 17 18:23:28.437629 env[1217]: time="2025-03-17T18:23:28.437614840Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Mar 17 18:23:28.437735 env[1217]: time="2025-03-17T18:23:28.437717800Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Mar 17 18:23:28.437811 env[1217]: time="2025-03-17T18:23:28.437796960Z" level=info msg="metadata content store policy set" policy=shared Mar 17 18:23:28.441278 env[1217]: time="2025-03-17T18:23:28.441255680Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Mar 17 18:23:28.441390 env[1217]: time="2025-03-17T18:23:28.441374360Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Mar 17 18:23:28.441450 env[1217]: time="2025-03-17T18:23:28.441436720Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Mar 17 18:23:28.441542 env[1217]: time="2025-03-17T18:23:28.441527040Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.441645 env[1217]: time="2025-03-17T18:23:28.441630760Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.441708 env[1217]: time="2025-03-17T18:23:28.441694400Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.441766 env[1217]: time="2025-03-17T18:23:28.441752480Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.442167 env[1217]: time="2025-03-17T18:23:28.442142360Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.442261 env[1217]: time="2025-03-17T18:23:28.442245360Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.442370 env[1217]: time="2025-03-17T18:23:28.442314680Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.442438 env[1217]: time="2025-03-17T18:23:28.442422440Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.442496 env[1217]: time="2025-03-17T18:23:28.442482880Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Mar 17 18:23:28.442649 env[1217]: time="2025-03-17T18:23:28.442629880Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Mar 17 18:23:28.442822 env[1217]: time="2025-03-17T18:23:28.442803480Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Mar 17 18:23:28.443097 env[1217]: time="2025-03-17T18:23:28.443077440Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Mar 17 18:23:28.443201 env[1217]: time="2025-03-17T18:23:28.443186320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443265 env[1217]: time="2025-03-17T18:23:28.443251520Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Mar 17 18:23:28.443489 env[1217]: time="2025-03-17T18:23:28.443473080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443600 env[1217]: time="2025-03-17T18:23:28.443584360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443664 env[1217]: time="2025-03-17T18:23:28.443650360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443733 env[1217]: time="2025-03-17T18:23:28.443718720Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443796 env[1217]: time="2025-03-17T18:23:28.443782520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443870 env[1217]: time="2025-03-17T18:23:28.443855480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.443942 env[1217]: time="2025-03-17T18:23:28.443927680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.444002 env[1217]: time="2025-03-17T18:23:28.443988640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.444068 env[1217]: time="2025-03-17T18:23:28.444054600Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Mar 17 18:23:28.444244 env[1217]: time="2025-03-17T18:23:28.444226360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.444324 env[1217]: time="2025-03-17T18:23:28.444309400Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.444406 env[1217]: time="2025-03-17T18:23:28.444391920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.444462 env[1217]: time="2025-03-17T18:23:28.444449080Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Mar 17 18:23:28.444545 env[1217]: time="2025-03-17T18:23:28.444528840Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Mar 17 18:23:28.444602 env[1217]: time="2025-03-17T18:23:28.444588680Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Mar 17 18:23:28.444673 env[1217]: time="2025-03-17T18:23:28.444658280Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Mar 17 18:23:28.444757 env[1217]: time="2025-03-17T18:23:28.444743520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Mar 17 18:23:28.445078 env[1217]: time="2025-03-17T18:23:28.445018080Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Mar 17 18:23:28.445761 env[1217]: time="2025-03-17T18:23:28.445520800Z" level=info msg="Connect containerd service" Mar 17 18:23:28.445761 env[1217]: time="2025-03-17T18:23:28.445560040Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Mar 17 18:23:28.446502 env[1217]: time="2025-03-17T18:23:28.446475000Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 18:23:28.446755 env[1217]: time="2025-03-17T18:23:28.446729400Z" level=info msg="Start subscribing containerd event" Mar 17 18:23:28.446954 env[1217]: time="2025-03-17T18:23:28.446932520Z" level=info msg="Start recovering state" Mar 17 18:23:28.447059 env[1217]: time="2025-03-17T18:23:28.447046680Z" level=info msg="Start event monitor" Mar 17 18:23:28.447134 env[1217]: time="2025-03-17T18:23:28.447120560Z" level=info msg="Start snapshots syncer" Mar 17 18:23:28.447191 env[1217]: time="2025-03-17T18:23:28.447177680Z" level=info msg="Start cni network conf syncer for default" Mar 17 18:23:28.447240 env[1217]: time="2025-03-17T18:23:28.447228760Z" level=info msg="Start streaming server" Mar 17 18:23:28.447638 env[1217]: time="2025-03-17T18:23:28.447596760Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Mar 17 18:23:28.447763 env[1217]: time="2025-03-17T18:23:28.447748640Z" level=info msg=serving... address=/run/containerd/containerd.sock Mar 17 18:23:28.447927 systemd[1]: Started containerd.service. Mar 17 18:23:28.450135 env[1217]: time="2025-03-17T18:23:28.449246560Z" level=info msg="containerd successfully booted in 0.032418s" Mar 17 18:23:28.463940 locksmithd[1246]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Mar 17 18:23:28.700021 tar[1216]: linux-arm64/LICENSE Mar 17 18:23:28.700141 tar[1216]: linux-arm64/README.md Mar 17 18:23:28.704528 systemd[1]: Finished prepare-helm.service. Mar 17 18:23:29.312582 systemd-networkd[1046]: eth0: Gained IPv6LL Mar 17 18:23:29.314641 systemd[1]: Finished systemd-networkd-wait-online.service. Mar 17 18:23:29.315836 systemd[1]: Reached target network-online.target. Mar 17 18:23:29.318086 systemd[1]: Starting kubelet.service... Mar 17 18:23:29.816468 systemd[1]: Started kubelet.service. Mar 17 18:23:30.357221 kubelet[1262]: E0317 18:23:30.357174 1262 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 18:23:30.359193 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 18:23:30.359338 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 18:23:31.650022 sshd_keygen[1215]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Mar 17 18:23:31.668760 systemd[1]: Finished sshd-keygen.service. Mar 17 18:23:31.671502 systemd[1]: Starting issuegen.service... Mar 17 18:23:31.677017 systemd[1]: issuegen.service: Deactivated successfully. Mar 17 18:23:31.677229 systemd[1]: Finished issuegen.service. Mar 17 18:23:31.680158 systemd[1]: Starting systemd-user-sessions.service... Mar 17 18:23:31.686175 systemd[1]: Finished systemd-user-sessions.service. Mar 17 18:23:31.688872 systemd[1]: Started getty@tty1.service. Mar 17 18:23:31.691153 systemd[1]: Started serial-getty@ttyAMA0.service. Mar 17 18:23:31.692383 systemd[1]: Reached target getty.target. Mar 17 18:23:31.693203 systemd[1]: Reached target multi-user.target. Mar 17 18:23:31.695618 systemd[1]: Starting systemd-update-utmp-runlevel.service... Mar 17 18:23:31.702367 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Mar 17 18:23:31.702542 systemd[1]: Finished systemd-update-utmp-runlevel.service. Mar 17 18:23:31.703641 systemd[1]: Startup finished in 569ms (kernel) + 4.389s (initrd) + 6.743s (userspace) = 11.702s. Mar 17 18:23:33.671262 systemd[1]: Created slice system-sshd.slice. Mar 17 18:23:33.672404 systemd[1]: Started sshd@0-10.0.0.89:22-10.0.0.1:46536.service. Mar 17 18:23:33.723185 sshd[1285]: Accepted publickey for core from 10.0.0.1 port 46536 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:23:33.725150 sshd[1285]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:23:33.732930 systemd[1]: Created slice user-500.slice. Mar 17 18:23:33.734063 systemd[1]: Starting user-runtime-dir@500.service... Mar 17 18:23:33.736101 systemd-logind[1210]: New session 1 of user core. Mar 17 18:23:33.742649 systemd[1]: Finished user-runtime-dir@500.service. Mar 17 18:23:33.744092 systemd[1]: Starting user@500.service... Mar 17 18:23:33.747520 (systemd)[1288]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:23:33.808341 systemd[1288]: Queued start job for default target default.target. Mar 17 18:23:33.808840 systemd[1288]: Reached target paths.target. Mar 17 18:23:33.808860 systemd[1288]: Reached target sockets.target. Mar 17 18:23:33.808871 systemd[1288]: Reached target timers.target. Mar 17 18:23:33.808881 systemd[1288]: Reached target basic.target. Mar 17 18:23:33.808934 systemd[1288]: Reached target default.target. Mar 17 18:23:33.808959 systemd[1288]: Startup finished in 55ms. Mar 17 18:23:33.809143 systemd[1]: Started user@500.service. Mar 17 18:23:33.810097 systemd[1]: Started session-1.scope. Mar 17 18:23:33.860922 systemd[1]: Started sshd@1-10.0.0.89:22-10.0.0.1:46548.service. Mar 17 18:23:33.903498 sshd[1297]: Accepted publickey for core from 10.0.0.1 port 46548 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:23:33.905032 sshd[1297]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:23:33.910267 systemd-logind[1210]: New session 2 of user core. Mar 17 18:23:33.910921 systemd[1]: Started session-2.scope. Mar 17 18:23:33.967782 sshd[1297]: pam_unix(sshd:session): session closed for user core Mar 17 18:23:33.970428 systemd[1]: sshd@1-10.0.0.89:22-10.0.0.1:46548.service: Deactivated successfully. Mar 17 18:23:33.970984 systemd[1]: session-2.scope: Deactivated successfully. Mar 17 18:23:33.971458 systemd-logind[1210]: Session 2 logged out. Waiting for processes to exit. Mar 17 18:23:33.972446 systemd[1]: Started sshd@2-10.0.0.89:22-10.0.0.1:46554.service. Mar 17 18:23:33.973111 systemd-logind[1210]: Removed session 2. Mar 17 18:23:34.015764 sshd[1303]: Accepted publickey for core from 10.0.0.1 port 46554 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:23:34.016975 sshd[1303]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:23:34.020346 systemd-logind[1210]: New session 3 of user core. Mar 17 18:23:34.021145 systemd[1]: Started session-3.scope. Mar 17 18:23:34.070248 sshd[1303]: pam_unix(sshd:session): session closed for user core Mar 17 18:23:34.074098 systemd[1]: sshd@2-10.0.0.89:22-10.0.0.1:46554.service: Deactivated successfully. Mar 17 18:23:34.074689 systemd[1]: session-3.scope: Deactivated successfully. Mar 17 18:23:34.075200 systemd-logind[1210]: Session 3 logged out. Waiting for processes to exit. Mar 17 18:23:34.076241 systemd[1]: Started sshd@3-10.0.0.89:22-10.0.0.1:46570.service. Mar 17 18:23:34.076926 systemd-logind[1210]: Removed session 3. Mar 17 18:23:34.118932 sshd[1309]: Accepted publickey for core from 10.0.0.1 port 46570 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:23:34.120126 sshd[1309]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:23:34.123570 systemd-logind[1210]: New session 4 of user core. Mar 17 18:23:34.124381 systemd[1]: Started session-4.scope. Mar 17 18:23:34.177136 sshd[1309]: pam_unix(sshd:session): session closed for user core Mar 17 18:23:34.180620 systemd[1]: Started sshd@4-10.0.0.89:22-10.0.0.1:46576.service. Mar 17 18:23:34.181104 systemd[1]: sshd@3-10.0.0.89:22-10.0.0.1:46570.service: Deactivated successfully. Mar 17 18:23:34.181672 systemd[1]: session-4.scope: Deactivated successfully. Mar 17 18:23:34.182189 systemd-logind[1210]: Session 4 logged out. Waiting for processes to exit. Mar 17 18:23:34.183141 systemd-logind[1210]: Removed session 4. Mar 17 18:23:34.222763 sshd[1314]: Accepted publickey for core from 10.0.0.1 port 46576 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:23:34.223831 sshd[1314]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:23:34.227813 systemd[1]: Started session-5.scope. Mar 17 18:23:34.228270 systemd-logind[1210]: New session 5 of user core. Mar 17 18:23:34.289307 sudo[1318]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Mar 17 18:23:34.289543 sudo[1318]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Mar 17 18:23:34.340215 systemd[1]: Starting docker.service... Mar 17 18:23:34.417995 env[1330]: time="2025-03-17T18:23:34.417942296Z" level=info msg="Starting up" Mar 17 18:23:34.419702 env[1330]: time="2025-03-17T18:23:34.419676846Z" level=info msg="parsed scheme: \"unix\"" module=grpc Mar 17 18:23:34.419784 env[1330]: time="2025-03-17T18:23:34.419770856Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Mar 17 18:23:34.419853 env[1330]: time="2025-03-17T18:23:34.419839006Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Mar 17 18:23:34.419905 env[1330]: time="2025-03-17T18:23:34.419893151Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Mar 17 18:23:34.425620 env[1330]: time="2025-03-17T18:23:34.425577999Z" level=info msg="parsed scheme: \"unix\"" module=grpc Mar 17 18:23:34.425620 env[1330]: time="2025-03-17T18:23:34.425601002Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Mar 17 18:23:34.425620 env[1330]: time="2025-03-17T18:23:34.425614812Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Mar 17 18:23:34.425620 env[1330]: time="2025-03-17T18:23:34.425623537Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Mar 17 18:23:34.579013 env[1330]: time="2025-03-17T18:23:34.578179100Z" level=info msg="Loading containers: start." Mar 17 18:23:34.700597 kernel: Initializing XFRM netlink socket Mar 17 18:23:34.723493 env[1330]: time="2025-03-17T18:23:34.723460995Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Mar 17 18:23:34.784553 systemd-networkd[1046]: docker0: Link UP Mar 17 18:23:34.799732 env[1330]: time="2025-03-17T18:23:34.799684221Z" level=info msg="Loading containers: done." Mar 17 18:23:34.818338 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck3911067555-merged.mount: Deactivated successfully. Mar 17 18:23:34.819558 env[1330]: time="2025-03-17T18:23:34.819514035Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Mar 17 18:23:34.819729 env[1330]: time="2025-03-17T18:23:34.819710858Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Mar 17 18:23:34.819823 env[1330]: time="2025-03-17T18:23:34.819808310Z" level=info msg="Daemon has completed initialization" Mar 17 18:23:34.832696 systemd[1]: Started docker.service. Mar 17 18:23:34.839572 env[1330]: time="2025-03-17T18:23:34.839526001Z" level=info msg="API listen on /run/docker.sock" Mar 17 18:23:35.533815 env[1217]: time="2025-03-17T18:23:35.533765752Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.11\"" Mar 17 18:23:36.141264 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3959163729.mount: Deactivated successfully. Mar 17 18:23:37.505842 env[1217]: time="2025-03-17T18:23:37.505799051Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:37.507638 env[1217]: time="2025-03-17T18:23:37.507612346Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fcbef283ab16167d1ca4acb66836af518e9fe445111fbc618fdbe196858f9530,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:37.508847 env[1217]: time="2025-03-17T18:23:37.508826387Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:37.510651 env[1217]: time="2025-03-17T18:23:37.510620331Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:77c54346965036acc7ac95c3200597ede36db9246179248dde21c1a3ecc1caf0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:37.511592 env[1217]: time="2025-03-17T18:23:37.511563258Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.11\" returns image reference \"sha256:fcbef283ab16167d1ca4acb66836af518e9fe445111fbc618fdbe196858f9530\"" Mar 17 18:23:37.520214 env[1217]: time="2025-03-17T18:23:37.520180601Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.11\"" Mar 17 18:23:39.191007 env[1217]: time="2025-03-17T18:23:39.190958847Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:39.192586 env[1217]: time="2025-03-17T18:23:39.192553409Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:9469d949b9e8c03b6cb06af513f683dd2975b57092f3deb2a9e125e0d05188d3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:39.194614 env[1217]: time="2025-03-17T18:23:39.194584481Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:39.196189 env[1217]: time="2025-03-17T18:23:39.196155313Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:d8874f3fb45591ecdac67a3035c730808f18b3ab13147495c7d77eb1960d4f6f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:39.197090 env[1217]: time="2025-03-17T18:23:39.197057164Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.11\" returns image reference \"sha256:9469d949b9e8c03b6cb06af513f683dd2975b57092f3deb2a9e125e0d05188d3\"" Mar 17 18:23:39.206063 env[1217]: time="2025-03-17T18:23:39.206035603Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.11\"" Mar 17 18:23:40.334284 env[1217]: time="2025-03-17T18:23:40.334227513Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:40.335411 env[1217]: time="2025-03-17T18:23:40.335376719Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3540cd10f52fac0a58ba43c004c6d3941e2a9f53e06440b982b9c130a72c0213,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:40.337654 env[1217]: time="2025-03-17T18:23:40.337624989Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:40.341429 env[1217]: time="2025-03-17T18:23:40.341395988Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:c699f8c97ae7ec819c8bd878d3db104ba72fc440d810d9030e09286b696017b5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:40.342284 env[1217]: time="2025-03-17T18:23:40.342233891Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.11\" returns image reference \"sha256:3540cd10f52fac0a58ba43c004c6d3941e2a9f53e06440b982b9c130a72c0213\"" Mar 17 18:23:40.353243 env[1217]: time="2025-03-17T18:23:40.353206952Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.11\"" Mar 17 18:23:40.521954 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Mar 17 18:23:40.522163 systemd[1]: Stopped kubelet.service. Mar 17 18:23:40.523664 systemd[1]: Starting kubelet.service... Mar 17 18:23:40.608480 systemd[1]: Started kubelet.service. Mar 17 18:23:40.713449 kubelet[1492]: E0317 18:23:40.713406 1492 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Mar 17 18:23:40.716071 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Mar 17 18:23:40.716262 systemd[1]: kubelet.service: Failed with result 'exit-code'. Mar 17 18:23:41.427301 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2163789089.mount: Deactivated successfully. Mar 17 18:23:41.849134 env[1217]: time="2025-03-17T18:23:41.849053123Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:41.850472 env[1217]: time="2025-03-17T18:23:41.850438675Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fe83790bf8a35411788b67fe5f0ce35309056c40530484d516af2ca01375220c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:41.851548 env[1217]: time="2025-03-17T18:23:41.851512062Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.30.11,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:41.853249 env[1217]: time="2025-03-17T18:23:41.853208551Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:ea4da798040a18ed3f302e8d5f67307c7275a2a53bcf3d51bcec223acda84a55,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:41.853652 env[1217]: time="2025-03-17T18:23:41.853620211Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.11\" returns image reference \"sha256:fe83790bf8a35411788b67fe5f0ce35309056c40530484d516af2ca01375220c\"" Mar 17 18:23:41.862652 env[1217]: time="2025-03-17T18:23:41.862618766Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Mar 17 18:23:42.400870 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2846366529.mount: Deactivated successfully. Mar 17 18:23:43.297770 env[1217]: time="2025-03-17T18:23:43.297606525Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.299501 env[1217]: time="2025-03-17T18:23:43.299456356Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.301281 env[1217]: time="2025-03-17T18:23:43.301253933Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.302883 env[1217]: time="2025-03-17T18:23:43.302848698Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.304641 env[1217]: time="2025-03-17T18:23:43.304600102Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Mar 17 18:23:43.314480 env[1217]: time="2025-03-17T18:23:43.314442074Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Mar 17 18:23:43.844153 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount438740319.mount: Deactivated successfully. Mar 17 18:23:43.863738 env[1217]: time="2025-03-17T18:23:43.863692744Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.866011 env[1217]: time="2025-03-17T18:23:43.865975498Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.867376 env[1217]: time="2025-03-17T18:23:43.867346708Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.868786 env[1217]: time="2025-03-17T18:23:43.868753283Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:43.869421 env[1217]: time="2025-03-17T18:23:43.869382980Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Mar 17 18:23:43.878127 env[1217]: time="2025-03-17T18:23:43.878089257Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Mar 17 18:23:44.386067 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2621020587.mount: Deactivated successfully. Mar 17 18:23:46.554574 env[1217]: time="2025-03-17T18:23:46.554516419Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:46.555868 env[1217]: time="2025-03-17T18:23:46.555835481Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:46.557708 env[1217]: time="2025-03-17T18:23:46.557677899Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.12-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:46.559696 env[1217]: time="2025-03-17T18:23:46.559670927Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:46.560626 env[1217]: time="2025-03-17T18:23:46.560600320Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" Mar 17 18:23:50.713199 systemd[1]: Stopped kubelet.service. Mar 17 18:23:50.715135 systemd[1]: Starting kubelet.service... Mar 17 18:23:50.728878 systemd[1]: Reloading. Mar 17 18:23:50.764042 /usr/lib/systemd/system-generators/torcx-generator[1623]: time="2025-03-17T18:23:50Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 18:23:50.764072 /usr/lib/systemd/system-generators/torcx-generator[1623]: time="2025-03-17T18:23:50Z" level=info msg="torcx already run" Mar 17 18:23:50.848853 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 18:23:50.849019 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 18:23:50.865507 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 18:23:50.929776 systemd[1]: Started kubelet.service. Mar 17 18:23:50.931157 systemd[1]: Stopping kubelet.service... Mar 17 18:23:50.931426 systemd[1]: kubelet.service: Deactivated successfully. Mar 17 18:23:50.931612 systemd[1]: Stopped kubelet.service. Mar 17 18:23:50.933005 systemd[1]: Starting kubelet.service... Mar 17 18:23:51.012199 systemd[1]: Started kubelet.service. Mar 17 18:23:51.049581 kubelet[1666]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:23:51.049581 kubelet[1666]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 18:23:51.049581 kubelet[1666]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:23:51.050506 kubelet[1666]: I0317 18:23:51.050458 1666 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 18:23:51.544316 kubelet[1666]: I0317 18:23:51.544260 1666 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Mar 17 18:23:51.544316 kubelet[1666]: I0317 18:23:51.544292 1666 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 18:23:51.544554 kubelet[1666]: I0317 18:23:51.544524 1666 server.go:927] "Client rotation is on, will bootstrap in background" Mar 17 18:23:51.564824 kubelet[1666]: E0317 18:23:51.564797 1666 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.89:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.565225 kubelet[1666]: I0317 18:23:51.564965 1666 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 18:23:51.783305 kubelet[1666]: I0317 18:23:51.783273 1666 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 18:23:51.784581 kubelet[1666]: I0317 18:23:51.784539 1666 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 18:23:51.784746 kubelet[1666]: I0317 18:23:51.784581 1666 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Mar 17 18:23:51.784828 kubelet[1666]: I0317 18:23:51.784809 1666 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 18:23:51.784828 kubelet[1666]: I0317 18:23:51.784819 1666 container_manager_linux.go:301] "Creating device plugin manager" Mar 17 18:23:51.785302 kubelet[1666]: I0317 18:23:51.785275 1666 state_mem.go:36] "Initialized new in-memory state store" Mar 17 18:23:51.787963 kubelet[1666]: I0317 18:23:51.787947 1666 kubelet.go:400] "Attempting to sync node with API server" Mar 17 18:23:51.788001 kubelet[1666]: I0317 18:23:51.787968 1666 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 18:23:51.788362 kubelet[1666]: I0317 18:23:51.788249 1666 kubelet.go:312] "Adding apiserver pod source" Mar 17 18:23:51.788497 kubelet[1666]: I0317 18:23:51.788439 1666 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 18:23:51.789986 kubelet[1666]: W0317 18:23:51.789937 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.89:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.790038 kubelet[1666]: I0317 18:23:51.789994 1666 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Mar 17 18:23:51.790038 kubelet[1666]: E0317 18:23:51.790014 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.89:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.790038 kubelet[1666]: W0317 18:23:51.789942 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.89:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.790108 kubelet[1666]: E0317 18:23:51.790050 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.89:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.790397 kubelet[1666]: I0317 18:23:51.790377 1666 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 18:23:51.790504 kubelet[1666]: W0317 18:23:51.790483 1666 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Mar 17 18:23:51.791343 kubelet[1666]: I0317 18:23:51.791318 1666 server.go:1264] "Started kubelet" Mar 17 18:23:51.794343 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Mar 17 18:23:51.794541 kubelet[1666]: I0317 18:23:51.794524 1666 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 18:23:51.800933 kubelet[1666]: I0317 18:23:51.800899 1666 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 18:23:51.801011 kubelet[1666]: I0317 18:23:51.800915 1666 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 18:23:51.801183 kubelet[1666]: I0317 18:23:51.801158 1666 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 18:23:51.801274 kubelet[1666]: E0317 18:23:51.801256 1666 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found" Mar 17 18:23:51.801379 kubelet[1666]: I0317 18:23:51.801364 1666 volume_manager.go:291] "Starting Kubelet Volume Manager" Mar 17 18:23:51.801484 kubelet[1666]: I0317 18:23:51.801467 1666 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 17 18:23:51.802050 kubelet[1666]: I0317 18:23:51.802018 1666 server.go:455] "Adding debug handlers to kubelet server" Mar 17 18:23:51.802514 kubelet[1666]: E0317 18:23:51.802273 1666 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.89:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.89:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.182daa421ce0a809 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-03-17 18:23:51.791298569 +0000 UTC m=+0.775773551,LastTimestamp:2025-03-17 18:23:51.791298569 +0000 UTC m=+0.775773551,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}" Mar 17 18:23:51.802839 kubelet[1666]: I0317 18:23:51.802810 1666 reconciler.go:26] "Reconciler: start to sync state" Mar 17 18:23:51.803571 kubelet[1666]: W0317 18:23:51.803522 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.89:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.803571 kubelet[1666]: E0317 18:23:51.803570 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.89:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.803665 kubelet[1666]: E0317 18:23:51.803634 1666 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.89:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.89:6443: connect: connection refused" interval="200ms" Mar 17 18:23:51.803729 kubelet[1666]: I0317 18:23:51.803701 1666 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 18:23:51.803851 kubelet[1666]: E0317 18:23:51.803821 1666 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 18:23:51.804985 kubelet[1666]: I0317 18:23:51.804955 1666 factory.go:221] Registration of the containerd container factory successfully Mar 17 18:23:51.804985 kubelet[1666]: I0317 18:23:51.804973 1666 factory.go:221] Registration of the systemd container factory successfully Mar 17 18:23:51.811509 kubelet[1666]: I0317 18:23:51.811479 1666 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 18:23:51.812809 kubelet[1666]: I0317 18:23:51.812783 1666 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 18:23:51.812809 kubelet[1666]: I0317 18:23:51.812807 1666 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 18:23:51.812949 kubelet[1666]: I0317 18:23:51.812825 1666 kubelet.go:2337] "Starting kubelet main sync loop" Mar 17 18:23:51.812949 kubelet[1666]: E0317 18:23:51.812875 1666 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 18:23:51.813817 kubelet[1666]: W0317 18:23:51.813789 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.89:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.813932 kubelet[1666]: E0317 18:23:51.813919 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.89:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:51.816648 kubelet[1666]: I0317 18:23:51.816610 1666 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 18:23:51.816648 kubelet[1666]: I0317 18:23:51.816631 1666 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 18:23:51.816648 kubelet[1666]: I0317 18:23:51.816646 1666 state_mem.go:36] "Initialized new in-memory state store" Mar 17 18:23:51.818976 kubelet[1666]: I0317 18:23:51.818950 1666 policy_none.go:49] "None policy: Start" Mar 17 18:23:51.819457 kubelet[1666]: I0317 18:23:51.819443 1666 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 18:23:51.819503 kubelet[1666]: I0317 18:23:51.819486 1666 state_mem.go:35] "Initializing new in-memory state store" Mar 17 18:23:51.823934 systemd[1]: Created slice kubepods.slice. Mar 17 18:23:51.827811 systemd[1]: Created slice kubepods-burstable.slice. Mar 17 18:23:51.830074 systemd[1]: Created slice kubepods-besteffort.slice. Mar 17 18:23:51.845011 kubelet[1666]: I0317 18:23:51.844980 1666 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 18:23:51.845237 kubelet[1666]: I0317 18:23:51.845120 1666 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 18:23:51.845292 kubelet[1666]: I0317 18:23:51.845243 1666 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 18:23:51.846575 kubelet[1666]: E0317 18:23:51.846549 1666 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Mar 17 18:23:51.902741 kubelet[1666]: I0317 18:23:51.902714 1666 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 18:23:51.903036 kubelet[1666]: E0317 18:23:51.903013 1666 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.89:6443/api/v1/nodes\": dial tcp 10.0.0.89:6443: connect: connection refused" node="localhost" Mar 17 18:23:51.913325 kubelet[1666]: I0317 18:23:51.913264 1666 topology_manager.go:215] "Topology Admit Handler" podUID="d79ab404294384d4bcc36fb5b5509bbb" podNamespace="kube-system" podName="kube-scheduler-localhost" Mar 17 18:23:51.914472 kubelet[1666]: I0317 18:23:51.914436 1666 topology_manager.go:215] "Topology Admit Handler" podUID="1518fccdb57a759873d8ac5bfb2e1095" podNamespace="kube-system" podName="kube-apiserver-localhost" Mar 17 18:23:51.915197 kubelet[1666]: I0317 18:23:51.915159 1666 topology_manager.go:215] "Topology Admit Handler" podUID="23a18e2dc14f395c5f1bea711a5a9344" podNamespace="kube-system" podName="kube-controller-manager-localhost" Mar 17 18:23:51.919780 systemd[1]: Created slice kubepods-burstable-podd79ab404294384d4bcc36fb5b5509bbb.slice. Mar 17 18:23:51.936202 systemd[1]: Created slice kubepods-burstable-pod1518fccdb57a759873d8ac5bfb2e1095.slice. Mar 17 18:23:51.948231 systemd[1]: Created slice kubepods-burstable-pod23a18e2dc14f395c5f1bea711a5a9344.slice. Mar 17 18:23:52.003489 kubelet[1666]: I0317 18:23:52.003448 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:52.003489 kubelet[1666]: I0317 18:23:52.003484 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:52.003623 kubelet[1666]: I0317 18:23:52.003508 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1518fccdb57a759873d8ac5bfb2e1095-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"1518fccdb57a759873d8ac5bfb2e1095\") " pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:52.003623 kubelet[1666]: I0317 18:23:52.003525 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1518fccdb57a759873d8ac5bfb2e1095-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"1518fccdb57a759873d8ac5bfb2e1095\") " pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:52.003623 kubelet[1666]: I0317 18:23:52.003544 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:52.003623 kubelet[1666]: I0317 18:23:52.003560 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:52.003623 kubelet[1666]: I0317 18:23:52.003578 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:52.003732 kubelet[1666]: I0317 18:23:52.003593 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d79ab404294384d4bcc36fb5b5509bbb-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d79ab404294384d4bcc36fb5b5509bbb\") " pod="kube-system/kube-scheduler-localhost" Mar 17 18:23:52.003732 kubelet[1666]: I0317 18:23:52.003605 1666 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1518fccdb57a759873d8ac5bfb2e1095-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"1518fccdb57a759873d8ac5bfb2e1095\") " pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:52.004035 kubelet[1666]: E0317 18:23:52.003985 1666 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.89:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.89:6443: connect: connection refused" interval="400ms" Mar 17 18:23:52.105104 kubelet[1666]: I0317 18:23:52.105007 1666 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 18:23:52.106435 kubelet[1666]: E0317 18:23:52.106405 1666 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.89:6443/api/v1/nodes\": dial tcp 10.0.0.89:6443: connect: connection refused" node="localhost" Mar 17 18:23:52.234133 kubelet[1666]: E0317 18:23:52.234092 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:52.234944 env[1217]: time="2025-03-17T18:23:52.234899702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d79ab404294384d4bcc36fb5b5509bbb,Namespace:kube-system,Attempt:0,}" Mar 17 18:23:52.247571 kubelet[1666]: E0317 18:23:52.247547 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:52.247950 env[1217]: time="2025-03-17T18:23:52.247913946Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:1518fccdb57a759873d8ac5bfb2e1095,Namespace:kube-system,Attempt:0,}" Mar 17 18:23:52.250463 kubelet[1666]: E0317 18:23:52.250416 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:52.250819 env[1217]: time="2025-03-17T18:23:52.250761684Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:23a18e2dc14f395c5f1bea711a5a9344,Namespace:kube-system,Attempt:0,}" Mar 17 18:23:52.404804 kubelet[1666]: E0317 18:23:52.404695 1666 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.89:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.89:6443: connect: connection refused" interval="800ms" Mar 17 18:23:52.508429 kubelet[1666]: I0317 18:23:52.508399 1666 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 18:23:52.508876 kubelet[1666]: E0317 18:23:52.508848 1666 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.89:6443/api/v1/nodes\": dial tcp 10.0.0.89:6443: connect: connection refused" node="localhost" Mar 17 18:23:52.605076 kubelet[1666]: W0317 18:23:52.604998 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.89:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:52.605076 kubelet[1666]: E0317 18:23:52.605074 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.89:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:52.726991 kubelet[1666]: W0317 18:23:52.726882 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.89:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:52.727156 kubelet[1666]: E0317 18:23:52.727141 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.89:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:52.798787 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3506403161.mount: Deactivated successfully. Mar 17 18:23:52.802358 env[1217]: time="2025-03-17T18:23:52.802302837Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.803745 env[1217]: time="2025-03-17T18:23:52.803668801Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.805850 env[1217]: time="2025-03-17T18:23:52.805820443Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.806551 env[1217]: time="2025-03-17T18:23:52.806525521Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.808348 env[1217]: time="2025-03-17T18:23:52.808311929Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.810853 env[1217]: time="2025-03-17T18:23:52.810822418Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.814310 env[1217]: time="2025-03-17T18:23:52.813998663Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.815120 env[1217]: time="2025-03-17T18:23:52.814963664Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.816245 kubelet[1666]: W0317 18:23:52.816160 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.89:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:52.816245 kubelet[1666]: E0317 18:23:52.816222 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.89:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:52.817910 env[1217]: time="2025-03-17T18:23:52.816956102Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.817910 env[1217]: time="2025-03-17T18:23:52.817650043Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.819495 env[1217]: time="2025-03-17T18:23:52.819466112Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.820391 env[1217]: time="2025-03-17T18:23:52.820359056Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:23:52.844450 env[1217]: time="2025-03-17T18:23:52.844390595Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:23:52.844649 env[1217]: time="2025-03-17T18:23:52.844430716Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:23:52.844649 env[1217]: time="2025-03-17T18:23:52.844441055Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:23:52.844765 env[1217]: time="2025-03-17T18:23:52.844651557Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/03b696f644db037f376562f848d75eab6d389a8c724901c433fa2cd4dc52efd8 pid=1708 runtime=io.containerd.runc.v2 Mar 17 18:23:52.853303 env[1217]: time="2025-03-17T18:23:52.851726170Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:23:52.853303 env[1217]: time="2025-03-17T18:23:52.851754554Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:23:52.853303 env[1217]: time="2025-03-17T18:23:52.851764215Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:23:52.853303 env[1217]: time="2025-03-17T18:23:52.851883737Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/af12061065aacc30f0550b224dcfa57e633fffa3a64f9a479dbd6fa0b19a72ae pid=1736 runtime=io.containerd.runc.v2 Mar 17 18:23:52.853506 env[1217]: time="2025-03-17T18:23:52.851216544Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:23:52.853506 env[1217]: time="2025-03-17T18:23:52.851253670Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:23:52.853506 env[1217]: time="2025-03-17T18:23:52.851264648Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:23:52.853506 env[1217]: time="2025-03-17T18:23:52.851481297Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/250cdca06c6d30527bb448f20ab31d8786ccabadb4ae1a2f8dad0ee40648487a pid=1731 runtime=io.containerd.runc.v2 Mar 17 18:23:52.858894 systemd[1]: Started cri-containerd-03b696f644db037f376562f848d75eab6d389a8c724901c433fa2cd4dc52efd8.scope. Mar 17 18:23:52.867503 systemd[1]: Started cri-containerd-af12061065aacc30f0550b224dcfa57e633fffa3a64f9a479dbd6fa0b19a72ae.scope. Mar 17 18:23:52.870725 systemd[1]: Started cri-containerd-250cdca06c6d30527bb448f20ab31d8786ccabadb4ae1a2f8dad0ee40648487a.scope. Mar 17 18:23:52.913098 env[1217]: time="2025-03-17T18:23:52.913052118Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:d79ab404294384d4bcc36fb5b5509bbb,Namespace:kube-system,Attempt:0,} returns sandbox id \"03b696f644db037f376562f848d75eab6d389a8c724901c433fa2cd4dc52efd8\"" Mar 17 18:23:52.914353 kubelet[1666]: E0317 18:23:52.914080 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:52.916603 env[1217]: time="2025-03-17T18:23:52.916563856Z" level=info msg="CreateContainer within sandbox \"03b696f644db037f376562f848d75eab6d389a8c724901c433fa2cd4dc52efd8\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Mar 17 18:23:52.922053 env[1217]: time="2025-03-17T18:23:52.922020687Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:23a18e2dc14f395c5f1bea711a5a9344,Namespace:kube-system,Attempt:0,} returns sandbox id \"af12061065aacc30f0550b224dcfa57e633fffa3a64f9a479dbd6fa0b19a72ae\"" Mar 17 18:23:52.922922 kubelet[1666]: E0317 18:23:52.922740 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:52.924410 env[1217]: time="2025-03-17T18:23:52.924376363Z" level=info msg="CreateContainer within sandbox \"af12061065aacc30f0550b224dcfa57e633fffa3a64f9a479dbd6fa0b19a72ae\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Mar 17 18:23:52.930529 env[1217]: time="2025-03-17T18:23:52.930495037Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:1518fccdb57a759873d8ac5bfb2e1095,Namespace:kube-system,Attempt:0,} returns sandbox id \"250cdca06c6d30527bb448f20ab31d8786ccabadb4ae1a2f8dad0ee40648487a\"" Mar 17 18:23:52.931232 kubelet[1666]: E0317 18:23:52.931053 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:52.932852 env[1217]: time="2025-03-17T18:23:52.932814146Z" level=info msg="CreateContainer within sandbox \"250cdca06c6d30527bb448f20ab31d8786ccabadb4ae1a2f8dad0ee40648487a\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Mar 17 18:23:52.933727 env[1217]: time="2025-03-17T18:23:52.933690444Z" level=info msg="CreateContainer within sandbox \"03b696f644db037f376562f848d75eab6d389a8c724901c433fa2cd4dc52efd8\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"bae16138d2f1ffdd31f1339eff50c9bd3a2d10f3b8e339d94b35fb8557784d77\"" Mar 17 18:23:52.934203 env[1217]: time="2025-03-17T18:23:52.934162665Z" level=info msg="StartContainer for \"bae16138d2f1ffdd31f1339eff50c9bd3a2d10f3b8e339d94b35fb8557784d77\"" Mar 17 18:23:52.936187 env[1217]: time="2025-03-17T18:23:52.936144764Z" level=info msg="CreateContainer within sandbox \"af12061065aacc30f0550b224dcfa57e633fffa3a64f9a479dbd6fa0b19a72ae\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"937eaadd8ca3bbc0526e46c66f6683c3db2db080eef74e5dc691a8944c80a1e6\"" Mar 17 18:23:52.936610 env[1217]: time="2025-03-17T18:23:52.936580179Z" level=info msg="StartContainer for \"937eaadd8ca3bbc0526e46c66f6683c3db2db080eef74e5dc691a8944c80a1e6\"" Mar 17 18:23:52.945439 env[1217]: time="2025-03-17T18:23:52.945396330Z" level=info msg="CreateContainer within sandbox \"250cdca06c6d30527bb448f20ab31d8786ccabadb4ae1a2f8dad0ee40648487a\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"287ed049c609b36544d89f1c0a7076a690d0229a3d16747e3d4d07462ccf9c26\"" Mar 17 18:23:52.945940 env[1217]: time="2025-03-17T18:23:52.945904479Z" level=info msg="StartContainer for \"287ed049c609b36544d89f1c0a7076a690d0229a3d16747e3d4d07462ccf9c26\"" Mar 17 18:23:52.954084 systemd[1]: Started cri-containerd-937eaadd8ca3bbc0526e46c66f6683c3db2db080eef74e5dc691a8944c80a1e6.scope. Mar 17 18:23:52.954859 systemd[1]: Started cri-containerd-bae16138d2f1ffdd31f1339eff50c9bd3a2d10f3b8e339d94b35fb8557784d77.scope. Mar 17 18:23:52.997280 systemd[1]: Started cri-containerd-287ed049c609b36544d89f1c0a7076a690d0229a3d16747e3d4d07462ccf9c26.scope. Mar 17 18:23:53.052414 env[1217]: time="2025-03-17T18:23:53.047498287Z" level=info msg="StartContainer for \"287ed049c609b36544d89f1c0a7076a690d0229a3d16747e3d4d07462ccf9c26\" returns successfully" Mar 17 18:23:53.054944 env[1217]: time="2025-03-17T18:23:53.054454229Z" level=info msg="StartContainer for \"bae16138d2f1ffdd31f1339eff50c9bd3a2d10f3b8e339d94b35fb8557784d77\" returns successfully" Mar 17 18:23:53.068506 kubelet[1666]: W0317 18:23:53.062674 1666 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.89:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:53.068506 kubelet[1666]: E0317 18:23:53.062740 1666 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.89:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.89:6443: connect: connection refused Mar 17 18:23:53.076397 env[1217]: time="2025-03-17T18:23:53.075425514Z" level=info msg="StartContainer for \"937eaadd8ca3bbc0526e46c66f6683c3db2db080eef74e5dc691a8944c80a1e6\" returns successfully" Mar 17 18:23:53.205842 kubelet[1666]: E0317 18:23:53.205772 1666 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.89:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.89:6443: connect: connection refused" interval="1.6s" Mar 17 18:23:53.312317 kubelet[1666]: I0317 18:23:53.312283 1666 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 18:23:53.820849 kubelet[1666]: E0317 18:23:53.820811 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:53.823087 kubelet[1666]: E0317 18:23:53.823067 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:53.824896 kubelet[1666]: E0317 18:23:53.824871 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:54.827574 kubelet[1666]: E0317 18:23:54.827541 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:54.827856 kubelet[1666]: E0317 18:23:54.827782 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:55.285154 kubelet[1666]: E0317 18:23:55.285103 1666 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Mar 17 18:23:55.378698 kubelet[1666]: I0317 18:23:55.378651 1666 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Mar 17 18:23:55.790385 kubelet[1666]: I0317 18:23:55.790343 1666 apiserver.go:52] "Watching apiserver" Mar 17 18:23:55.802263 kubelet[1666]: I0317 18:23:55.802229 1666 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 17 18:23:55.832942 kubelet[1666]: E0317 18:23:55.832883 1666 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:55.833358 kubelet[1666]: E0317 18:23:55.833314 1666 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:57.462868 systemd[1]: Reloading. Mar 17 18:23:57.505781 /usr/lib/systemd/system-generators/torcx-generator[1969]: time="2025-03-17T18:23:57Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" Mar 17 18:23:57.506135 /usr/lib/systemd/system-generators/torcx-generator[1969]: time="2025-03-17T18:23:57Z" level=info msg="torcx already run" Mar 17 18:23:57.559206 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Mar 17 18:23:57.559377 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Mar 17 18:23:57.575007 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Mar 17 18:23:57.656243 systemd[1]: Stopping kubelet.service... Mar 17 18:23:57.669758 systemd[1]: kubelet.service: Deactivated successfully. Mar 17 18:23:57.669933 systemd[1]: Stopped kubelet.service. Mar 17 18:23:57.671493 systemd[1]: Starting kubelet.service... Mar 17 18:23:57.756253 systemd[1]: Started kubelet.service. Mar 17 18:23:57.795809 kubelet[2012]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:23:57.795809 kubelet[2012]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 18:23:57.795809 kubelet[2012]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 18:23:57.796147 kubelet[2012]: I0317 18:23:57.795861 2012 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 18:23:57.799931 kubelet[2012]: I0317 18:23:57.799897 2012 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Mar 17 18:23:57.799931 kubelet[2012]: I0317 18:23:57.799923 2012 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 18:23:57.800088 kubelet[2012]: I0317 18:23:57.800074 2012 server.go:927] "Client rotation is on, will bootstrap in background" Mar 17 18:23:57.801362 kubelet[2012]: I0317 18:23:57.801322 2012 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 17 18:23:57.802724 kubelet[2012]: I0317 18:23:57.802476 2012 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Mar 17 18:23:57.807952 kubelet[2012]: I0317 18:23:57.807928 2012 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Mar 17 18:23:57.808257 kubelet[2012]: I0317 18:23:57.808231 2012 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 18:23:57.808568 kubelet[2012]: I0317 18:23:57.808367 2012 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Mar 17 18:23:57.808713 kubelet[2012]: I0317 18:23:57.808698 2012 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 18:23:57.808774 kubelet[2012]: I0317 18:23:57.808764 2012 container_manager_linux.go:301] "Creating device plugin manager" Mar 17 18:23:57.808858 kubelet[2012]: I0317 18:23:57.808847 2012 state_mem.go:36] "Initialized new in-memory state store" Mar 17 18:23:57.809015 kubelet[2012]: I0317 18:23:57.809000 2012 kubelet.go:400] "Attempting to sync node with API server" Mar 17 18:23:57.809098 kubelet[2012]: I0317 18:23:57.809086 2012 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 18:23:57.809171 kubelet[2012]: I0317 18:23:57.809162 2012 kubelet.go:312] "Adding apiserver pod source" Mar 17 18:23:57.809243 kubelet[2012]: I0317 18:23:57.809233 2012 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.813478 2012 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.813977 2012 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.814319 2012 server.go:1264] "Started kubelet" Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.814691 2012 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.814907 2012 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.814937 2012 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.815153 2012 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 18:23:57.818868 kubelet[2012]: I0317 18:23:57.815706 2012 server.go:455] "Adding debug handlers to kubelet server" Mar 17 18:23:57.819432 kubelet[2012]: I0317 18:23:57.819411 2012 volume_manager.go:291] "Starting Kubelet Volume Manager" Mar 17 18:23:57.819599 kubelet[2012]: I0317 18:23:57.819585 2012 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Mar 17 18:23:57.819879 kubelet[2012]: I0317 18:23:57.819865 2012 reconciler.go:26] "Reconciler: start to sync state" Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.821069 2012 factory.go:221] Registration of the systemd container factory successfully Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.821164 2012 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.823145 2012 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.823360 2012 factory.go:221] Registration of the containerd container factory successfully Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.824066 2012 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.824093 2012 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 18:23:57.832371 kubelet[2012]: I0317 18:23:57.824106 2012 kubelet.go:2337] "Starting kubelet main sync loop" Mar 17 18:23:57.832371 kubelet[2012]: E0317 18:23:57.824159 2012 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 18:23:57.832371 kubelet[2012]: E0317 18:23:57.828720 2012 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Mar 17 18:23:57.869075 kubelet[2012]: I0317 18:23:57.869048 2012 cpu_manager.go:214] "Starting CPU manager" policy="none" Mar 17 18:23:57.869177 kubelet[2012]: I0317 18:23:57.869074 2012 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Mar 17 18:23:57.869177 kubelet[2012]: I0317 18:23:57.869126 2012 state_mem.go:36] "Initialized new in-memory state store" Mar 17 18:23:57.869318 kubelet[2012]: I0317 18:23:57.869301 2012 state_mem.go:88] "Updated default CPUSet" cpuSet="" Mar 17 18:23:57.869369 kubelet[2012]: I0317 18:23:57.869318 2012 state_mem.go:96] "Updated CPUSet assignments" assignments={} Mar 17 18:23:57.869369 kubelet[2012]: I0317 18:23:57.869357 2012 policy_none.go:49] "None policy: Start" Mar 17 18:23:57.869827 kubelet[2012]: I0317 18:23:57.869806 2012 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 18:23:57.869911 kubelet[2012]: I0317 18:23:57.869899 2012 state_mem.go:35] "Initializing new in-memory state store" Mar 17 18:23:57.870079 kubelet[2012]: I0317 18:23:57.870065 2012 state_mem.go:75] "Updated machine memory state" Mar 17 18:23:57.873955 kubelet[2012]: I0317 18:23:57.873935 2012 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 18:23:57.874117 kubelet[2012]: I0317 18:23:57.874084 2012 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 18:23:57.874196 kubelet[2012]: I0317 18:23:57.874185 2012 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 18:23:57.922821 kubelet[2012]: I0317 18:23:57.922789 2012 kubelet_node_status.go:73] "Attempting to register node" node="localhost" Mar 17 18:23:57.924735 kubelet[2012]: I0317 18:23:57.924673 2012 topology_manager.go:215] "Topology Admit Handler" podUID="1518fccdb57a759873d8ac5bfb2e1095" podNamespace="kube-system" podName="kube-apiserver-localhost" Mar 17 18:23:57.928843 kubelet[2012]: I0317 18:23:57.924833 2012 topology_manager.go:215] "Topology Admit Handler" podUID="23a18e2dc14f395c5f1bea711a5a9344" podNamespace="kube-system" podName="kube-controller-manager-localhost" Mar 17 18:23:57.928932 kubelet[2012]: I0317 18:23:57.928904 2012 topology_manager.go:215] "Topology Admit Handler" podUID="d79ab404294384d4bcc36fb5b5509bbb" podNamespace="kube-system" podName="kube-scheduler-localhost" Mar 17 18:23:57.931420 kubelet[2012]: I0317 18:23:57.931399 2012 kubelet_node_status.go:112] "Node was previously registered" node="localhost" Mar 17 18:23:57.931569 kubelet[2012]: I0317 18:23:57.931556 2012 kubelet_node_status.go:76] "Successfully registered node" node="localhost" Mar 17 18:23:58.120780 kubelet[2012]: I0317 18:23:58.120740 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/1518fccdb57a759873d8ac5bfb2e1095-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"1518fccdb57a759873d8ac5bfb2e1095\") " pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:58.120780 kubelet[2012]: I0317 18:23:58.120779 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/1518fccdb57a759873d8ac5bfb2e1095-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"1518fccdb57a759873d8ac5bfb2e1095\") " pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:58.121019 kubelet[2012]: I0317 18:23:58.120802 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/1518fccdb57a759873d8ac5bfb2e1095-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"1518fccdb57a759873d8ac5bfb2e1095\") " pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:58.121019 kubelet[2012]: I0317 18:23:58.120832 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:58.121019 kubelet[2012]: I0317 18:23:58.120850 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:58.121019 kubelet[2012]: I0317 18:23:58.120867 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d79ab404294384d4bcc36fb5b5509bbb-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"d79ab404294384d4bcc36fb5b5509bbb\") " pod="kube-system/kube-scheduler-localhost" Mar 17 18:23:58.121019 kubelet[2012]: I0317 18:23:58.120891 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:58.121160 kubelet[2012]: I0317 18:23:58.120905 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:58.121160 kubelet[2012]: I0317 18:23:58.121056 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/23a18e2dc14f395c5f1bea711a5a9344-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"23a18e2dc14f395c5f1bea711a5a9344\") " pod="kube-system/kube-controller-manager-localhost" Mar 17 18:23:58.233574 kubelet[2012]: E0317 18:23:58.233538 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:58.235005 kubelet[2012]: E0317 18:23:58.234958 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:58.235184 kubelet[2012]: E0317 18:23:58.235159 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:58.461312 sudo[2048]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Mar 17 18:23:58.462028 sudo[2048]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Mar 17 18:23:58.809767 kubelet[2012]: I0317 18:23:58.809730 2012 apiserver.go:52] "Watching apiserver" Mar 17 18:23:58.820266 kubelet[2012]: I0317 18:23:58.820244 2012 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Mar 17 18:23:58.848387 kubelet[2012]: E0317 18:23:58.848082 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:58.850267 kubelet[2012]: E0317 18:23:58.848802 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:58.853265 kubelet[2012]: E0317 18:23:58.853232 2012 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Mar 17 18:23:58.853768 kubelet[2012]: E0317 18:23:58.853749 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:23:58.865693 kubelet[2012]: I0317 18:23:58.865647 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.8656183880000001 podStartE2EDuration="1.865618388s" podCreationTimestamp="2025-03-17 18:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:23:58.865462407 +0000 UTC m=+1.105584572" watchObservedRunningTime="2025-03-17 18:23:58.865618388 +0000 UTC m=+1.105740513" Mar 17 18:23:58.871772 kubelet[2012]: I0317 18:23:58.871733 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.871723187 podStartE2EDuration="1.871723187s" podCreationTimestamp="2025-03-17 18:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:23:58.871598778 +0000 UTC m=+1.111720943" watchObservedRunningTime="2025-03-17 18:23:58.871723187 +0000 UTC m=+1.111845352" Mar 17 18:23:58.904524 kubelet[2012]: I0317 18:23:58.904473 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.904459927 podStartE2EDuration="1.904459927s" podCreationTimestamp="2025-03-17 18:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:23:58.895408715 +0000 UTC m=+1.135530880" watchObservedRunningTime="2025-03-17 18:23:58.904459927 +0000 UTC m=+1.144582092" Mar 17 18:23:58.935782 sudo[2048]: pam_unix(sudo:session): session closed for user root Mar 17 18:23:59.849108 kubelet[2012]: E0317 18:23:59.849069 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:00.850199 kubelet[2012]: E0317 18:24:00.850160 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:01.911608 sudo[1318]: pam_unix(sudo:session): session closed for user root Mar 17 18:24:01.915214 sshd[1314]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:01.917213 systemd[1]: sshd@4-10.0.0.89:22-10.0.0.1:46576.service: Deactivated successfully. Mar 17 18:24:01.917916 systemd[1]: session-5.scope: Deactivated successfully. Mar 17 18:24:01.918076 systemd[1]: session-5.scope: Consumed 7.522s CPU time. Mar 17 18:24:01.918466 systemd-logind[1210]: Session 5 logged out. Waiting for processes to exit. Mar 17 18:24:01.919351 systemd-logind[1210]: Removed session 5. Mar 17 18:24:05.835255 kubelet[2012]: E0317 18:24:05.835220 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:05.857192 kubelet[2012]: E0317 18:24:05.857163 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:07.448684 kubelet[2012]: E0317 18:24:07.448650 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:07.859596 kubelet[2012]: E0317 18:24:07.859566 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:08.913375 kubelet[2012]: E0317 18:24:08.911579 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:09.861218 kubelet[2012]: E0317 18:24:09.861182 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:11.793457 kubelet[2012]: I0317 18:24:11.793419 2012 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Mar 17 18:24:11.793884 env[1217]: time="2025-03-17T18:24:11.793848281Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Mar 17 18:24:11.794145 kubelet[2012]: I0317 18:24:11.794116 2012 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Mar 17 18:24:12.176781 kubelet[2012]: I0317 18:24:12.176662 2012 topology_manager.go:215] "Topology Admit Handler" podUID="604697b2-9548-42a5-8d91-c81497085ee2" podNamespace="kube-system" podName="kube-proxy-2wq8r" Mar 17 18:24:12.181875 systemd[1]: Created slice kubepods-besteffort-pod604697b2_9548_42a5_8d91_c81497085ee2.slice. Mar 17 18:24:12.184337 kubelet[2012]: I0317 18:24:12.184291 2012 topology_manager.go:215] "Topology Admit Handler" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" podNamespace="kube-system" podName="cilium-5hhfz" Mar 17 18:24:12.195741 systemd[1]: Created slice kubepods-burstable-pod247563cd_d76b_4510_a0e9_9809c1eb949d.slice. Mar 17 18:24:12.312359 kubelet[2012]: I0317 18:24:12.312284 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-lib-modules\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312359 kubelet[2012]: I0317 18:24:12.312320 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/604697b2-9548-42a5-8d91-c81497085ee2-kube-proxy\") pod \"kube-proxy-2wq8r\" (UID: \"604697b2-9548-42a5-8d91-c81497085ee2\") " pod="kube-system/kube-proxy-2wq8r" Mar 17 18:24:12.312359 kubelet[2012]: I0317 18:24:12.312364 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-cgroup\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312570 kubelet[2012]: I0317 18:24:12.312381 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-kernel\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312570 kubelet[2012]: I0317 18:24:12.312404 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-422hk\" (UniqueName: \"kubernetes.io/projected/604697b2-9548-42a5-8d91-c81497085ee2-kube-api-access-422hk\") pod \"kube-proxy-2wq8r\" (UID: \"604697b2-9548-42a5-8d91-c81497085ee2\") " pod="kube-system/kube-proxy-2wq8r" Mar 17 18:24:12.312570 kubelet[2012]: I0317 18:24:12.312424 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-run\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312570 kubelet[2012]: I0317 18:24:12.312441 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-hostproc\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312570 kubelet[2012]: I0317 18:24:12.312461 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cni-path\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312570 kubelet[2012]: I0317 18:24:12.312476 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-net\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312723 kubelet[2012]: I0317 18:24:12.312491 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/604697b2-9548-42a5-8d91-c81497085ee2-lib-modules\") pod \"kube-proxy-2wq8r\" (UID: \"604697b2-9548-42a5-8d91-c81497085ee2\") " pod="kube-system/kube-proxy-2wq8r" Mar 17 18:24:12.312723 kubelet[2012]: I0317 18:24:12.312505 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-xtables-lock\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312723 kubelet[2012]: I0317 18:24:12.312522 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-config-path\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312723 kubelet[2012]: I0317 18:24:12.312536 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-hubble-tls\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312723 kubelet[2012]: I0317 18:24:12.312555 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8lck\" (UniqueName: \"kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-kube-api-access-j8lck\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312723 kubelet[2012]: I0317 18:24:12.312577 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-bpf-maps\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312859 kubelet[2012]: I0317 18:24:12.312593 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-etc-cni-netd\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.312859 kubelet[2012]: I0317 18:24:12.312608 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/604697b2-9548-42a5-8d91-c81497085ee2-xtables-lock\") pod \"kube-proxy-2wq8r\" (UID: \"604697b2-9548-42a5-8d91-c81497085ee2\") " pod="kube-system/kube-proxy-2wq8r" Mar 17 18:24:12.312859 kubelet[2012]: I0317 18:24:12.312640 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/247563cd-d76b-4510-a0e9-9809c1eb949d-clustermesh-secrets\") pod \"cilium-5hhfz\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " pod="kube-system/cilium-5hhfz" Mar 17 18:24:12.427423 kubelet[2012]: E0317 18:24:12.426837 2012 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Mar 17 18:24:12.427423 kubelet[2012]: E0317 18:24:12.426869 2012 projected.go:200] Error preparing data for projected volume kube-api-access-j8lck for pod kube-system/cilium-5hhfz: configmap "kube-root-ca.crt" not found Mar 17 18:24:12.427423 kubelet[2012]: E0317 18:24:12.426927 2012 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-kube-api-access-j8lck podName:247563cd-d76b-4510-a0e9-9809c1eb949d nodeName:}" failed. No retries permitted until 2025-03-17 18:24:12.92690959 +0000 UTC m=+15.167031755 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-j8lck" (UniqueName: "kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-kube-api-access-j8lck") pod "cilium-5hhfz" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d") : configmap "kube-root-ca.crt" not found Mar 17 18:24:12.429780 kubelet[2012]: E0317 18:24:12.429744 2012 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Mar 17 18:24:12.429780 kubelet[2012]: E0317 18:24:12.429774 2012 projected.go:200] Error preparing data for projected volume kube-api-access-422hk for pod kube-system/kube-proxy-2wq8r: configmap "kube-root-ca.crt" not found Mar 17 18:24:12.429885 kubelet[2012]: E0317 18:24:12.429818 2012 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/604697b2-9548-42a5-8d91-c81497085ee2-kube-api-access-422hk podName:604697b2-9548-42a5-8d91-c81497085ee2 nodeName:}" failed. No retries permitted until 2025-03-17 18:24:12.929803791 +0000 UTC m=+15.169925956 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-422hk" (UniqueName: "kubernetes.io/projected/604697b2-9548-42a5-8d91-c81497085ee2-kube-api-access-422hk") pod "kube-proxy-2wq8r" (UID: "604697b2-9548-42a5-8d91-c81497085ee2") : configmap "kube-root-ca.crt" not found Mar 17 18:24:12.925995 kubelet[2012]: I0317 18:24:12.925929 2012 topology_manager.go:215] "Topology Admit Handler" podUID="bfd8cef5-6135-47d2-86de-bf89377bb2fe" podNamespace="kube-system" podName="cilium-operator-599987898-g6x5p" Mar 17 18:24:12.931539 systemd[1]: Created slice kubepods-besteffort-podbfd8cef5_6135_47d2_86de_bf89377bb2fe.slice. Mar 17 18:24:13.092438 kubelet[2012]: E0317 18:24:13.092395 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:13.092996 env[1217]: time="2025-03-17T18:24:13.092941706Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-2wq8r,Uid:604697b2-9548-42a5-8d91-c81497085ee2,Namespace:kube-system,Attempt:0,}" Mar 17 18:24:13.097801 kubelet[2012]: E0317 18:24:13.097768 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:13.098390 env[1217]: time="2025-03-17T18:24:13.098124520Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-5hhfz,Uid:247563cd-d76b-4510-a0e9-9809c1eb949d,Namespace:kube-system,Attempt:0,}" Mar 17 18:24:13.110053 env[1217]: time="2025-03-17T18:24:13.109980603Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:24:13.110053 env[1217]: time="2025-03-17T18:24:13.110019448Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:24:13.110053 env[1217]: time="2025-03-17T18:24:13.110030809Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:24:13.110192 env[1217]: time="2025-03-17T18:24:13.110161064Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/998d61dd811cb74ff8e7f79e581121309835d2ddf4ef6a52bc414d5f4c7ec60e pid=2110 runtime=io.containerd.runc.v2 Mar 17 18:24:13.112817 env[1217]: time="2025-03-17T18:24:13.112748010Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:24:13.112817 env[1217]: time="2025-03-17T18:24:13.112782575Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:24:13.112817 env[1217]: time="2025-03-17T18:24:13.112792536Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:24:13.113177 env[1217]: time="2025-03-17T18:24:13.113129736Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50 pid=2127 runtime=io.containerd.runc.v2 Mar 17 18:24:13.118135 kubelet[2012]: I0317 18:24:13.118047 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzz8n\" (UniqueName: \"kubernetes.io/projected/bfd8cef5-6135-47d2-86de-bf89377bb2fe-kube-api-access-pzz8n\") pod \"cilium-operator-599987898-g6x5p\" (UID: \"bfd8cef5-6135-47d2-86de-bf89377bb2fe\") " pod="kube-system/cilium-operator-599987898-g6x5p" Mar 17 18:24:13.118135 kubelet[2012]: I0317 18:24:13.118091 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/bfd8cef5-6135-47d2-86de-bf89377bb2fe-cilium-config-path\") pod \"cilium-operator-599987898-g6x5p\" (UID: \"bfd8cef5-6135-47d2-86de-bf89377bb2fe\") " pod="kube-system/cilium-operator-599987898-g6x5p" Mar 17 18:24:13.121262 systemd[1]: Started cri-containerd-998d61dd811cb74ff8e7f79e581121309835d2ddf4ef6a52bc414d5f4c7ec60e.scope. Mar 17 18:24:13.135420 systemd[1]: Started cri-containerd-2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50.scope. Mar 17 18:24:13.161500 env[1217]: time="2025-03-17T18:24:13.160881268Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-2wq8r,Uid:604697b2-9548-42a5-8d91-c81497085ee2,Namespace:kube-system,Attempt:0,} returns sandbox id \"998d61dd811cb74ff8e7f79e581121309835d2ddf4ef6a52bc414d5f4c7ec60e\"" Mar 17 18:24:13.162915 kubelet[2012]: E0317 18:24:13.161792 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:13.165723 env[1217]: time="2025-03-17T18:24:13.165686037Z" level=info msg="CreateContainer within sandbox \"998d61dd811cb74ff8e7f79e581121309835d2ddf4ef6a52bc414d5f4c7ec60e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Mar 17 18:24:13.168153 env[1217]: time="2025-03-17T18:24:13.167360435Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-5hhfz,Uid:247563cd-d76b-4510-a0e9-9809c1eb949d,Namespace:kube-system,Attempt:0,} returns sandbox id \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\"" Mar 17 18:24:13.168236 kubelet[2012]: E0317 18:24:13.167923 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:13.169064 env[1217]: time="2025-03-17T18:24:13.169029712Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Mar 17 18:24:13.181939 env[1217]: time="2025-03-17T18:24:13.181365012Z" level=info msg="CreateContainer within sandbox \"998d61dd811cb74ff8e7f79e581121309835d2ddf4ef6a52bc414d5f4c7ec60e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"ac242d9d5137ad68b1cd37622e2d9d7ce0c9c17ef0e1a160f35a8a46e2ca840e\"" Mar 17 18:24:13.182291 env[1217]: time="2025-03-17T18:24:13.182185670Z" level=info msg="StartContainer for \"ac242d9d5137ad68b1cd37622e2d9d7ce0c9c17ef0e1a160f35a8a46e2ca840e\"" Mar 17 18:24:13.196170 systemd[1]: Started cri-containerd-ac242d9d5137ad68b1cd37622e2d9d7ce0c9c17ef0e1a160f35a8a46e2ca840e.scope. Mar 17 18:24:13.234552 kubelet[2012]: E0317 18:24:13.234525 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:13.235183 env[1217]: time="2025-03-17T18:24:13.235148018Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-g6x5p,Uid:bfd8cef5-6135-47d2-86de-bf89377bb2fe,Namespace:kube-system,Attempt:0,}" Mar 17 18:24:13.244228 env[1217]: time="2025-03-17T18:24:13.244184048Z" level=info msg="StartContainer for \"ac242d9d5137ad68b1cd37622e2d9d7ce0c9c17ef0e1a160f35a8a46e2ca840e\" returns successfully" Mar 17 18:24:13.252255 env[1217]: time="2025-03-17T18:24:13.252056380Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:24:13.252255 env[1217]: time="2025-03-17T18:24:13.252110466Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:24:13.252255 env[1217]: time="2025-03-17T18:24:13.252124908Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:24:13.252456 env[1217]: time="2025-03-17T18:24:13.252300369Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb pid=2223 runtime=io.containerd.runc.v2 Mar 17 18:24:13.263025 systemd[1]: Started cri-containerd-0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb.scope. Mar 17 18:24:13.308731 env[1217]: time="2025-03-17T18:24:13.308689763Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-g6x5p,Uid:bfd8cef5-6135-47d2-86de-bf89377bb2fe,Namespace:kube-system,Attempt:0,} returns sandbox id \"0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb\"" Mar 17 18:24:13.309586 kubelet[2012]: E0317 18:24:13.309562 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:13.869685 kubelet[2012]: E0317 18:24:13.869646 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:14.002348 update_engine[1211]: I0317 18:24:14.002286 1211 update_attempter.cc:509] Updating boot flags... Mar 17 18:24:17.050463 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1629507012.mount: Deactivated successfully. Mar 17 18:24:17.845766 kubelet[2012]: I0317 18:24:17.845672 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-2wq8r" podStartSLOduration=5.845657582 podStartE2EDuration="5.845657582s" podCreationTimestamp="2025-03-17 18:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:24:13.87791146 +0000 UTC m=+16.118033625" watchObservedRunningTime="2025-03-17 18:24:17.845657582 +0000 UTC m=+20.085779747" Mar 17 18:24:19.305751 env[1217]: time="2025-03-17T18:24:19.305678123Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:24:19.307006 env[1217]: time="2025-03-17T18:24:19.306980999Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:24:19.308580 env[1217]: time="2025-03-17T18:24:19.308548537Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:24:19.309077 env[1217]: time="2025-03-17T18:24:19.309052622Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Mar 17 18:24:19.312723 env[1217]: time="2025-03-17T18:24:19.312530890Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Mar 17 18:24:19.313836 env[1217]: time="2025-03-17T18:24:19.313790001Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 17 18:24:19.324156 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1525022405.mount: Deactivated successfully. Mar 17 18:24:19.326619 env[1217]: time="2025-03-17T18:24:19.326572332Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\"" Mar 17 18:24:19.327102 env[1217]: time="2025-03-17T18:24:19.327053455Z" level=info msg="StartContainer for \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\"" Mar 17 18:24:19.346548 systemd[1]: Started cri-containerd-2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b.scope. Mar 17 18:24:19.422092 env[1217]: time="2025-03-17T18:24:19.421104295Z" level=info msg="StartContainer for \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\" returns successfully" Mar 17 18:24:19.459257 systemd[1]: cri-containerd-2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b.scope: Deactivated successfully. Mar 17 18:24:19.565036 env[1217]: time="2025-03-17T18:24:19.564557267Z" level=info msg="shim disconnected" id=2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b Mar 17 18:24:19.565036 env[1217]: time="2025-03-17T18:24:19.564603071Z" level=warning msg="cleaning up after shim disconnected" id=2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b namespace=k8s.io Mar 17 18:24:19.565036 env[1217]: time="2025-03-17T18:24:19.564611712Z" level=info msg="cleaning up dead shim" Mar 17 18:24:19.572463 env[1217]: time="2025-03-17T18:24:19.572427243Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:24:19Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2443 runtime=io.containerd.runc.v2\n" Mar 17 18:24:19.879934 kubelet[2012]: E0317 18:24:19.879669 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:19.883229 env[1217]: time="2025-03-17T18:24:19.883098249Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 17 18:24:19.929690 env[1217]: time="2025-03-17T18:24:19.929508035Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\"" Mar 17 18:24:19.930904 env[1217]: time="2025-03-17T18:24:19.930576930Z" level=info msg="StartContainer for \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\"" Mar 17 18:24:19.944893 systemd[1]: Started cri-containerd-dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839.scope. Mar 17 18:24:19.988565 env[1217]: time="2025-03-17T18:24:19.988515696Z" level=info msg="StartContainer for \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\" returns successfully" Mar 17 18:24:19.998554 systemd[1]: systemd-sysctl.service: Deactivated successfully. Mar 17 18:24:19.998789 systemd[1]: Stopped systemd-sysctl.service. Mar 17 18:24:19.999352 systemd[1]: Stopping systemd-sysctl.service... Mar 17 18:24:20.001064 systemd[1]: Starting systemd-sysctl.service... Mar 17 18:24:20.002037 systemd[1]: cri-containerd-dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839.scope: Deactivated successfully. Mar 17 18:24:20.012083 systemd[1]: Finished systemd-sysctl.service. Mar 17 18:24:20.026626 env[1217]: time="2025-03-17T18:24:20.026581600Z" level=info msg="shim disconnected" id=dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839 Mar 17 18:24:20.026829 env[1217]: time="2025-03-17T18:24:20.026811419Z" level=warning msg="cleaning up after shim disconnected" id=dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839 namespace=k8s.io Mar 17 18:24:20.026913 env[1217]: time="2025-03-17T18:24:20.026899667Z" level=info msg="cleaning up dead shim" Mar 17 18:24:20.034093 env[1217]: time="2025-03-17T18:24:20.034053991Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:24:20Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2504 runtime=io.containerd.runc.v2\n" Mar 17 18:24:20.322312 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b-rootfs.mount: Deactivated successfully. Mar 17 18:24:20.415904 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount79146658.mount: Deactivated successfully. Mar 17 18:24:20.865687 env[1217]: time="2025-03-17T18:24:20.865641878Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:24:20.868025 env[1217]: time="2025-03-17T18:24:20.867977196Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:24:20.870602 env[1217]: time="2025-03-17T18:24:20.870559054Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Mar 17 18:24:20.871178 env[1217]: time="2025-03-17T18:24:20.871130702Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Mar 17 18:24:20.874371 env[1217]: time="2025-03-17T18:24:20.874308251Z" level=info msg="CreateContainer within sandbox \"0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Mar 17 18:24:20.882273 kubelet[2012]: E0317 18:24:20.882234 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:20.884337 env[1217]: time="2025-03-17T18:24:20.884295335Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 17 18:24:20.889052 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1608985172.mount: Deactivated successfully. Mar 17 18:24:20.895257 env[1217]: time="2025-03-17T18:24:20.895211738Z" level=info msg="CreateContainer within sandbox \"0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\"" Mar 17 18:24:20.895834 env[1217]: time="2025-03-17T18:24:20.895805268Z" level=info msg="StartContainer for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\"" Mar 17 18:24:20.904818 env[1217]: time="2025-03-17T18:24:20.904775866Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\"" Mar 17 18:24:20.905713 env[1217]: time="2025-03-17T18:24:20.905283189Z" level=info msg="StartContainer for \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\"" Mar 17 18:24:20.919507 systemd[1]: Started cri-containerd-bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825.scope. Mar 17 18:24:20.927732 systemd[1]: Started cri-containerd-24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717.scope. Mar 17 18:24:20.975902 env[1217]: time="2025-03-17T18:24:20.975855794Z" level=info msg="StartContainer for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" returns successfully" Mar 17 18:24:20.978405 env[1217]: time="2025-03-17T18:24:20.978372126Z" level=info msg="StartContainer for \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\" returns successfully" Mar 17 18:24:20.994463 systemd[1]: cri-containerd-24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717.scope: Deactivated successfully. Mar 17 18:24:21.081133 env[1217]: time="2025-03-17T18:24:21.081081308Z" level=info msg="shim disconnected" id=24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717 Mar 17 18:24:21.081133 env[1217]: time="2025-03-17T18:24:21.081130192Z" level=warning msg="cleaning up after shim disconnected" id=24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717 namespace=k8s.io Mar 17 18:24:21.081133 env[1217]: time="2025-03-17T18:24:21.081141673Z" level=info msg="cleaning up dead shim" Mar 17 18:24:21.087803 env[1217]: time="2025-03-17T18:24:21.087756288Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:24:21Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2600 runtime=io.containerd.runc.v2\n" Mar 17 18:24:21.886051 kubelet[2012]: E0317 18:24:21.886021 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:21.888126 env[1217]: time="2025-03-17T18:24:21.888083688Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 17 18:24:21.888397 kubelet[2012]: E0317 18:24:21.888162 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:21.901496 env[1217]: time="2025-03-17T18:24:21.901439728Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\"" Mar 17 18:24:21.902281 env[1217]: time="2025-03-17T18:24:21.902083660Z" level=info msg="StartContainer for \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\"" Mar 17 18:24:21.917183 kubelet[2012]: I0317 18:24:21.917096 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-g6x5p" podStartSLOduration=2.35586639 podStartE2EDuration="9.917076911s" podCreationTimestamp="2025-03-17 18:24:12 +0000 UTC" firstStartedPulling="2025-03-17 18:24:13.310895344 +0000 UTC m=+15.551017509" lastFinishedPulling="2025-03-17 18:24:20.872105865 +0000 UTC m=+23.112228030" observedRunningTime="2025-03-17 18:24:21.916754605 +0000 UTC m=+24.156876770" watchObservedRunningTime="2025-03-17 18:24:21.917076911 +0000 UTC m=+24.157199116" Mar 17 18:24:21.924254 systemd[1]: Started cri-containerd-ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac.scope. Mar 17 18:24:21.992348 env[1217]: time="2025-03-17T18:24:21.992283269Z" level=info msg="StartContainer for \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\" returns successfully" Mar 17 18:24:21.993244 systemd[1]: cri-containerd-ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac.scope: Deactivated successfully. Mar 17 18:24:22.013844 env[1217]: time="2025-03-17T18:24:22.013780560Z" level=info msg="shim disconnected" id=ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac Mar 17 18:24:22.013844 env[1217]: time="2025-03-17T18:24:22.013826244Z" level=warning msg="cleaning up after shim disconnected" id=ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac namespace=k8s.io Mar 17 18:24:22.013844 env[1217]: time="2025-03-17T18:24:22.013835444Z" level=info msg="cleaning up dead shim" Mar 17 18:24:22.024158 env[1217]: time="2025-03-17T18:24:22.023564557Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:24:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2658 runtime=io.containerd.runc.v2\n" Mar 17 18:24:22.322437 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac-rootfs.mount: Deactivated successfully. Mar 17 18:24:22.892212 kubelet[2012]: E0317 18:24:22.892184 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:22.895074 env[1217]: time="2025-03-17T18:24:22.895040041Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 17 18:24:22.898217 kubelet[2012]: E0317 18:24:22.898185 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:22.908630 env[1217]: time="2025-03-17T18:24:22.908577088Z" level=info msg="CreateContainer within sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\"" Mar 17 18:24:22.909315 env[1217]: time="2025-03-17T18:24:22.909283903Z" level=info msg="StartContainer for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\"" Mar 17 18:24:22.927360 systemd[1]: Started cri-containerd-75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256.scope. Mar 17 18:24:22.976692 env[1217]: time="2025-03-17T18:24:22.976585628Z" level=info msg="StartContainer for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" returns successfully" Mar 17 18:24:23.156036 kubelet[2012]: I0317 18:24:23.155767 2012 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Mar 17 18:24:23.175898 kubelet[2012]: I0317 18:24:23.175461 2012 topology_manager.go:215] "Topology Admit Handler" podUID="60ae35a1-d903-46d9-970b-1da361fda73a" podNamespace="kube-system" podName="coredns-7db6d8ff4d-rwcqc" Mar 17 18:24:23.176985 kubelet[2012]: I0317 18:24:23.176953 2012 topology_manager.go:215] "Topology Admit Handler" podUID="86a13f02-4036-43db-87c2-de3e8c20a13d" podNamespace="kube-system" podName="coredns-7db6d8ff4d-z5jfq" Mar 17 18:24:23.185379 systemd[1]: Created slice kubepods-burstable-pod60ae35a1_d903_46d9_970b_1da361fda73a.slice. Mar 17 18:24:23.186178 kubelet[2012]: I0317 18:24:23.186150 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7nrs\" (UniqueName: \"kubernetes.io/projected/60ae35a1-d903-46d9-970b-1da361fda73a-kube-api-access-g7nrs\") pod \"coredns-7db6d8ff4d-rwcqc\" (UID: \"60ae35a1-d903-46d9-970b-1da361fda73a\") " pod="kube-system/coredns-7db6d8ff4d-rwcqc" Mar 17 18:24:23.186342 kubelet[2012]: I0317 18:24:23.186308 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/60ae35a1-d903-46d9-970b-1da361fda73a-config-volume\") pod \"coredns-7db6d8ff4d-rwcqc\" (UID: \"60ae35a1-d903-46d9-970b-1da361fda73a\") " pod="kube-system/coredns-7db6d8ff4d-rwcqc" Mar 17 18:24:23.186446 kubelet[2012]: I0317 18:24:23.186431 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86a13f02-4036-43db-87c2-de3e8c20a13d-config-volume\") pod \"coredns-7db6d8ff4d-z5jfq\" (UID: \"86a13f02-4036-43db-87c2-de3e8c20a13d\") " pod="kube-system/coredns-7db6d8ff4d-z5jfq" Mar 17 18:24:23.186656 kubelet[2012]: I0317 18:24:23.186628 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkxm7\" (UniqueName: \"kubernetes.io/projected/86a13f02-4036-43db-87c2-de3e8c20a13d-kube-api-access-rkxm7\") pod \"coredns-7db6d8ff4d-z5jfq\" (UID: \"86a13f02-4036-43db-87c2-de3e8c20a13d\") " pod="kube-system/coredns-7db6d8ff4d-z5jfq" Mar 17 18:24:23.196850 systemd[1]: Created slice kubepods-burstable-pod86a13f02_4036_43db_87c2_de3e8c20a13d.slice. Mar 17 18:24:23.278382 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Mar 17 18:24:23.492703 kubelet[2012]: E0317 18:24:23.492595 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:23.493248 env[1217]: time="2025-03-17T18:24:23.493205301Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-rwcqc,Uid:60ae35a1-d903-46d9-970b-1da361fda73a,Namespace:kube-system,Attempt:0,}" Mar 17 18:24:23.503247 kubelet[2012]: E0317 18:24:23.503214 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:23.504021 env[1217]: time="2025-03-17T18:24:23.503973059Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-z5jfq,Uid:86a13f02-4036-43db-87c2-de3e8c20a13d,Namespace:kube-system,Attempt:0,}" Mar 17 18:24:23.509352 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Mar 17 18:24:23.895863 kubelet[2012]: E0317 18:24:23.895825 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:23.910657 kubelet[2012]: I0317 18:24:23.910581 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-5hhfz" podStartSLOduration=5.766713879 podStartE2EDuration="11.910554102s" podCreationTimestamp="2025-03-17 18:24:12 +0000 UTC" firstStartedPulling="2025-03-17 18:24:13.16849917 +0000 UTC m=+15.408621335" lastFinishedPulling="2025-03-17 18:24:19.312339393 +0000 UTC m=+21.552461558" observedRunningTime="2025-03-17 18:24:23.909213963 +0000 UTC m=+26.149336128" watchObservedRunningTime="2025-03-17 18:24:23.910554102 +0000 UTC m=+26.150676227" Mar 17 18:24:24.520919 systemd[1]: Started sshd@5-10.0.0.89:22-10.0.0.1:33198.service. Mar 17 18:24:24.568749 sshd[2831]: Accepted publickey for core from 10.0.0.1 port 33198 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:24.570618 sshd[2831]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:24.574071 systemd-logind[1210]: New session 6 of user core. Mar 17 18:24:24.574947 systemd[1]: Started session-6.scope. Mar 17 18:24:24.697906 sshd[2831]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:24.702461 systemd[1]: sshd@5-10.0.0.89:22-10.0.0.1:33198.service: Deactivated successfully. Mar 17 18:24:24.703148 systemd[1]: session-6.scope: Deactivated successfully. Mar 17 18:24:24.703635 systemd-logind[1210]: Session 6 logged out. Waiting for processes to exit. Mar 17 18:24:24.704325 systemd-logind[1210]: Removed session 6. Mar 17 18:24:24.899242 kubelet[2012]: E0317 18:24:24.899205 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:25.167256 systemd-networkd[1046]: cilium_host: Link UP Mar 17 18:24:25.168179 systemd-networkd[1046]: cilium_net: Link UP Mar 17 18:24:25.169925 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Mar 17 18:24:25.169991 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Mar 17 18:24:25.170086 systemd-networkd[1046]: cilium_net: Gained carrier Mar 17 18:24:25.170272 systemd-networkd[1046]: cilium_host: Gained carrier Mar 17 18:24:25.240465 systemd-networkd[1046]: cilium_net: Gained IPv6LL Mar 17 18:24:25.254967 systemd-networkd[1046]: cilium_vxlan: Link UP Mar 17 18:24:25.254976 systemd-networkd[1046]: cilium_vxlan: Gained carrier Mar 17 18:24:25.280639 systemd-networkd[1046]: cilium_host: Gained IPv6LL Mar 17 18:24:25.571371 kernel: NET: Registered PF_ALG protocol family Mar 17 18:24:25.900804 kubelet[2012]: E0317 18:24:25.900701 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:26.171760 systemd-networkd[1046]: lxc_health: Link UP Mar 17 18:24:26.182452 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Mar 17 18:24:26.180048 systemd-networkd[1046]: lxc_health: Gained carrier Mar 17 18:24:26.578853 systemd-networkd[1046]: lxcb061505b78a8: Link UP Mar 17 18:24:26.586837 systemd-networkd[1046]: lxc92b7124b4d91: Link UP Mar 17 18:24:26.596370 kernel: eth0: renamed from tmp84e9a Mar 17 18:24:26.608365 kernel: eth0: renamed from tmp88f0f Mar 17 18:24:26.614741 systemd-networkd[1046]: lxc92b7124b4d91: Gained carrier Mar 17 18:24:26.615359 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc92b7124b4d91: link becomes ready Mar 17 18:24:26.616142 systemd-networkd[1046]: lxcb061505b78a8: Gained carrier Mar 17 18:24:26.616395 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcb061505b78a8: link becomes ready Mar 17 18:24:26.849784 systemd-networkd[1046]: cilium_vxlan: Gained IPv6LL Mar 17 18:24:27.104436 kubelet[2012]: E0317 18:24:27.104307 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:27.809491 systemd-networkd[1046]: lxc92b7124b4d91: Gained IPv6LL Mar 17 18:24:27.936470 systemd-networkd[1046]: lxc_health: Gained IPv6LL Mar 17 18:24:28.512501 systemd-networkd[1046]: lxcb061505b78a8: Gained IPv6LL Mar 17 18:24:29.708190 systemd[1]: Started sshd@6-10.0.0.89:22-10.0.0.1:33208.service. Mar 17 18:24:29.764422 sshd[3229]: Accepted publickey for core from 10.0.0.1 port 33208 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:29.765872 sshd[3229]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:29.770306 systemd-logind[1210]: New session 7 of user core. Mar 17 18:24:29.771910 systemd[1]: Started session-7.scope. Mar 17 18:24:29.912385 sshd[3229]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:29.915326 systemd[1]: sshd@6-10.0.0.89:22-10.0.0.1:33208.service: Deactivated successfully. Mar 17 18:24:29.916051 systemd[1]: session-7.scope: Deactivated successfully. Mar 17 18:24:29.916879 systemd-logind[1210]: Session 7 logged out. Waiting for processes to exit. Mar 17 18:24:29.917649 systemd-logind[1210]: Removed session 7. Mar 17 18:24:30.242291 env[1217]: time="2025-03-17T18:24:30.242207713Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:24:30.242291 env[1217]: time="2025-03-17T18:24:30.242253636Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:24:30.242291 env[1217]: time="2025-03-17T18:24:30.242265076Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:24:30.242779 env[1217]: time="2025-03-17T18:24:30.242535812Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/88f0ffd52ee2f200236fdcb663f6cf6b473417ed666806ca3e9ccbeed888dbb9 pid=3255 runtime=io.containerd.runc.v2 Mar 17 18:24:30.249839 env[1217]: time="2025-03-17T18:24:30.249752898Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:24:30.249839 env[1217]: time="2025-03-17T18:24:30.249803821Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:24:30.249839 env[1217]: time="2025-03-17T18:24:30.249814902Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:24:30.250219 env[1217]: time="2025-03-17T18:24:30.250113198Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/84e9a7271a2c3c09ba724f97a5d10a8b38a42ecec314b208d904d5f0533ba18b pid=3273 runtime=io.containerd.runc.v2 Mar 17 18:24:30.260106 systemd[1]: Started cri-containerd-88f0ffd52ee2f200236fdcb663f6cf6b473417ed666806ca3e9ccbeed888dbb9.scope. Mar 17 18:24:30.268236 systemd[1]: Started cri-containerd-84e9a7271a2c3c09ba724f97a5d10a8b38a42ecec314b208d904d5f0533ba18b.scope. Mar 17 18:24:30.327267 systemd-resolved[1160]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 18:24:30.343010 systemd-resolved[1160]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Mar 17 18:24:30.345571 env[1217]: time="2025-03-17T18:24:30.345531134Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-rwcqc,Uid:60ae35a1-d903-46d9-970b-1da361fda73a,Namespace:kube-system,Attempt:0,} returns sandbox id \"84e9a7271a2c3c09ba724f97a5d10a8b38a42ecec314b208d904d5f0533ba18b\"" Mar 17 18:24:30.346250 kubelet[2012]: E0317 18:24:30.346230 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:30.348791 env[1217]: time="2025-03-17T18:24:30.348756996Z" level=info msg="CreateContainer within sandbox \"84e9a7271a2c3c09ba724f97a5d10a8b38a42ecec314b208d904d5f0533ba18b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 17 18:24:30.361727 env[1217]: time="2025-03-17T18:24:30.361683284Z" level=info msg="CreateContainer within sandbox \"84e9a7271a2c3c09ba724f97a5d10a8b38a42ecec314b208d904d5f0533ba18b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"37a896cf0d3157b012946877c33c278249c0080687430f8de4bb8b5ae2ce41c9\"" Mar 17 18:24:30.362581 env[1217]: time="2025-03-17T18:24:30.362542692Z" level=info msg="StartContainer for \"37a896cf0d3157b012946877c33c278249c0080687430f8de4bb8b5ae2ce41c9\"" Mar 17 18:24:30.364667 env[1217]: time="2025-03-17T18:24:30.364631130Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-z5jfq,Uid:86a13f02-4036-43db-87c2-de3e8c20a13d,Namespace:kube-system,Attempt:0,} returns sandbox id \"88f0ffd52ee2f200236fdcb663f6cf6b473417ed666806ca3e9ccbeed888dbb9\"" Mar 17 18:24:30.365466 kubelet[2012]: E0317 18:24:30.365440 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:30.370136 env[1217]: time="2025-03-17T18:24:30.370105599Z" level=info msg="CreateContainer within sandbox \"88f0ffd52ee2f200236fdcb663f6cf6b473417ed666806ca3e9ccbeed888dbb9\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Mar 17 18:24:30.384116 env[1217]: time="2025-03-17T18:24:30.383373146Z" level=info msg="CreateContainer within sandbox \"88f0ffd52ee2f200236fdcb663f6cf6b473417ed666806ca3e9ccbeed888dbb9\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3ee6e94dee6fdc040ac28a6a01425bbef8046a1a5f986039068f6d224f2028e6\"" Mar 17 18:24:30.383681 systemd[1]: Started cri-containerd-37a896cf0d3157b012946877c33c278249c0080687430f8de4bb8b5ae2ce41c9.scope. Mar 17 18:24:30.384776 env[1217]: time="2025-03-17T18:24:30.384749784Z" level=info msg="StartContainer for \"3ee6e94dee6fdc040ac28a6a01425bbef8046a1a5f986039068f6d224f2028e6\"" Mar 17 18:24:30.411656 systemd[1]: Started cri-containerd-3ee6e94dee6fdc040ac28a6a01425bbef8046a1a5f986039068f6d224f2028e6.scope. Mar 17 18:24:30.440891 env[1217]: time="2025-03-17T18:24:30.439915331Z" level=info msg="StartContainer for \"37a896cf0d3157b012946877c33c278249c0080687430f8de4bb8b5ae2ce41c9\" returns successfully" Mar 17 18:24:30.449800 env[1217]: time="2025-03-17T18:24:30.449754686Z" level=info msg="StartContainer for \"3ee6e94dee6fdc040ac28a6a01425bbef8046a1a5f986039068f6d224f2028e6\" returns successfully" Mar 17 18:24:30.910529 kubelet[2012]: E0317 18:24:30.910499 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:30.913009 kubelet[2012]: E0317 18:24:30.912950 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:30.941731 kubelet[2012]: I0317 18:24:30.941662 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-z5jfq" podStartSLOduration=18.941647478 podStartE2EDuration="18.941647478s" podCreationTimestamp="2025-03-17 18:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:24:30.925473567 +0000 UTC m=+33.165595732" watchObservedRunningTime="2025-03-17 18:24:30.941647478 +0000 UTC m=+33.181769643" Mar 17 18:24:30.953051 kubelet[2012]: I0317 18:24:30.953003 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-rwcqc" podStartSLOduration=18.952987357 podStartE2EDuration="18.952987357s" podCreationTimestamp="2025-03-17 18:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:24:30.942644254 +0000 UTC m=+33.182766459" watchObservedRunningTime="2025-03-17 18:24:30.952987357 +0000 UTC m=+33.193109522" Mar 17 18:24:31.915200 kubelet[2012]: E0317 18:24:31.915148 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:31.915753 kubelet[2012]: E0317 18:24:31.915662 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:32.916384 kubelet[2012]: E0317 18:24:32.916351 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:32.916721 kubelet[2012]: E0317 18:24:32.916410 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:34.917286 systemd[1]: Started sshd@7-10.0.0.89:22-10.0.0.1:52394.service. Mar 17 18:24:34.967647 sshd[3413]: Accepted publickey for core from 10.0.0.1 port 52394 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:34.969437 sshd[3413]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:34.973660 systemd-logind[1210]: New session 8 of user core. Mar 17 18:24:34.974192 systemd[1]: Started session-8.scope. Mar 17 18:24:35.087032 sshd[3413]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:35.091141 systemd[1]: sshd@7-10.0.0.89:22-10.0.0.1:52394.service: Deactivated successfully. Mar 17 18:24:35.092019 systemd[1]: session-8.scope: Deactivated successfully. Mar 17 18:24:35.092672 systemd-logind[1210]: Session 8 logged out. Waiting for processes to exit. Mar 17 18:24:35.093640 systemd-logind[1210]: Removed session 8. Mar 17 18:24:36.790204 kubelet[2012]: I0317 18:24:36.790170 2012 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 18:24:36.791775 kubelet[2012]: E0317 18:24:36.791754 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:36.924134 kubelet[2012]: E0317 18:24:36.924103 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:24:40.091519 systemd[1]: Started sshd@8-10.0.0.89:22-10.0.0.1:52402.service. Mar 17 18:24:40.134548 sshd[3428]: Accepted publickey for core from 10.0.0.1 port 52402 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:40.136317 sshd[3428]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:40.140414 systemd-logind[1210]: New session 9 of user core. Mar 17 18:24:40.140821 systemd[1]: Started session-9.scope. Mar 17 18:24:40.265778 sshd[3428]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:40.268346 systemd[1]: sshd@8-10.0.0.89:22-10.0.0.1:52402.service: Deactivated successfully. Mar 17 18:24:40.269102 systemd[1]: session-9.scope: Deactivated successfully. Mar 17 18:24:40.269754 systemd-logind[1210]: Session 9 logged out. Waiting for processes to exit. Mar 17 18:24:40.270480 systemd-logind[1210]: Removed session 9. Mar 17 18:24:45.269975 systemd[1]: Started sshd@9-10.0.0.89:22-10.0.0.1:39668.service. Mar 17 18:24:45.314647 sshd[3447]: Accepted publickey for core from 10.0.0.1 port 39668 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:45.316534 sshd[3447]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:45.319815 systemd-logind[1210]: New session 10 of user core. Mar 17 18:24:45.320674 systemd[1]: Started session-10.scope. Mar 17 18:24:45.432053 sshd[3447]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:45.435635 systemd[1]: Started sshd@10-10.0.0.89:22-10.0.0.1:39674.service. Mar 17 18:24:45.436154 systemd[1]: sshd@9-10.0.0.89:22-10.0.0.1:39668.service: Deactivated successfully. Mar 17 18:24:45.436911 systemd[1]: session-10.scope: Deactivated successfully. Mar 17 18:24:45.437525 systemd-logind[1210]: Session 10 logged out. Waiting for processes to exit. Mar 17 18:24:45.438240 systemd-logind[1210]: Removed session 10. Mar 17 18:24:45.482007 sshd[3461]: Accepted publickey for core from 10.0.0.1 port 39674 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:45.483237 sshd[3461]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:45.486575 systemd-logind[1210]: New session 11 of user core. Mar 17 18:24:45.488486 systemd[1]: Started session-11.scope. Mar 17 18:24:45.636678 sshd[3461]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:45.642036 systemd[1]: Started sshd@11-10.0.0.89:22-10.0.0.1:39688.service. Mar 17 18:24:45.645084 systemd[1]: sshd@10-10.0.0.89:22-10.0.0.1:39674.service: Deactivated successfully. Mar 17 18:24:45.645774 systemd[1]: session-11.scope: Deactivated successfully. Mar 17 18:24:45.647054 systemd-logind[1210]: Session 11 logged out. Waiting for processes to exit. Mar 17 18:24:45.653628 systemd-logind[1210]: Removed session 11. Mar 17 18:24:45.696666 sshd[3473]: Accepted publickey for core from 10.0.0.1 port 39688 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:45.697952 sshd[3473]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:45.701447 systemd-logind[1210]: New session 12 of user core. Mar 17 18:24:45.702243 systemd[1]: Started session-12.scope. Mar 17 18:24:45.815964 sshd[3473]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:45.818439 systemd[1]: sshd@11-10.0.0.89:22-10.0.0.1:39688.service: Deactivated successfully. Mar 17 18:24:45.819099 systemd[1]: session-12.scope: Deactivated successfully. Mar 17 18:24:45.819619 systemd-logind[1210]: Session 12 logged out. Waiting for processes to exit. Mar 17 18:24:45.820243 systemd-logind[1210]: Removed session 12. Mar 17 18:24:50.820529 systemd[1]: Started sshd@12-10.0.0.89:22-10.0.0.1:39694.service. Mar 17 18:24:50.864403 sshd[3491]: Accepted publickey for core from 10.0.0.1 port 39694 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:50.865881 sshd[3491]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:50.870237 systemd-logind[1210]: New session 13 of user core. Mar 17 18:24:50.871168 systemd[1]: Started session-13.scope. Mar 17 18:24:50.989342 sshd[3491]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:50.992908 systemd[1]: sshd@12-10.0.0.89:22-10.0.0.1:39694.service: Deactivated successfully. Mar 17 18:24:50.993681 systemd[1]: session-13.scope: Deactivated successfully. Mar 17 18:24:50.994388 systemd-logind[1210]: Session 13 logged out. Waiting for processes to exit. Mar 17 18:24:50.995582 systemd-logind[1210]: Removed session 13. Mar 17 18:24:55.994196 systemd[1]: Started sshd@13-10.0.0.89:22-10.0.0.1:60452.service. Mar 17 18:24:56.040866 sshd[3504]: Accepted publickey for core from 10.0.0.1 port 60452 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:56.042213 sshd[3504]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:56.046467 systemd-logind[1210]: New session 14 of user core. Mar 17 18:24:56.047355 systemd[1]: Started session-14.scope. Mar 17 18:24:56.154015 sshd[3504]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:56.157829 systemd[1]: Started sshd@14-10.0.0.89:22-10.0.0.1:60456.service. Mar 17 18:24:56.158381 systemd[1]: sshd@13-10.0.0.89:22-10.0.0.1:60452.service: Deactivated successfully. Mar 17 18:24:56.159042 systemd[1]: session-14.scope: Deactivated successfully. Mar 17 18:24:56.159565 systemd-logind[1210]: Session 14 logged out. Waiting for processes to exit. Mar 17 18:24:56.160318 systemd-logind[1210]: Removed session 14. Mar 17 18:24:56.201471 sshd[3517]: Accepted publickey for core from 10.0.0.1 port 60456 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:56.202707 sshd[3517]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:56.206133 systemd-logind[1210]: New session 15 of user core. Mar 17 18:24:56.206976 systemd[1]: Started session-15.scope. Mar 17 18:24:56.403320 sshd[3517]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:56.406777 systemd[1]: Started sshd@15-10.0.0.89:22-10.0.0.1:60468.service. Mar 17 18:24:56.407324 systemd[1]: sshd@14-10.0.0.89:22-10.0.0.1:60456.service: Deactivated successfully. Mar 17 18:24:56.408099 systemd[1]: session-15.scope: Deactivated successfully. Mar 17 18:24:56.408641 systemd-logind[1210]: Session 15 logged out. Waiting for processes to exit. Mar 17 18:24:56.409748 systemd-logind[1210]: Removed session 15. Mar 17 18:24:56.453588 sshd[3528]: Accepted publickey for core from 10.0.0.1 port 60468 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:56.454885 sshd[3528]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:56.458338 systemd-logind[1210]: New session 16 of user core. Mar 17 18:24:56.459177 systemd[1]: Started session-16.scope. Mar 17 18:24:57.800984 sshd[3528]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:57.804560 systemd[1]: Started sshd@16-10.0.0.89:22-10.0.0.1:60478.service. Mar 17 18:24:57.805112 systemd[1]: sshd@15-10.0.0.89:22-10.0.0.1:60468.service: Deactivated successfully. Mar 17 18:24:57.805998 systemd[1]: session-16.scope: Deactivated successfully. Mar 17 18:24:57.806711 systemd-logind[1210]: Session 16 logged out. Waiting for processes to exit. Mar 17 18:24:57.807882 systemd-logind[1210]: Removed session 16. Mar 17 18:24:57.859963 sshd[3547]: Accepted publickey for core from 10.0.0.1 port 60478 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:57.861219 sshd[3547]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:57.864749 systemd-logind[1210]: New session 17 of user core. Mar 17 18:24:57.865617 systemd[1]: Started session-17.scope. Mar 17 18:24:58.097865 sshd[3547]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:58.101728 systemd[1]: Started sshd@17-10.0.0.89:22-10.0.0.1:60484.service. Mar 17 18:24:58.102285 systemd[1]: sshd@16-10.0.0.89:22-10.0.0.1:60478.service: Deactivated successfully. Mar 17 18:24:58.103051 systemd[1]: session-17.scope: Deactivated successfully. Mar 17 18:24:58.106500 systemd-logind[1210]: Session 17 logged out. Waiting for processes to exit. Mar 17 18:24:58.121423 systemd-logind[1210]: Removed session 17. Mar 17 18:24:58.160855 sshd[3563]: Accepted publickey for core from 10.0.0.1 port 60484 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:24:58.162514 sshd[3563]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:24:58.166011 systemd-logind[1210]: New session 18 of user core. Mar 17 18:24:58.166899 systemd[1]: Started session-18.scope. Mar 17 18:24:58.286828 sshd[3563]: pam_unix(sshd:session): session closed for user core Mar 17 18:24:58.289445 systemd-logind[1210]: Session 18 logged out. Waiting for processes to exit. Mar 17 18:24:58.289695 systemd[1]: sshd@17-10.0.0.89:22-10.0.0.1:60484.service: Deactivated successfully. Mar 17 18:24:58.290360 systemd[1]: session-18.scope: Deactivated successfully. Mar 17 18:24:58.290923 systemd-logind[1210]: Removed session 18. Mar 17 18:25:03.291948 systemd[1]: Started sshd@18-10.0.0.89:22-10.0.0.1:42252.service. Mar 17 18:25:03.334703 sshd[3582]: Accepted publickey for core from 10.0.0.1 port 42252 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:03.335898 sshd[3582]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:03.339023 systemd-logind[1210]: New session 19 of user core. Mar 17 18:25:03.339876 systemd[1]: Started session-19.scope. Mar 17 18:25:03.443214 sshd[3582]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:03.445667 systemd[1]: sshd@18-10.0.0.89:22-10.0.0.1:42252.service: Deactivated successfully. Mar 17 18:25:03.446401 systemd[1]: session-19.scope: Deactivated successfully. Mar 17 18:25:03.446906 systemd-logind[1210]: Session 19 logged out. Waiting for processes to exit. Mar 17 18:25:03.447538 systemd-logind[1210]: Removed session 19. Mar 17 18:25:08.448465 systemd[1]: Started sshd@19-10.0.0.89:22-10.0.0.1:42266.service. Mar 17 18:25:08.491808 sshd[3595]: Accepted publickey for core from 10.0.0.1 port 42266 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:08.493056 sshd[3595]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:08.497190 systemd[1]: Started session-20.scope. Mar 17 18:25:08.497629 systemd-logind[1210]: New session 20 of user core. Mar 17 18:25:08.602220 sshd[3595]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:08.604539 systemd[1]: sshd@19-10.0.0.89:22-10.0.0.1:42266.service: Deactivated successfully. Mar 17 18:25:08.605271 systemd[1]: session-20.scope: Deactivated successfully. Mar 17 18:25:08.605848 systemd-logind[1210]: Session 20 logged out. Waiting for processes to exit. Mar 17 18:25:08.606604 systemd-logind[1210]: Removed session 20. Mar 17 18:25:13.606602 systemd[1]: Started sshd@20-10.0.0.89:22-10.0.0.1:42234.service. Mar 17 18:25:13.650662 sshd[3610]: Accepted publickey for core from 10.0.0.1 port 42234 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:13.652396 sshd[3610]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:13.656396 systemd-logind[1210]: New session 21 of user core. Mar 17 18:25:13.656863 systemd[1]: Started session-21.scope. Mar 17 18:25:13.765714 sshd[3610]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:13.769890 systemd[1]: Started sshd@21-10.0.0.89:22-10.0.0.1:42238.service. Mar 17 18:25:13.770529 systemd[1]: sshd@20-10.0.0.89:22-10.0.0.1:42234.service: Deactivated successfully. Mar 17 18:25:13.771434 systemd[1]: session-21.scope: Deactivated successfully. Mar 17 18:25:13.772007 systemd-logind[1210]: Session 21 logged out. Waiting for processes to exit. Mar 17 18:25:13.772778 systemd-logind[1210]: Removed session 21. Mar 17 18:25:13.813315 sshd[3622]: Accepted publickey for core from 10.0.0.1 port 42238 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:13.814864 sshd[3622]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:13.818413 systemd-logind[1210]: New session 22 of user core. Mar 17 18:25:13.819215 systemd[1]: Started session-22.scope. Mar 17 18:25:15.526325 env[1217]: time="2025-03-17T18:25:15.526276565Z" level=info msg="StopContainer for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" with timeout 30 (s)" Mar 17 18:25:15.526943 env[1217]: time="2025-03-17T18:25:15.526853015Z" level=info msg="Stop container \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" with signal terminated" Mar 17 18:25:15.539707 systemd[1]: cri-containerd-bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825.scope: Deactivated successfully. Mar 17 18:25:15.558827 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825-rootfs.mount: Deactivated successfully. Mar 17 18:25:15.562527 env[1217]: time="2025-03-17T18:25:15.562452024Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Mar 17 18:25:15.568045 env[1217]: time="2025-03-17T18:25:15.565783044Z" level=info msg="shim disconnected" id=bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825 Mar 17 18:25:15.568045 env[1217]: time="2025-03-17T18:25:15.565829685Z" level=warning msg="cleaning up after shim disconnected" id=bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825 namespace=k8s.io Mar 17 18:25:15.568045 env[1217]: time="2025-03-17T18:25:15.565837685Z" level=info msg="cleaning up dead shim" Mar 17 18:25:15.572238 env[1217]: time="2025-03-17T18:25:15.572202281Z" level=info msg="StopContainer for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" with timeout 2 (s)" Mar 17 18:25:15.572449 env[1217]: time="2025-03-17T18:25:15.572419725Z" level=info msg="Stop container \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" with signal terminated" Mar 17 18:25:15.574824 env[1217]: time="2025-03-17T18:25:15.574772208Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3676 runtime=io.containerd.runc.v2\n" Mar 17 18:25:15.577247 env[1217]: time="2025-03-17T18:25:15.577215292Z" level=info msg="StopContainer for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" returns successfully" Mar 17 18:25:15.577792 env[1217]: time="2025-03-17T18:25:15.577759422Z" level=info msg="StopPodSandbox for \"0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb\"" Mar 17 18:25:15.577968 env[1217]: time="2025-03-17T18:25:15.577934386Z" level=info msg="Container to stop \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 18:25:15.580035 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb-shm.mount: Deactivated successfully. Mar 17 18:25:15.582776 systemd-networkd[1046]: lxc_health: Link DOWN Mar 17 18:25:15.582783 systemd-networkd[1046]: lxc_health: Lost carrier Mar 17 18:25:15.591627 systemd[1]: cri-containerd-0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb.scope: Deactivated successfully. Mar 17 18:25:15.605672 systemd[1]: cri-containerd-75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256.scope: Deactivated successfully. Mar 17 18:25:15.605978 systemd[1]: cri-containerd-75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256.scope: Consumed 6.590s CPU time. Mar 17 18:25:15.614141 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb-rootfs.mount: Deactivated successfully. Mar 17 18:25:15.619953 env[1217]: time="2025-03-17T18:25:15.619903150Z" level=info msg="shim disconnected" id=0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb Mar 17 18:25:15.620618 env[1217]: time="2025-03-17T18:25:15.620589562Z" level=warning msg="cleaning up after shim disconnected" id=0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb namespace=k8s.io Mar 17 18:25:15.620717 env[1217]: time="2025-03-17T18:25:15.620701844Z" level=info msg="cleaning up dead shim" Mar 17 18:25:15.629007 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256-rootfs.mount: Deactivated successfully. Mar 17 18:25:15.630291 env[1217]: time="2025-03-17T18:25:15.630259658Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3723 runtime=io.containerd.runc.v2\n" Mar 17 18:25:15.630723 env[1217]: time="2025-03-17T18:25:15.630693626Z" level=info msg="TearDown network for sandbox \"0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb\" successfully" Mar 17 18:25:15.630845 env[1217]: time="2025-03-17T18:25:15.630823789Z" level=info msg="StopPodSandbox for \"0480f72af5fe6fad7cdc125b134693d94232acad1e9f6a0cd5181f9ceea7dadb\" returns successfully" Mar 17 18:25:15.632634 env[1217]: time="2025-03-17T18:25:15.632602101Z" level=info msg="shim disconnected" id=75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256 Mar 17 18:25:15.632762 env[1217]: time="2025-03-17T18:25:15.632741224Z" level=warning msg="cleaning up after shim disconnected" id=75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256 namespace=k8s.io Mar 17 18:25:15.632838 env[1217]: time="2025-03-17T18:25:15.632821785Z" level=info msg="cleaning up dead shim" Mar 17 18:25:15.641676 env[1217]: time="2025-03-17T18:25:15.641642946Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3741 runtime=io.containerd.runc.v2\n" Mar 17 18:25:15.644201 env[1217]: time="2025-03-17T18:25:15.644168632Z" level=info msg="StopContainer for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" returns successfully" Mar 17 18:25:15.644800 env[1217]: time="2025-03-17T18:25:15.644775163Z" level=info msg="StopPodSandbox for \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\"" Mar 17 18:25:15.644858 env[1217]: time="2025-03-17T18:25:15.644840124Z" level=info msg="Container to stop \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 18:25:15.644901 env[1217]: time="2025-03-17T18:25:15.644856164Z" level=info msg="Container to stop \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 18:25:15.644901 env[1217]: time="2025-03-17T18:25:15.644872365Z" level=info msg="Container to stop \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 18:25:15.644901 env[1217]: time="2025-03-17T18:25:15.644886205Z" level=info msg="Container to stop \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 18:25:15.644901 env[1217]: time="2025-03-17T18:25:15.644897885Z" level=info msg="Container to stop \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Mar 17 18:25:15.650214 systemd[1]: cri-containerd-2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50.scope: Deactivated successfully. Mar 17 18:25:15.672442 env[1217]: time="2025-03-17T18:25:15.672397586Z" level=info msg="shim disconnected" id=2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50 Mar 17 18:25:15.672668 env[1217]: time="2025-03-17T18:25:15.672650271Z" level=warning msg="cleaning up after shim disconnected" id=2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50 namespace=k8s.io Mar 17 18:25:15.672733 env[1217]: time="2025-03-17T18:25:15.672719872Z" level=info msg="cleaning up dead shim" Mar 17 18:25:15.682155 env[1217]: time="2025-03-17T18:25:15.682116203Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3772 runtime=io.containerd.runc.v2\n" Mar 17 18:25:15.682738 env[1217]: time="2025-03-17T18:25:15.682705894Z" level=info msg="TearDown network for sandbox \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" successfully" Mar 17 18:25:15.682853 env[1217]: time="2025-03-17T18:25:15.682834056Z" level=info msg="StopPodSandbox for \"2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50\" returns successfully" Mar 17 18:25:15.803499 kubelet[2012]: I0317 18:25:15.803374 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzz8n\" (UniqueName: \"kubernetes.io/projected/bfd8cef5-6135-47d2-86de-bf89377bb2fe-kube-api-access-pzz8n\") pod \"bfd8cef5-6135-47d2-86de-bf89377bb2fe\" (UID: \"bfd8cef5-6135-47d2-86de-bf89377bb2fe\") " Mar 17 18:25:15.803499 kubelet[2012]: I0317 18:25:15.803417 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-cgroup\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.803499 kubelet[2012]: I0317 18:25:15.803435 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-net\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.803499 kubelet[2012]: I0317 18:25:15.803452 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-bpf-maps\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.803499 kubelet[2012]: I0317 18:25:15.803466 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-etc-cni-netd\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.803499 kubelet[2012]: I0317 18:25:15.803483 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/bfd8cef5-6135-47d2-86de-bf89377bb2fe-cilium-config-path\") pod \"bfd8cef5-6135-47d2-86de-bf89377bb2fe\" (UID: \"bfd8cef5-6135-47d2-86de-bf89377bb2fe\") " Mar 17 18:25:15.804016 kubelet[2012]: I0317 18:25:15.803500 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/247563cd-d76b-4510-a0e9-9809c1eb949d-clustermesh-secrets\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804016 kubelet[2012]: I0317 18:25:15.803514 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-kernel\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804016 kubelet[2012]: I0317 18:25:15.803529 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-hostproc\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804016 kubelet[2012]: I0317 18:25:15.803544 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-run\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804016 kubelet[2012]: I0317 18:25:15.803556 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cni-path\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804016 kubelet[2012]: I0317 18:25:15.803571 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-lib-modules\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804152 kubelet[2012]: I0317 18:25:15.803588 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-hubble-tls\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804152 kubelet[2012]: I0317 18:25:15.803604 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8lck\" (UniqueName: \"kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-kube-api-access-j8lck\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804152 kubelet[2012]: I0317 18:25:15.803621 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-config-path\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.804152 kubelet[2012]: I0317 18:25:15.803636 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-xtables-lock\") pod \"247563cd-d76b-4510-a0e9-9809c1eb949d\" (UID: \"247563cd-d76b-4510-a0e9-9809c1eb949d\") " Mar 17 18:25:15.805496 kubelet[2012]: I0317 18:25:15.805381 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805589 kubelet[2012]: I0317 18:25:15.805503 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805589 kubelet[2012]: I0317 18:25:15.805520 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805589 kubelet[2012]: I0317 18:25:15.805536 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805589 kubelet[2012]: I0317 18:25:15.805550 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-hostproc" (OuterVolumeSpecName: "hostproc") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805589 kubelet[2012]: I0317 18:25:15.805564 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805714 kubelet[2012]: I0317 18:25:15.805577 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cni-path" (OuterVolumeSpecName: "cni-path") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805714 kubelet[2012]: I0317 18:25:15.805626 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805761 kubelet[2012]: I0317 18:25:15.805740 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.805799 kubelet[2012]: I0317 18:25:15.805779 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:15.810079 kubelet[2012]: I0317 18:25:15.810043 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfd8cef5-6135-47d2-86de-bf89377bb2fe-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "bfd8cef5-6135-47d2-86de-bf89377bb2fe" (UID: "bfd8cef5-6135-47d2-86de-bf89377bb2fe"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 18:25:15.811793 kubelet[2012]: I0317 18:25:15.811755 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/247563cd-d76b-4510-a0e9-9809c1eb949d-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 18:25:15.812461 kubelet[2012]: I0317 18:25:15.812433 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-kube-api-access-j8lck" (OuterVolumeSpecName: "kube-api-access-j8lck") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "kube-api-access-j8lck". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 18:25:15.812570 kubelet[2012]: I0317 18:25:15.812537 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfd8cef5-6135-47d2-86de-bf89377bb2fe-kube-api-access-pzz8n" (OuterVolumeSpecName: "kube-api-access-pzz8n") pod "bfd8cef5-6135-47d2-86de-bf89377bb2fe" (UID: "bfd8cef5-6135-47d2-86de-bf89377bb2fe"). InnerVolumeSpecName "kube-api-access-pzz8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 18:25:15.813456 kubelet[2012]: I0317 18:25:15.813314 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 18:25:15.813762 kubelet[2012]: I0317 18:25:15.813719 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "247563cd-d76b-4510-a0e9-9809c1eb949d" (UID: "247563cd-d76b-4510-a0e9-9809c1eb949d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 18:25:15.831857 systemd[1]: Removed slice kubepods-burstable-pod247563cd_d76b_4510_a0e9_9809c1eb949d.slice. Mar 17 18:25:15.831941 systemd[1]: kubepods-burstable-pod247563cd_d76b_4510_a0e9_9809c1eb949d.slice: Consumed 6.858s CPU time. Mar 17 18:25:15.832774 systemd[1]: Removed slice kubepods-besteffort-podbfd8cef5_6135_47d2_86de_bf89377bb2fe.slice. Mar 17 18:25:15.904850 kubelet[2012]: I0317 18:25:15.904796 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/bfd8cef5-6135-47d2-86de-bf89377bb2fe-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.904850 kubelet[2012]: I0317 18:25:15.904829 2012 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-bpf-maps\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.904850 kubelet[2012]: I0317 18:25:15.904839 2012 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.904850 kubelet[2012]: I0317 18:25:15.904847 2012 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/247563cd-d76b-4510-a0e9-9809c1eb949d-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.904850 kubelet[2012]: I0317 18:25:15.904855 2012 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.904850 kubelet[2012]: I0317 18:25:15.904863 2012 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-hostproc\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904870 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-run\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904878 2012 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cni-path\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904885 2012 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-lib-modules\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904892 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904899 2012 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-hubble-tls\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904908 2012 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-j8lck\" (UniqueName: \"kubernetes.io/projected/247563cd-d76b-4510-a0e9-9809c1eb949d-kube-api-access-j8lck\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904916 2012 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-xtables-lock\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905071 kubelet[2012]: I0317 18:25:15.904923 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905242 kubelet[2012]: I0317 18:25:15.904930 2012 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/247563cd-d76b-4510-a0e9-9809c1eb949d-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.905242 kubelet[2012]: I0317 18:25:15.904937 2012 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-pzz8n\" (UniqueName: \"kubernetes.io/projected/bfd8cef5-6135-47d2-86de-bf89377bb2fe-kube-api-access-pzz8n\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:15.993473 kubelet[2012]: I0317 18:25:15.993450 2012 scope.go:117] "RemoveContainer" containerID="75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256" Mar 17 18:25:15.996157 env[1217]: time="2025-03-17T18:25:15.996122002Z" level=info msg="RemoveContainer for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\"" Mar 17 18:25:16.000498 env[1217]: time="2025-03-17T18:25:16.000466201Z" level=info msg="RemoveContainer for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" returns successfully" Mar 17 18:25:16.000698 kubelet[2012]: I0317 18:25:16.000678 2012 scope.go:117] "RemoveContainer" containerID="ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac" Mar 17 18:25:16.002465 env[1217]: time="2025-03-17T18:25:16.002421197Z" level=info msg="RemoveContainer for \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\"" Mar 17 18:25:16.004752 env[1217]: time="2025-03-17T18:25:16.004723599Z" level=info msg="RemoveContainer for \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\" returns successfully" Mar 17 18:25:16.004937 kubelet[2012]: I0317 18:25:16.004914 2012 scope.go:117] "RemoveContainer" containerID="24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717" Mar 17 18:25:16.005904 env[1217]: time="2025-03-17T18:25:16.005879860Z" level=info msg="RemoveContainer for \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\"" Mar 17 18:25:16.008737 env[1217]: time="2025-03-17T18:25:16.008708552Z" level=info msg="RemoveContainer for \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\" returns successfully" Mar 17 18:25:16.008948 kubelet[2012]: I0317 18:25:16.008927 2012 scope.go:117] "RemoveContainer" containerID="dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839" Mar 17 18:25:16.013978 env[1217]: time="2025-03-17T18:25:16.012677105Z" level=info msg="RemoveContainer for \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\"" Mar 17 18:25:16.019609 env[1217]: time="2025-03-17T18:25:16.019569952Z" level=info msg="RemoveContainer for \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\" returns successfully" Mar 17 18:25:16.019805 kubelet[2012]: I0317 18:25:16.019785 2012 scope.go:117] "RemoveContainer" containerID="2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b" Mar 17 18:25:16.021797 env[1217]: time="2025-03-17T18:25:16.021770033Z" level=info msg="RemoveContainer for \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\"" Mar 17 18:25:16.023879 env[1217]: time="2025-03-17T18:25:16.023849071Z" level=info msg="RemoveContainer for \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\" returns successfully" Mar 17 18:25:16.024071 kubelet[2012]: I0317 18:25:16.024052 2012 scope.go:117] "RemoveContainer" containerID="75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256" Mar 17 18:25:16.024372 env[1217]: time="2025-03-17T18:25:16.024296799Z" level=error msg="ContainerStatus for \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\": not found" Mar 17 18:25:16.024478 kubelet[2012]: E0317 18:25:16.024455 2012 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\": not found" containerID="75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256" Mar 17 18:25:16.024560 kubelet[2012]: I0317 18:25:16.024487 2012 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256"} err="failed to get container status \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\": rpc error: code = NotFound desc = an error occurred when try to find container \"75fe37da61d511b807c356cc34152cd4d539df3c3f4dbec94730c4c02cd78256\": not found" Mar 17 18:25:16.024597 kubelet[2012]: I0317 18:25:16.024561 2012 scope.go:117] "RemoveContainer" containerID="ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac" Mar 17 18:25:16.024760 env[1217]: time="2025-03-17T18:25:16.024711807Z" level=error msg="ContainerStatus for \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\": not found" Mar 17 18:25:16.024952 kubelet[2012]: E0317 18:25:16.024852 2012 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\": not found" containerID="ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac" Mar 17 18:25:16.024952 kubelet[2012]: I0317 18:25:16.024877 2012 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac"} err="failed to get container status \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\": rpc error: code = NotFound desc = an error occurred when try to find container \"ed0adb5e16af78a8db813b1958acadd0d41b000905ce35c8d66fd1797dbc40ac\": not found" Mar 17 18:25:16.024952 kubelet[2012]: I0317 18:25:16.024892 2012 scope.go:117] "RemoveContainer" containerID="24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717" Mar 17 18:25:16.025140 env[1217]: time="2025-03-17T18:25:16.025021172Z" level=error msg="ContainerStatus for \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\": not found" Mar 17 18:25:16.025228 kubelet[2012]: E0317 18:25:16.025203 2012 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\": not found" containerID="24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717" Mar 17 18:25:16.025273 kubelet[2012]: I0317 18:25:16.025245 2012 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717"} err="failed to get container status \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\": rpc error: code = NotFound desc = an error occurred when try to find container \"24f16b43c792dcfa1f32408ce438dfc3c99ae2df221a9dd469bd21bd7d158717\": not found" Mar 17 18:25:16.025273 kubelet[2012]: I0317 18:25:16.025266 2012 scope.go:117] "RemoveContainer" containerID="dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839" Mar 17 18:25:16.025494 env[1217]: time="2025-03-17T18:25:16.025447020Z" level=error msg="ContainerStatus for \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\": not found" Mar 17 18:25:16.025579 kubelet[2012]: E0317 18:25:16.025559 2012 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\": not found" containerID="dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839" Mar 17 18:25:16.025620 kubelet[2012]: I0317 18:25:16.025581 2012 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839"} err="failed to get container status \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\": rpc error: code = NotFound desc = an error occurred when try to find container \"dfc4596e25c787807bc3cde8f3e817f7b27bad50945e20fa8d569b2273564839\": not found" Mar 17 18:25:16.025620 kubelet[2012]: I0317 18:25:16.025595 2012 scope.go:117] "RemoveContainer" containerID="2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b" Mar 17 18:25:16.025837 env[1217]: time="2025-03-17T18:25:16.025786307Z" level=error msg="ContainerStatus for \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\": not found" Mar 17 18:25:16.025977 kubelet[2012]: E0317 18:25:16.025959 2012 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\": not found" containerID="2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b" Mar 17 18:25:16.026027 kubelet[2012]: I0317 18:25:16.025978 2012 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b"} err="failed to get container status \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\": rpc error: code = NotFound desc = an error occurred when try to find container \"2edc014df57ac680ae05a0f6229bb8c91792650bde4acba50c494000ddb6d38b\": not found" Mar 17 18:25:16.026027 kubelet[2012]: I0317 18:25:16.025990 2012 scope.go:117] "RemoveContainer" containerID="bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825" Mar 17 18:25:16.026859 env[1217]: time="2025-03-17T18:25:16.026829086Z" level=info msg="RemoveContainer for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\"" Mar 17 18:25:16.028946 env[1217]: time="2025-03-17T18:25:16.028906764Z" level=info msg="RemoveContainer for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" returns successfully" Mar 17 18:25:16.029118 kubelet[2012]: I0317 18:25:16.029096 2012 scope.go:117] "RemoveContainer" containerID="bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825" Mar 17 18:25:16.029391 env[1217]: time="2025-03-17T18:25:16.029311171Z" level=error msg="ContainerStatus for \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\": not found" Mar 17 18:25:16.029544 kubelet[2012]: E0317 18:25:16.029522 2012 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\": not found" containerID="bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825" Mar 17 18:25:16.029630 kubelet[2012]: I0317 18:25:16.029610 2012 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825"} err="failed to get container status \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\": rpc error: code = NotFound desc = an error occurred when try to find container \"bdc20df0db019a42e8157e1e2caeabc55de37de57e1e6e8d73fea5763c871825\": not found" Mar 17 18:25:16.534825 systemd[1]: var-lib-kubelet-pods-bfd8cef5\x2d6135\x2d47d2\x2d86de\x2dbf89377bb2fe-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dpzz8n.mount: Deactivated successfully. Mar 17 18:25:16.534922 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50-rootfs.mount: Deactivated successfully. Mar 17 18:25:16.534983 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2714e7f0ebd5cfa96d337dfe155816faae61d6cd375612e4b37c4b30cf1f7b50-shm.mount: Deactivated successfully. Mar 17 18:25:16.535042 systemd[1]: var-lib-kubelet-pods-247563cd\x2dd76b\x2d4510\x2da0e9\x2d9809c1eb949d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dj8lck.mount: Deactivated successfully. Mar 17 18:25:16.535092 systemd[1]: var-lib-kubelet-pods-247563cd\x2dd76b\x2d4510\x2da0e9\x2d9809c1eb949d-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 17 18:25:16.535142 systemd[1]: var-lib-kubelet-pods-247563cd\x2dd76b\x2d4510\x2da0e9\x2d9809c1eb949d-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 17 18:25:17.483574 sshd[3622]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:17.487369 systemd[1]: Started sshd@22-10.0.0.89:22-10.0.0.1:42250.service. Mar 17 18:25:17.488721 systemd[1]: sshd@21-10.0.0.89:22-10.0.0.1:42238.service: Deactivated successfully. Mar 17 18:25:17.489428 systemd[1]: session-22.scope: Deactivated successfully. Mar 17 18:25:17.489586 systemd[1]: session-22.scope: Consumed 1.037s CPU time. Mar 17 18:25:17.490401 systemd-logind[1210]: Session 22 logged out. Waiting for processes to exit. Mar 17 18:25:17.491432 systemd-logind[1210]: Removed session 22. Mar 17 18:25:17.531080 sshd[3790]: Accepted publickey for core from 10.0.0.1 port 42250 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:17.532945 sshd[3790]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:17.536384 systemd-logind[1210]: New session 23 of user core. Mar 17 18:25:17.536734 systemd[1]: Started session-23.scope. Mar 17 18:25:17.827556 kubelet[2012]: I0317 18:25:17.827521 2012 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" path="/var/lib/kubelet/pods/247563cd-d76b-4510-a0e9-9809c1eb949d/volumes" Mar 17 18:25:17.828102 kubelet[2012]: I0317 18:25:17.828078 2012 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfd8cef5-6135-47d2-86de-bf89377bb2fe" path="/var/lib/kubelet/pods/bfd8cef5-6135-47d2-86de-bf89377bb2fe/volumes" Mar 17 18:25:17.888040 kubelet[2012]: E0317 18:25:17.888001 2012 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 17 18:25:18.640894 sshd[3790]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:18.650951 kubelet[2012]: I0317 18:25:18.650898 2012 topology_manager.go:215] "Topology Admit Handler" podUID="15d58983-4ce1-4014-93f3-08d3c86a8148" podNamespace="kube-system" podName="cilium-v2fgd" Mar 17 18:25:18.651066 kubelet[2012]: E0317 18:25:18.651040 2012 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" containerName="mount-cgroup" Mar 17 18:25:18.651066 kubelet[2012]: E0317 18:25:18.651051 2012 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="bfd8cef5-6135-47d2-86de-bf89377bb2fe" containerName="cilium-operator" Mar 17 18:25:18.651066 kubelet[2012]: E0317 18:25:18.651057 2012 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" containerName="clean-cilium-state" Mar 17 18:25:18.651066 kubelet[2012]: E0317 18:25:18.651063 2012 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" containerName="cilium-agent" Mar 17 18:25:18.651188 kubelet[2012]: E0317 18:25:18.651074 2012 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" containerName="apply-sysctl-overwrites" Mar 17 18:25:18.651188 kubelet[2012]: E0317 18:25:18.651080 2012 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" containerName="mount-bpf-fs" Mar 17 18:25:18.651188 kubelet[2012]: I0317 18:25:18.651105 2012 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfd8cef5-6135-47d2-86de-bf89377bb2fe" containerName="cilium-operator" Mar 17 18:25:18.651188 kubelet[2012]: I0317 18:25:18.651122 2012 memory_manager.go:354] "RemoveStaleState removing state" podUID="247563cd-d76b-4510-a0e9-9809c1eb949d" containerName="cilium-agent" Mar 17 18:25:18.651109 systemd[1]: Started sshd@23-10.0.0.89:22-10.0.0.1:42256.service. Mar 17 18:25:18.660018 systemd[1]: sshd@22-10.0.0.89:22-10.0.0.1:42250.service: Deactivated successfully. Mar 17 18:25:18.660755 systemd[1]: session-23.scope: Deactivated successfully. Mar 17 18:25:18.660916 systemd[1]: session-23.scope: Consumed 1.002s CPU time. Mar 17 18:25:18.663688 systemd-logind[1210]: Session 23 logged out. Waiting for processes to exit. Mar 17 18:25:18.664108 systemd[1]: Created slice kubepods-burstable-pod15d58983_4ce1_4014_93f3_08d3c86a8148.slice. Mar 17 18:25:18.665048 systemd-logind[1210]: Removed session 23. Mar 17 18:25:18.703912 sshd[3802]: Accepted publickey for core from 10.0.0.1 port 42256 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:18.705422 sshd[3802]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:18.709430 systemd-logind[1210]: New session 24 of user core. Mar 17 18:25:18.709840 systemd[1]: Started session-24.scope. Mar 17 18:25:18.819250 kubelet[2012]: I0317 18:25:18.819198 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-bpf-maps\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819250 kubelet[2012]: I0317 18:25:18.819249 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-kernel\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819445 kubelet[2012]: I0317 18:25:18.819270 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-hubble-tls\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819445 kubelet[2012]: I0317 18:25:18.819292 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cni-path\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819445 kubelet[2012]: I0317 18:25:18.819321 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-hostproc\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819445 kubelet[2012]: I0317 18:25:18.819359 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-lib-modules\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819445 kubelet[2012]: I0317 18:25:18.819379 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-clustermesh-secrets\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819445 kubelet[2012]: I0317 18:25:18.819396 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-config-path\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819622 kubelet[2012]: I0317 18:25:18.819411 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-run\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819622 kubelet[2012]: I0317 18:25:18.819440 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-cgroup\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819622 kubelet[2012]: I0317 18:25:18.819466 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k89h\" (UniqueName: \"kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-kube-api-access-9k89h\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819622 kubelet[2012]: I0317 18:25:18.819487 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-ipsec-secrets\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819622 kubelet[2012]: I0317 18:25:18.819515 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-net\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819622 kubelet[2012]: I0317 18:25:18.819530 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-etc-cni-netd\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.819752 kubelet[2012]: I0317 18:25:18.819546 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-xtables-lock\") pod \"cilium-v2fgd\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " pod="kube-system/cilium-v2fgd" Mar 17 18:25:18.834595 sshd[3802]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:18.839309 systemd[1]: Started sshd@24-10.0.0.89:22-10.0.0.1:42264.service. Mar 17 18:25:18.842987 kubelet[2012]: E0317 18:25:18.840959 2012 pod_workers.go:1298] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-ipsec-secrets cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-9k89h lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-v2fgd" podUID="15d58983-4ce1-4014-93f3-08d3c86a8148" Mar 17 18:25:18.842621 systemd[1]: sshd@23-10.0.0.89:22-10.0.0.1:42256.service: Deactivated successfully. Mar 17 18:25:18.843286 systemd[1]: session-24.scope: Deactivated successfully. Mar 17 18:25:18.844284 systemd-logind[1210]: Session 24 logged out. Waiting for processes to exit. Mar 17 18:25:18.846930 systemd-logind[1210]: Removed session 24. Mar 17 18:25:18.887436 sshd[3815]: Accepted publickey for core from 10.0.0.1 port 42264 ssh2: RSA SHA256:hoQCPKafrT/V1URQ18ch5K7mLY85DMM2OIJJf47c8zQ Mar 17 18:25:18.888667 sshd[3815]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Mar 17 18:25:18.892765 systemd-logind[1210]: New session 25 of user core. Mar 17 18:25:18.892974 systemd[1]: Started session-25.scope. Mar 17 18:25:19.121040 kubelet[2012]: I0317 18:25:19.120752 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-hubble-tls\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121040 kubelet[2012]: I0317 18:25:19.121048 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k89h\" (UniqueName: \"kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-kube-api-access-9k89h\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121246 kubelet[2012]: I0317 18:25:19.121069 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-xtables-lock\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121246 kubelet[2012]: I0317 18:25:19.121183 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.121246 kubelet[2012]: I0317 18:25:19.121216 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-hostproc\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121387 kubelet[2012]: I0317 18:25:19.121311 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-etc-cni-netd\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121387 kubelet[2012]: I0317 18:25:19.121357 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cni-path\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121387 kubelet[2012]: I0317 18:25:19.121374 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-net\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121456 kubelet[2012]: I0317 18:25:19.121389 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-cgroup\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121456 kubelet[2012]: I0317 18:25:19.121420 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-clustermesh-secrets\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121456 kubelet[2012]: I0317 18:25:19.121440 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-run\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121522 kubelet[2012]: I0317 18:25:19.121457 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-ipsec-secrets\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121522 kubelet[2012]: I0317 18:25:19.121488 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-kernel\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121522 kubelet[2012]: I0317 18:25:19.121504 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-lib-modules\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121522 kubelet[2012]: I0317 18:25:19.121519 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-bpf-maps\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121608 kubelet[2012]: I0317 18:25:19.121542 2012 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-config-path\") pod \"15d58983-4ce1-4014-93f3-08d3c86a8148\" (UID: \"15d58983-4ce1-4014-93f3-08d3c86a8148\") " Mar 17 18:25:19.121608 kubelet[2012]: I0317 18:25:19.121587 2012 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-xtables-lock\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.122292 kubelet[2012]: I0317 18:25:19.121733 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122292 kubelet[2012]: I0317 18:25:19.121766 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122292 kubelet[2012]: I0317 18:25:19.122211 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-hostproc" (OuterVolumeSpecName: "hostproc") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122292 kubelet[2012]: I0317 18:25:19.122241 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122292 kubelet[2012]: I0317 18:25:19.122256 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cni-path" (OuterVolumeSpecName: "cni-path") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122504 kubelet[2012]: I0317 18:25:19.122290 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122504 kubelet[2012]: I0317 18:25:19.122307 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122613 kubelet[2012]: I0317 18:25:19.122591 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.122693 kubelet[2012]: I0317 18:25:19.122680 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 18:25:19.123809 kubelet[2012]: I0317 18:25:19.123783 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-kube-api-access-9k89h" (OuterVolumeSpecName: "kube-api-access-9k89h") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "kube-api-access-9k89h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 18:25:19.124343 kubelet[2012]: I0317 18:25:19.124290 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 18:25:19.124938 systemd[1]: var-lib-kubelet-pods-15d58983\x2d4ce1\x2d4014\x2d93f3\x2d08d3c86a8148-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d9k89h.mount: Deactivated successfully. Mar 17 18:25:19.126083 kubelet[2012]: I0317 18:25:19.126056 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 18:25:19.126222 kubelet[2012]: I0317 18:25:19.126171 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 18:25:19.126299 kubelet[2012]: I0317 18:25:19.126254 2012 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "15d58983-4ce1-4014-93f3-08d3c86a8148" (UID: "15d58983-4ce1-4014-93f3-08d3c86a8148"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 18:25:19.126786 systemd[1]: var-lib-kubelet-pods-15d58983\x2d4ce1\x2d4014\x2d93f3\x2d08d3c86a8148-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Mar 17 18:25:19.126881 systemd[1]: var-lib-kubelet-pods-15d58983\x2d4ce1\x2d4014\x2d93f3\x2d08d3c86a8148-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Mar 17 18:25:19.126941 systemd[1]: var-lib-kubelet-pods-15d58983\x2d4ce1\x2d4014\x2d93f3\x2d08d3c86a8148-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Mar 17 18:25:19.222270 kubelet[2012]: I0317 18:25:19.222182 2012 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222270 kubelet[2012]: I0317 18:25:19.222213 2012 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-hostproc\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222270 kubelet[2012]: I0317 18:25:19.222230 2012 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222270 kubelet[2012]: I0317 18:25:19.222250 2012 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cni-path\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222270 kubelet[2012]: I0317 18:25:19.222264 2012 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222278 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-run\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222294 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222308 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-ipsec-secrets\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222322 2012 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222357 2012 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-lib-modules\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222371 2012 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/15d58983-4ce1-4014-93f3-08d3c86a8148-bpf-maps\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222385 2012 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15d58983-4ce1-4014-93f3-08d3c86a8148-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222459 kubelet[2012]: I0317 18:25:19.222400 2012 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-9k89h\" (UniqueName: \"kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-kube-api-access-9k89h\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.222622 kubelet[2012]: I0317 18:25:19.222414 2012 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/15d58983-4ce1-4014-93f3-08d3c86a8148-hubble-tls\") on node \"localhost\" DevicePath \"\"" Mar 17 18:25:19.831393 systemd[1]: Removed slice kubepods-burstable-pod15d58983_4ce1_4014_93f3_08d3c86a8148.slice. Mar 17 18:25:20.009874 kubelet[2012]: I0317 18:25:20.009813 2012 setters.go:580] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-03-17T18:25:20Z","lastTransitionTime":"2025-03-17T18:25:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Mar 17 18:25:20.036557 kubelet[2012]: I0317 18:25:20.036508 2012 topology_manager.go:215] "Topology Admit Handler" podUID="3aa6550d-e7d3-41c6-9e1b-084f78e59672" podNamespace="kube-system" podName="cilium-6vgbg" Mar 17 18:25:20.042287 systemd[1]: Created slice kubepods-burstable-pod3aa6550d_e7d3_41c6_9e1b_084f78e59672.slice. Mar 17 18:25:20.226515 kubelet[2012]: I0317 18:25:20.226401 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-host-proc-sys-net\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226515 kubelet[2012]: I0317 18:25:20.226446 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3aa6550d-e7d3-41c6-9e1b-084f78e59672-hubble-tls\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226515 kubelet[2012]: I0317 18:25:20.226467 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-bpf-maps\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226515 kubelet[2012]: I0317 18:25:20.226484 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-lib-modules\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226515 kubelet[2012]: I0317 18:25:20.226502 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-host-proc-sys-kernel\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226515 kubelet[2012]: I0317 18:25:20.226517 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72jq8\" (UniqueName: \"kubernetes.io/projected/3aa6550d-e7d3-41c6-9e1b-084f78e59672-kube-api-access-72jq8\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226762 kubelet[2012]: I0317 18:25:20.226534 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3aa6550d-e7d3-41c6-9e1b-084f78e59672-clustermesh-secrets\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226762 kubelet[2012]: I0317 18:25:20.226553 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-hostproc\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226762 kubelet[2012]: I0317 18:25:20.226569 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-etc-cni-netd\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226762 kubelet[2012]: I0317 18:25:20.226586 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/3aa6550d-e7d3-41c6-9e1b-084f78e59672-cilium-ipsec-secrets\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226762 kubelet[2012]: I0317 18:25:20.226600 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-cni-path\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226762 kubelet[2012]: I0317 18:25:20.226617 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-cilium-cgroup\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226919 kubelet[2012]: I0317 18:25:20.226632 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-xtables-lock\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226919 kubelet[2012]: I0317 18:25:20.226646 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3aa6550d-e7d3-41c6-9e1b-084f78e59672-cilium-config-path\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.226919 kubelet[2012]: I0317 18:25:20.226663 2012 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3aa6550d-e7d3-41c6-9e1b-084f78e59672-cilium-run\") pod \"cilium-6vgbg\" (UID: \"3aa6550d-e7d3-41c6-9e1b-084f78e59672\") " pod="kube-system/cilium-6vgbg" Mar 17 18:25:20.344055 kubelet[2012]: E0317 18:25:20.344006 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:20.344541 env[1217]: time="2025-03-17T18:25:20.344498557Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6vgbg,Uid:3aa6550d-e7d3-41c6-9e1b-084f78e59672,Namespace:kube-system,Attempt:0,}" Mar 17 18:25:20.358253 env[1217]: time="2025-03-17T18:25:20.358182339Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Mar 17 18:25:20.358253 env[1217]: time="2025-03-17T18:25:20.358223380Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Mar 17 18:25:20.358253 env[1217]: time="2025-03-17T18:25:20.358233700Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Mar 17 18:25:20.358715 env[1217]: time="2025-03-17T18:25:20.358678908Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb pid=3849 runtime=io.containerd.runc.v2 Mar 17 18:25:20.370146 systemd[1]: Started cri-containerd-ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb.scope. Mar 17 18:25:20.402377 env[1217]: time="2025-03-17T18:25:20.402060377Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6vgbg,Uid:3aa6550d-e7d3-41c6-9e1b-084f78e59672,Namespace:kube-system,Attempt:0,} returns sandbox id \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\"" Mar 17 18:25:20.403237 kubelet[2012]: E0317 18:25:20.403213 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:20.407191 env[1217]: time="2025-03-17T18:25:20.407126113Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Mar 17 18:25:20.417085 env[1217]: time="2025-03-17T18:25:20.417049663Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c\"" Mar 17 18:25:20.417525 env[1217]: time="2025-03-17T18:25:20.417495351Z" level=info msg="StartContainer for \"1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c\"" Mar 17 18:25:20.431149 systemd[1]: Started cri-containerd-1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c.scope. Mar 17 18:25:20.460937 env[1217]: time="2025-03-17T18:25:20.460881140Z" level=info msg="StartContainer for \"1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c\" returns successfully" Mar 17 18:25:20.469319 systemd[1]: cri-containerd-1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c.scope: Deactivated successfully. Mar 17 18:25:20.503214 env[1217]: time="2025-03-17T18:25:20.503087266Z" level=info msg="shim disconnected" id=1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c Mar 17 18:25:20.503214 env[1217]: time="2025-03-17T18:25:20.503132626Z" level=warning msg="cleaning up after shim disconnected" id=1ad9c218ee9f2007beaff966fbde1f32cdc5f0f959c079fb713d1f8d0668385c namespace=k8s.io Mar 17 18:25:20.503214 env[1217]: time="2025-03-17T18:25:20.503143347Z" level=info msg="cleaning up dead shim" Mar 17 18:25:20.511659 env[1217]: time="2025-03-17T18:25:20.511613828Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:20Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3935 runtime=io.containerd.runc.v2\n" Mar 17 18:25:21.008233 kubelet[2012]: E0317 18:25:21.008192 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:21.010190 env[1217]: time="2025-03-17T18:25:21.010150789Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Mar 17 18:25:21.020162 env[1217]: time="2025-03-17T18:25:21.020114221Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07\"" Mar 17 18:25:21.020636 env[1217]: time="2025-03-17T18:25:21.020612910Z" level=info msg="StartContainer for \"7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07\"" Mar 17 18:25:21.046271 systemd[1]: Started cri-containerd-7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07.scope. Mar 17 18:25:21.081252 env[1217]: time="2025-03-17T18:25:21.081208237Z" level=info msg="StartContainer for \"7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07\" returns successfully" Mar 17 18:25:21.090732 systemd[1]: cri-containerd-7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07.scope: Deactivated successfully. Mar 17 18:25:21.116788 env[1217]: time="2025-03-17T18:25:21.116745361Z" level=info msg="shim disconnected" id=7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07 Mar 17 18:25:21.117064 env[1217]: time="2025-03-17T18:25:21.117040967Z" level=warning msg="cleaning up after shim disconnected" id=7541c93fea745f4931dd0123dd7771f4aee0f13b83175c9d291a797043b62b07 namespace=k8s.io Mar 17 18:25:21.117132 env[1217]: time="2025-03-17T18:25:21.117118889Z" level=info msg="cleaning up dead shim" Mar 17 18:25:21.124081 env[1217]: time="2025-03-17T18:25:21.124045662Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:21Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3998 runtime=io.containerd.runc.v2\n" Mar 17 18:25:21.825214 kubelet[2012]: E0317 18:25:21.825174 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:21.828075 kubelet[2012]: I0317 18:25:21.828050 2012 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15d58983-4ce1-4014-93f3-08d3c86a8148" path="/var/lib/kubelet/pods/15d58983-4ce1-4014-93f3-08d3c86a8148/volumes" Mar 17 18:25:22.011882 kubelet[2012]: E0317 18:25:22.011719 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:22.016984 env[1217]: time="2025-03-17T18:25:22.013845436Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Mar 17 18:25:22.026357 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2957626542.mount: Deactivated successfully. Mar 17 18:25:22.037960 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1588462923.mount: Deactivated successfully. Mar 17 18:25:22.041733 env[1217]: time="2025-03-17T18:25:22.041695017Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22\"" Mar 17 18:25:22.042258 env[1217]: time="2025-03-17T18:25:22.042224267Z" level=info msg="StartContainer for \"6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22\"" Mar 17 18:25:22.058130 systemd[1]: Started cri-containerd-6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22.scope. Mar 17 18:25:22.093554 systemd[1]: cri-containerd-6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22.scope: Deactivated successfully. Mar 17 18:25:22.094815 env[1217]: time="2025-03-17T18:25:22.094761927Z" level=info msg="StartContainer for \"6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22\" returns successfully" Mar 17 18:25:22.114888 env[1217]: time="2025-03-17T18:25:22.114838717Z" level=info msg="shim disconnected" id=6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22 Mar 17 18:25:22.115047 env[1217]: time="2025-03-17T18:25:22.114889198Z" level=warning msg="cleaning up after shim disconnected" id=6730849b4ec6416457b4976525504879ca7f5cb3fa5e21275d86bb9fec1f1f22 namespace=k8s.io Mar 17 18:25:22.115047 env[1217]: time="2025-03-17T18:25:22.114900358Z" level=info msg="cleaning up dead shim" Mar 17 18:25:22.121135 env[1217]: time="2025-03-17T18:25:22.121098678Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4057 runtime=io.containerd.runc.v2\n" Mar 17 18:25:22.889375 kubelet[2012]: E0317 18:25:22.889317 2012 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Mar 17 18:25:23.016004 kubelet[2012]: E0317 18:25:23.015963 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:23.017906 env[1217]: time="2025-03-17T18:25:23.017844726Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Mar 17 18:25:23.031301 env[1217]: time="2025-03-17T18:25:23.031255668Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138\"" Mar 17 18:25:23.031883 env[1217]: time="2025-03-17T18:25:23.031844440Z" level=info msg="StartContainer for \"b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138\"" Mar 17 18:25:23.049321 systemd[1]: Started cri-containerd-b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138.scope. Mar 17 18:25:23.090295 systemd[1]: cri-containerd-b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138.scope: Deactivated successfully. Mar 17 18:25:23.092608 env[1217]: time="2025-03-17T18:25:23.092564708Z" level=info msg="StartContainer for \"b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138\" returns successfully" Mar 17 18:25:23.113184 env[1217]: time="2025-03-17T18:25:23.113125830Z" level=info msg="shim disconnected" id=b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138 Mar 17 18:25:23.113184 env[1217]: time="2025-03-17T18:25:23.113174511Z" level=warning msg="cleaning up after shim disconnected" id=b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138 namespace=k8s.io Mar 17 18:25:23.113184 env[1217]: time="2025-03-17T18:25:23.113184631Z" level=info msg="cleaning up dead shim" Mar 17 18:25:23.120091 env[1217]: time="2025-03-17T18:25:23.120002084Z" level=warning msg="cleanup warnings time=\"2025-03-17T18:25:23Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4110 runtime=io.containerd.runc.v2\n" Mar 17 18:25:23.332514 systemd[1]: run-containerd-runc-k8s.io-b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138-runc.McpzNs.mount: Deactivated successfully. Mar 17 18:25:23.332608 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b17cc220c86024da6645e069fd4d53226a3cd9bce0757b24082bcd76e98ec138-rootfs.mount: Deactivated successfully. Mar 17 18:25:24.020476 kubelet[2012]: E0317 18:25:24.020303 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:24.023391 env[1217]: time="2025-03-17T18:25:24.023348717Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Mar 17 18:25:24.036224 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1432509522.mount: Deactivated successfully. Mar 17 18:25:24.038453 env[1217]: time="2025-03-17T18:25:24.038379333Z" level=info msg="CreateContainer within sandbox \"ba2c5fa59c7d8ee10e83659a72bff1d710fa783ff919a695511ca0cdf73178bb\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0\"" Mar 17 18:25:24.038844 env[1217]: time="2025-03-17T18:25:24.038816102Z" level=info msg="StartContainer for \"cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0\"" Mar 17 18:25:24.059341 systemd[1]: Started cri-containerd-cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0.scope. Mar 17 18:25:24.095183 env[1217]: time="2025-03-17T18:25:24.095134891Z" level=info msg="StartContainer for \"cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0\" returns successfully" Mar 17 18:25:24.332623 systemd[1]: run-containerd-runc-k8s.io-cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0-runc.dMWbwa.mount: Deactivated successfully. Mar 17 18:25:24.349346 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106(gcm-aes-ce))) Mar 17 18:25:25.025201 kubelet[2012]: E0317 18:25:25.024721 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:25.038445 kubelet[2012]: I0317 18:25:25.038393 2012 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-6vgbg" podStartSLOduration=5.038380004 podStartE2EDuration="5.038380004s" podCreationTimestamp="2025-03-17 18:25:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-03-17 18:25:25.038047277 +0000 UTC m=+87.278169442" watchObservedRunningTime="2025-03-17 18:25:25.038380004 +0000 UTC m=+87.278502169" Mar 17 18:25:25.825757 kubelet[2012]: E0317 18:25:25.825371 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:26.345585 kubelet[2012]: E0317 18:25:26.345544 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:27.068539 systemd-networkd[1046]: lxc_health: Link UP Mar 17 18:25:27.075559 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Mar 17 18:25:27.074550 systemd-networkd[1046]: lxc_health: Gained carrier Mar 17 18:25:27.240852 systemd[1]: run-containerd-runc-k8s.io-cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0-runc.Qd5vp8.mount: Deactivated successfully. Mar 17 18:25:28.346344 kubelet[2012]: E0317 18:25:28.346277 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:28.480503 systemd-networkd[1046]: lxc_health: Gained IPv6LL Mar 17 18:25:29.032239 kubelet[2012]: E0317 18:25:29.032195 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:29.371046 systemd[1]: run-containerd-runc-k8s.io-cb5b7ef06c47ad24e11bac43eada2bb82d336eca2e11d0b2fa6123cfd8132cf0-runc.tzQJUW.mount: Deactivated successfully. Mar 17 18:25:29.826573 kubelet[2012]: E0317 18:25:29.826541 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:30.033626 kubelet[2012]: E0317 18:25:30.033595 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:30.825999 kubelet[2012]: E0317 18:25:30.825961 2012 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Mar 17 18:25:33.690219 sshd[3815]: pam_unix(sshd:session): session closed for user core Mar 17 18:25:33.692502 systemd[1]: sshd@24-10.0.0.89:22-10.0.0.1:42264.service: Deactivated successfully. Mar 17 18:25:33.693170 systemd[1]: session-25.scope: Deactivated successfully. Mar 17 18:25:33.693758 systemd-logind[1210]: Session 25 logged out. Waiting for processes to exit. Mar 17 18:25:33.694387 systemd-logind[1210]: Removed session 25.