Dec 13 14:09:37.705288 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Dec 13 14:09:37.705307 kernel: Linux version 5.15.173-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Fri Dec 13 12:58:58 -00 2024 Dec 13 14:09:37.705315 kernel: efi: EFI v2.70 by EDK II Dec 13 14:09:37.705321 kernel: efi: SMBIOS 3.0=0xd9260000 ACPI 2.0=0xd9240000 MEMATTR=0xda32b018 RNG=0xd9220018 MEMRESERVE=0xd9521c18 Dec 13 14:09:37.705326 kernel: random: crng init done Dec 13 14:09:37.705332 kernel: ACPI: Early table checksum verification disabled Dec 13 14:09:37.705338 kernel: ACPI: RSDP 0x00000000D9240000 000024 (v02 BOCHS ) Dec 13 14:09:37.705344 kernel: ACPI: XSDT 0x00000000D9230000 000064 (v01 BOCHS BXPC 00000001 01000013) Dec 13 14:09:37.705350 kernel: ACPI: FACP 0x00000000D91E0000 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705355 kernel: ACPI: DSDT 0x00000000D91F0000 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705361 kernel: ACPI: APIC 0x00000000D91D0000 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705366 kernel: ACPI: PPTT 0x00000000D91C0000 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705371 kernel: ACPI: GTDT 0x00000000D91B0000 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705377 kernel: ACPI: MCFG 0x00000000D91A0000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705384 kernel: ACPI: SPCR 0x00000000D9190000 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705390 kernel: ACPI: DBG2 0x00000000D9180000 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705396 kernel: ACPI: IORT 0x00000000D9170000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:09:37.705402 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Dec 13 14:09:37.705408 kernel: NUMA: Failed to initialise from firmware Dec 13 14:09:37.705413 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Dec 13 14:09:37.705419 kernel: NUMA: NODE_DATA [mem 0xdcb0b900-0xdcb10fff] Dec 13 14:09:37.705425 kernel: Zone ranges: Dec 13 14:09:37.705430 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Dec 13 14:09:37.705437 kernel: DMA32 empty Dec 13 14:09:37.705443 kernel: Normal empty Dec 13 14:09:37.705448 kernel: Movable zone start for each node Dec 13 14:09:37.705454 kernel: Early memory node ranges Dec 13 14:09:37.705460 kernel: node 0: [mem 0x0000000040000000-0x00000000d924ffff] Dec 13 14:09:37.705465 kernel: node 0: [mem 0x00000000d9250000-0x00000000d951ffff] Dec 13 14:09:37.705471 kernel: node 0: [mem 0x00000000d9520000-0x00000000dc7fffff] Dec 13 14:09:37.705477 kernel: node 0: [mem 0x00000000dc800000-0x00000000dc88ffff] Dec 13 14:09:37.705483 kernel: node 0: [mem 0x00000000dc890000-0x00000000dc89ffff] Dec 13 14:09:37.705488 kernel: node 0: [mem 0x00000000dc8a0000-0x00000000dc9bffff] Dec 13 14:09:37.705494 kernel: node 0: [mem 0x00000000dc9c0000-0x00000000dcffffff] Dec 13 14:09:37.705499 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Dec 13 14:09:37.705506 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Dec 13 14:09:37.705512 kernel: psci: probing for conduit method from ACPI. Dec 13 14:09:37.705518 kernel: psci: PSCIv1.1 detected in firmware. Dec 13 14:09:37.705523 kernel: psci: Using standard PSCI v0.2 function IDs Dec 13 14:09:37.705529 kernel: psci: Trusted OS migration not required Dec 13 14:09:37.705537 kernel: psci: SMC Calling Convention v1.1 Dec 13 14:09:37.705543 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Dec 13 14:09:37.705550 kernel: ACPI: SRAT not present Dec 13 14:09:37.705557 kernel: percpu: Embedded 30 pages/cpu s83032 r8192 d31656 u122880 Dec 13 14:09:37.705563 kernel: pcpu-alloc: s83032 r8192 d31656 u122880 alloc=30*4096 Dec 13 14:09:37.705570 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Dec 13 14:09:37.705576 kernel: Detected PIPT I-cache on CPU0 Dec 13 14:09:37.705582 kernel: CPU features: detected: GIC system register CPU interface Dec 13 14:09:37.705588 kernel: CPU features: detected: Hardware dirty bit management Dec 13 14:09:37.705594 kernel: CPU features: detected: Spectre-v4 Dec 13 14:09:37.705600 kernel: CPU features: detected: Spectre-BHB Dec 13 14:09:37.705607 kernel: CPU features: kernel page table isolation forced ON by KASLR Dec 13 14:09:37.705613 kernel: CPU features: detected: Kernel page table isolation (KPTI) Dec 13 14:09:37.705619 kernel: CPU features: detected: ARM erratum 1418040 Dec 13 14:09:37.705625 kernel: CPU features: detected: SSBS not fully self-synchronizing Dec 13 14:09:37.705631 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Dec 13 14:09:37.705637 kernel: Policy zone: DMA Dec 13 14:09:37.705644 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=5997a8cf94b1df1856dc785f0a7074604bbf4c21fdcca24a1996021471a77601 Dec 13 14:09:37.705651 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 13 14:09:37.705657 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Dec 13 14:09:37.705663 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 13 14:09:37.705669 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 13 14:09:37.705677 kernel: Memory: 2457404K/2572288K available (9792K kernel code, 2092K rwdata, 7576K rodata, 36416K init, 777K bss, 114884K reserved, 0K cma-reserved) Dec 13 14:09:37.705683 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Dec 13 14:09:37.705689 kernel: trace event string verifier disabled Dec 13 14:09:37.705695 kernel: rcu: Preemptible hierarchical RCU implementation. Dec 13 14:09:37.705701 kernel: rcu: RCU event tracing is enabled. Dec 13 14:09:37.705707 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Dec 13 14:09:37.705714 kernel: Trampoline variant of Tasks RCU enabled. Dec 13 14:09:37.705720 kernel: Tracing variant of Tasks RCU enabled. Dec 13 14:09:37.705726 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 13 14:09:37.705732 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Dec 13 14:09:37.705738 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Dec 13 14:09:37.705745 kernel: GICv3: 256 SPIs implemented Dec 13 14:09:37.705751 kernel: GICv3: 0 Extended SPIs implemented Dec 13 14:09:37.705757 kernel: GICv3: Distributor has no Range Selector support Dec 13 14:09:37.705763 kernel: Root IRQ handler: gic_handle_irq Dec 13 14:09:37.705769 kernel: GICv3: 16 PPIs implemented Dec 13 14:09:37.705775 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Dec 13 14:09:37.705781 kernel: ACPI: SRAT not present Dec 13 14:09:37.705787 kernel: ITS [mem 0x08080000-0x0809ffff] Dec 13 14:09:37.705793 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400b0000 (indirect, esz 8, psz 64K, shr 1) Dec 13 14:09:37.705799 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400c0000 (flat, esz 8, psz 64K, shr 1) Dec 13 14:09:37.705805 kernel: GICv3: using LPI property table @0x00000000400d0000 Dec 13 14:09:37.705811 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000000400e0000 Dec 13 14:09:37.705818 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Dec 13 14:09:37.705825 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Dec 13 14:09:37.705831 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Dec 13 14:09:37.705837 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Dec 13 14:09:37.705843 kernel: arm-pv: using stolen time PV Dec 13 14:09:37.705849 kernel: Console: colour dummy device 80x25 Dec 13 14:09:37.705856 kernel: ACPI: Core revision 20210730 Dec 13 14:09:37.705862 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Dec 13 14:09:37.705868 kernel: pid_max: default: 32768 minimum: 301 Dec 13 14:09:37.705875 kernel: LSM: Security Framework initializing Dec 13 14:09:37.705892 kernel: SELinux: Initializing. Dec 13 14:09:37.705900 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 14:09:37.705906 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 14:09:37.705912 kernel: rcu: Hierarchical SRCU implementation. Dec 13 14:09:37.705919 kernel: Platform MSI: ITS@0x8080000 domain created Dec 13 14:09:37.705925 kernel: PCI/MSI: ITS@0x8080000 domain created Dec 13 14:09:37.705931 kernel: Remapping and enabling EFI services. Dec 13 14:09:37.705937 kernel: smp: Bringing up secondary CPUs ... Dec 13 14:09:37.705949 kernel: Detected PIPT I-cache on CPU1 Dec 13 14:09:37.705958 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Dec 13 14:09:37.705965 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000000400f0000 Dec 13 14:09:37.705971 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Dec 13 14:09:37.705977 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Dec 13 14:09:37.705983 kernel: Detected PIPT I-cache on CPU2 Dec 13 14:09:37.705990 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Dec 13 14:09:37.705996 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040100000 Dec 13 14:09:37.706002 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Dec 13 14:09:37.706009 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Dec 13 14:09:37.706015 kernel: Detected PIPT I-cache on CPU3 Dec 13 14:09:37.706022 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Dec 13 14:09:37.706028 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040110000 Dec 13 14:09:37.706035 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Dec 13 14:09:37.706041 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Dec 13 14:09:37.706051 kernel: smp: Brought up 1 node, 4 CPUs Dec 13 14:09:37.706059 kernel: SMP: Total of 4 processors activated. Dec 13 14:09:37.706065 kernel: CPU features: detected: 32-bit EL0 Support Dec 13 14:09:37.706072 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Dec 13 14:09:37.706078 kernel: CPU features: detected: Common not Private translations Dec 13 14:09:37.706085 kernel: CPU features: detected: CRC32 instructions Dec 13 14:09:37.706092 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Dec 13 14:09:37.706098 kernel: CPU features: detected: LSE atomic instructions Dec 13 14:09:37.706106 kernel: CPU features: detected: Privileged Access Never Dec 13 14:09:37.706112 kernel: CPU features: detected: RAS Extension Support Dec 13 14:09:37.706119 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Dec 13 14:09:37.706126 kernel: CPU: All CPU(s) started at EL1 Dec 13 14:09:37.706132 kernel: alternatives: patching kernel code Dec 13 14:09:37.706140 kernel: devtmpfs: initialized Dec 13 14:09:37.706146 kernel: KASLR enabled Dec 13 14:09:37.706153 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 13 14:09:37.706159 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Dec 13 14:09:37.706166 kernel: pinctrl core: initialized pinctrl subsystem Dec 13 14:09:37.706172 kernel: SMBIOS 3.0.0 present. Dec 13 14:09:37.706179 kernel: DMI: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 Dec 13 14:09:37.706185 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 13 14:09:37.706192 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Dec 13 14:09:37.706200 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Dec 13 14:09:37.706210 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Dec 13 14:09:37.706217 kernel: audit: initializing netlink subsys (disabled) Dec 13 14:09:37.706224 kernel: audit: type=2000 audit(0.030:1): state=initialized audit_enabled=0 res=1 Dec 13 14:09:37.706231 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 13 14:09:37.706237 kernel: cpuidle: using governor menu Dec 13 14:09:37.706244 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Dec 13 14:09:37.706250 kernel: ASID allocator initialised with 32768 entries Dec 13 14:09:37.706257 kernel: ACPI: bus type PCI registered Dec 13 14:09:37.706265 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 13 14:09:37.706271 kernel: Serial: AMBA PL011 UART driver Dec 13 14:09:37.706278 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Dec 13 14:09:37.706284 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages Dec 13 14:09:37.706291 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Dec 13 14:09:37.706297 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages Dec 13 14:09:37.706304 kernel: cryptd: max_cpu_qlen set to 1000 Dec 13 14:09:37.706311 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Dec 13 14:09:37.706317 kernel: ACPI: Added _OSI(Module Device) Dec 13 14:09:37.706325 kernel: ACPI: Added _OSI(Processor Device) Dec 13 14:09:37.706331 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 13 14:09:37.706338 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 13 14:09:37.706344 kernel: ACPI: Added _OSI(Linux-Dell-Video) Dec 13 14:09:37.706351 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Dec 13 14:09:37.706357 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Dec 13 14:09:37.706364 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 13 14:09:37.706370 kernel: ACPI: Interpreter enabled Dec 13 14:09:37.706377 kernel: ACPI: Using GIC for interrupt routing Dec 13 14:09:37.706384 kernel: ACPI: MCFG table detected, 1 entries Dec 13 14:09:37.706391 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Dec 13 14:09:37.706398 kernel: printk: console [ttyAMA0] enabled Dec 13 14:09:37.706404 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 13 14:09:37.706519 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 13 14:09:37.706579 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Dec 13 14:09:37.706635 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Dec 13 14:09:37.706693 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Dec 13 14:09:37.706755 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Dec 13 14:09:37.706764 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Dec 13 14:09:37.706771 kernel: PCI host bridge to bus 0000:00 Dec 13 14:09:37.706839 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Dec 13 14:09:37.706935 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Dec 13 14:09:37.707006 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Dec 13 14:09:37.707063 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 13 14:09:37.707180 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Dec 13 14:09:37.707266 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Dec 13 14:09:37.707331 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Dec 13 14:09:37.707393 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Dec 13 14:09:37.707454 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Dec 13 14:09:37.707522 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Dec 13 14:09:37.707590 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Dec 13 14:09:37.707653 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Dec 13 14:09:37.707713 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Dec 13 14:09:37.707771 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Dec 13 14:09:37.707823 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Dec 13 14:09:37.707831 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Dec 13 14:09:37.707838 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Dec 13 14:09:37.707845 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Dec 13 14:09:37.707853 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Dec 13 14:09:37.707860 kernel: iommu: Default domain type: Translated Dec 13 14:09:37.707866 kernel: iommu: DMA domain TLB invalidation policy: strict mode Dec 13 14:09:37.707873 kernel: vgaarb: loaded Dec 13 14:09:37.707880 kernel: pps_core: LinuxPPS API ver. 1 registered Dec 13 14:09:37.707899 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Dec 13 14:09:37.707906 kernel: PTP clock support registered Dec 13 14:09:37.707913 kernel: Registered efivars operations Dec 13 14:09:37.707920 kernel: clocksource: Switched to clocksource arch_sys_counter Dec 13 14:09:37.707928 kernel: VFS: Disk quotas dquot_6.6.0 Dec 13 14:09:37.707935 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 13 14:09:37.707941 kernel: pnp: PnP ACPI init Dec 13 14:09:37.708019 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Dec 13 14:09:37.708030 kernel: pnp: PnP ACPI: found 1 devices Dec 13 14:09:37.708036 kernel: NET: Registered PF_INET protocol family Dec 13 14:09:37.708043 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Dec 13 14:09:37.708050 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Dec 13 14:09:37.708058 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 13 14:09:37.708065 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 13 14:09:37.708071 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Dec 13 14:09:37.708078 kernel: TCP: Hash tables configured (established 32768 bind 32768) Dec 13 14:09:37.708084 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 14:09:37.708091 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 14:09:37.708097 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 13 14:09:37.708104 kernel: PCI: CLS 0 bytes, default 64 Dec 13 14:09:37.708111 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Dec 13 14:09:37.708118 kernel: kvm [1]: HYP mode not available Dec 13 14:09:37.708125 kernel: Initialise system trusted keyrings Dec 13 14:09:37.708132 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Dec 13 14:09:37.708138 kernel: Key type asymmetric registered Dec 13 14:09:37.708144 kernel: Asymmetric key parser 'x509' registered Dec 13 14:09:37.708151 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Dec 13 14:09:37.708157 kernel: io scheduler mq-deadline registered Dec 13 14:09:37.708164 kernel: io scheduler kyber registered Dec 13 14:09:37.708171 kernel: io scheduler bfq registered Dec 13 14:09:37.708179 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Dec 13 14:09:37.708186 kernel: ACPI: button: Power Button [PWRB] Dec 13 14:09:37.708193 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Dec 13 14:09:37.708258 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Dec 13 14:09:37.708268 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 13 14:09:37.708274 kernel: thunder_xcv, ver 1.0 Dec 13 14:09:37.708281 kernel: thunder_bgx, ver 1.0 Dec 13 14:09:37.708287 kernel: nicpf, ver 1.0 Dec 13 14:09:37.708294 kernel: nicvf, ver 1.0 Dec 13 14:09:37.708363 kernel: rtc-efi rtc-efi.0: registered as rtc0 Dec 13 14:09:37.708417 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-12-13T14:09:37 UTC (1734098977) Dec 13 14:09:37.708425 kernel: hid: raw HID events driver (C) Jiri Kosina Dec 13 14:09:37.708432 kernel: NET: Registered PF_INET6 protocol family Dec 13 14:09:37.708438 kernel: Segment Routing with IPv6 Dec 13 14:09:37.708445 kernel: In-situ OAM (IOAM) with IPv6 Dec 13 14:09:37.708452 kernel: NET: Registered PF_PACKET protocol family Dec 13 14:09:37.708458 kernel: Key type dns_resolver registered Dec 13 14:09:37.708466 kernel: registered taskstats version 1 Dec 13 14:09:37.708473 kernel: Loading compiled-in X.509 certificates Dec 13 14:09:37.708480 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.173-flatcar: e011ba9949ade5a6d03f7a5e28171f7f59e70f8a' Dec 13 14:09:37.708486 kernel: Key type .fscrypt registered Dec 13 14:09:37.708493 kernel: Key type fscrypt-provisioning registered Dec 13 14:09:37.708499 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 13 14:09:37.708506 kernel: ima: Allocated hash algorithm: sha1 Dec 13 14:09:37.708513 kernel: ima: No architecture policies found Dec 13 14:09:37.708519 kernel: clk: Disabling unused clocks Dec 13 14:09:37.708527 kernel: Freeing unused kernel memory: 36416K Dec 13 14:09:37.708533 kernel: Run /init as init process Dec 13 14:09:37.708540 kernel: with arguments: Dec 13 14:09:37.708547 kernel: /init Dec 13 14:09:37.708553 kernel: with environment: Dec 13 14:09:37.708560 kernel: HOME=/ Dec 13 14:09:37.708566 kernel: TERM=linux Dec 13 14:09:37.708573 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 13 14:09:37.708581 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Dec 13 14:09:37.708590 systemd[1]: Detected virtualization kvm. Dec 13 14:09:37.708598 systemd[1]: Detected architecture arm64. Dec 13 14:09:37.708605 systemd[1]: Running in initrd. Dec 13 14:09:37.708612 systemd[1]: No hostname configured, using default hostname. Dec 13 14:09:37.708619 systemd[1]: Hostname set to . Dec 13 14:09:37.708626 systemd[1]: Initializing machine ID from VM UUID. Dec 13 14:09:37.708633 systemd[1]: Queued start job for default target initrd.target. Dec 13 14:09:37.708641 systemd[1]: Started systemd-ask-password-console.path. Dec 13 14:09:37.708648 systemd[1]: Reached target cryptsetup.target. Dec 13 14:09:37.708654 systemd[1]: Reached target paths.target. Dec 13 14:09:37.708662 systemd[1]: Reached target slices.target. Dec 13 14:09:37.708668 systemd[1]: Reached target swap.target. Dec 13 14:09:37.708675 systemd[1]: Reached target timers.target. Dec 13 14:09:37.708688 systemd[1]: Listening on iscsid.socket. Dec 13 14:09:37.708697 systemd[1]: Listening on iscsiuio.socket. Dec 13 14:09:37.708704 systemd[1]: Listening on systemd-journald-audit.socket. Dec 13 14:09:37.708711 systemd[1]: Listening on systemd-journald-dev-log.socket. Dec 13 14:09:37.709341 systemd[1]: Listening on systemd-journald.socket. Dec 13 14:09:37.709351 systemd[1]: Listening on systemd-networkd.socket. Dec 13 14:09:37.709359 systemd[1]: Listening on systemd-udevd-control.socket. Dec 13 14:09:37.709366 systemd[1]: Listening on systemd-udevd-kernel.socket. Dec 13 14:09:37.709373 systemd[1]: Reached target sockets.target. Dec 13 14:09:37.709380 systemd[1]: Starting kmod-static-nodes.service... Dec 13 14:09:37.709391 systemd[1]: Finished network-cleanup.service. Dec 13 14:09:37.709398 systemd[1]: Starting systemd-fsck-usr.service... Dec 13 14:09:37.709405 systemd[1]: Starting systemd-journald.service... Dec 13 14:09:37.709412 systemd[1]: Starting systemd-modules-load.service... Dec 13 14:09:37.709419 systemd[1]: Starting systemd-resolved.service... Dec 13 14:09:37.709426 systemd[1]: Starting systemd-vconsole-setup.service... Dec 13 14:09:37.709432 systemd[1]: Finished kmod-static-nodes.service. Dec 13 14:09:37.709439 systemd[1]: Finished systemd-fsck-usr.service. Dec 13 14:09:37.709446 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Dec 13 14:09:37.709454 systemd[1]: Finished systemd-vconsole-setup.service. Dec 13 14:09:37.709462 kernel: audit: type=1130 audit(1734098977.705:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.709470 systemd[1]: Starting dracut-cmdline-ask.service... Dec 13 14:09:37.709479 systemd-journald[289]: Journal started Dec 13 14:09:37.709534 systemd-journald[289]: Runtime Journal (/run/log/journal/63af08d6234f452d88df29a5cb10701d) is 6.0M, max 48.7M, 42.6M free. Dec 13 14:09:37.705000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.701239 systemd-modules-load[290]: Inserted module 'overlay' Dec 13 14:09:37.714212 systemd[1]: Started systemd-journald.service. Dec 13 14:09:37.714244 kernel: audit: type=1130 audit(1734098977.711:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.711000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.714000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.715205 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Dec 13 14:09:37.718265 kernel: audit: type=1130 audit(1734098977.714:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.722904 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 13 14:09:37.723497 systemd-resolved[291]: Positive Trust Anchors: Dec 13 14:09:37.723511 systemd-resolved[291]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 14:09:37.723537 systemd-resolved[291]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Dec 13 14:09:37.732986 kernel: Bridge firewalling registered Dec 13 14:09:37.733014 kernel: audit: type=1130 audit(1734098977.728:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.727542 systemd-resolved[291]: Defaulting to hostname 'linux'. Dec 13 14:09:37.732000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.735912 kernel: audit: type=1130 audit(1734098977.732:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.728835 systemd[1]: Started systemd-resolved.service. Dec 13 14:09:37.729822 systemd-modules-load[290]: Inserted module 'br_netfilter' Dec 13 14:09:37.730023 systemd[1]: Finished dracut-cmdline-ask.service. Dec 13 14:09:37.733553 systemd[1]: Reached target nss-lookup.target. Dec 13 14:09:37.737083 systemd[1]: Starting dracut-cmdline.service... Dec 13 14:09:37.741903 kernel: SCSI subsystem initialized Dec 13 14:09:37.745591 dracut-cmdline[308]: dracut-dracut-053 Dec 13 14:09:37.747730 dracut-cmdline[308]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=5997a8cf94b1df1856dc785f0a7074604bbf4c21fdcca24a1996021471a77601 Dec 13 14:09:37.752825 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 13 14:09:37.752843 kernel: device-mapper: uevent: version 1.0.3 Dec 13 14:09:37.752852 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Dec 13 14:09:37.755724 systemd-modules-load[290]: Inserted module 'dm_multipath' Dec 13 14:09:37.760992 kernel: audit: type=1130 audit(1734098977.756:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.756000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.756692 systemd[1]: Finished systemd-modules-load.service. Dec 13 14:09:37.758301 systemd[1]: Starting systemd-sysctl.service... Dec 13 14:09:37.764747 systemd[1]: Finished systemd-sysctl.service. Dec 13 14:09:37.764000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.767904 kernel: audit: type=1130 audit(1734098977.764:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.807910 kernel: Loading iSCSI transport class v2.0-870. Dec 13 14:09:37.821909 kernel: iscsi: registered transport (tcp) Dec 13 14:09:37.835905 kernel: iscsi: registered transport (qla4xxx) Dec 13 14:09:37.835922 kernel: QLogic iSCSI HBA Driver Dec 13 14:09:37.868963 systemd[1]: Finished dracut-cmdline.service. Dec 13 14:09:37.868000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.870354 systemd[1]: Starting dracut-pre-udev.service... Dec 13 14:09:37.872630 kernel: audit: type=1130 audit(1734098977.868:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:37.913908 kernel: raid6: neonx8 gen() 13732 MB/s Dec 13 14:09:37.930896 kernel: raid6: neonx8 xor() 10771 MB/s Dec 13 14:09:37.947898 kernel: raid6: neonx4 gen() 13491 MB/s Dec 13 14:09:37.964899 kernel: raid6: neonx4 xor() 11160 MB/s Dec 13 14:09:37.981895 kernel: raid6: neonx2 gen() 12971 MB/s Dec 13 14:09:37.998904 kernel: raid6: neonx2 xor() 10530 MB/s Dec 13 14:09:38.015898 kernel: raid6: neonx1 gen() 10442 MB/s Dec 13 14:09:38.032896 kernel: raid6: neonx1 xor() 8734 MB/s Dec 13 14:09:38.049905 kernel: raid6: int64x8 gen() 6235 MB/s Dec 13 14:09:38.066897 kernel: raid6: int64x8 xor() 3516 MB/s Dec 13 14:09:38.083897 kernel: raid6: int64x4 gen() 7212 MB/s Dec 13 14:09:38.100895 kernel: raid6: int64x4 xor() 3837 MB/s Dec 13 14:09:38.117896 kernel: raid6: int64x2 gen() 6117 MB/s Dec 13 14:09:38.134894 kernel: raid6: int64x2 xor() 3295 MB/s Dec 13 14:09:38.151902 kernel: raid6: int64x1 gen() 5030 MB/s Dec 13 14:09:38.169088 kernel: raid6: int64x1 xor() 2645 MB/s Dec 13 14:09:38.169110 kernel: raid6: using algorithm neonx8 gen() 13732 MB/s Dec 13 14:09:38.169127 kernel: raid6: .... xor() 10771 MB/s, rmw enabled Dec 13 14:09:38.169143 kernel: raid6: using neon recovery algorithm Dec 13 14:09:38.180239 kernel: xor: measuring software checksum speed Dec 13 14:09:38.180253 kernel: 8regs : 17224 MB/sec Dec 13 14:09:38.180262 kernel: 32regs : 20728 MB/sec Dec 13 14:09:38.181173 kernel: arm64_neon : 27832 MB/sec Dec 13 14:09:38.181200 kernel: xor: using function: arm64_neon (27832 MB/sec) Dec 13 14:09:38.235906 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no Dec 13 14:09:38.245659 systemd[1]: Finished dracut-pre-udev.service. Dec 13 14:09:38.245000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:38.247156 systemd[1]: Starting systemd-udevd.service... Dec 13 14:09:38.249532 kernel: audit: type=1130 audit(1734098978.245:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:38.245000 audit: BPF prog-id=7 op=LOAD Dec 13 14:09:38.245000 audit: BPF prog-id=8 op=LOAD Dec 13 14:09:38.262264 systemd-udevd[492]: Using default interface naming scheme 'v252'. Dec 13 14:09:38.266454 systemd[1]: Started systemd-udevd.service. Dec 13 14:09:38.267000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:38.268285 systemd[1]: Starting dracut-pre-trigger.service... Dec 13 14:09:38.278571 dracut-pre-trigger[499]: rd.md=0: removing MD RAID activation Dec 13 14:09:38.304683 systemd[1]: Finished dracut-pre-trigger.service. Dec 13 14:09:38.304000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:38.306073 systemd[1]: Starting systemd-udev-trigger.service... Dec 13 14:09:38.338599 systemd[1]: Finished systemd-udev-trigger.service. Dec 13 14:09:38.338000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:38.368020 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Dec 13 14:09:38.374561 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 13 14:09:38.374575 kernel: GPT:9289727 != 19775487 Dec 13 14:09:38.374584 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 13 14:09:38.374593 kernel: GPT:9289727 != 19775487 Dec 13 14:09:38.374600 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 13 14:09:38.374609 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:09:38.386724 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Dec 13 14:09:38.388903 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (550) Dec 13 14:09:38.392463 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Dec 13 14:09:38.397620 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Dec 13 14:09:38.398438 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Dec 13 14:09:38.404192 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Dec 13 14:09:38.405630 systemd[1]: Starting disk-uuid.service... Dec 13 14:09:38.411624 disk-uuid[564]: Primary Header is updated. Dec 13 14:09:38.411624 disk-uuid[564]: Secondary Entries is updated. Dec 13 14:09:38.411624 disk-uuid[564]: Secondary Header is updated. Dec 13 14:09:38.414182 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:09:38.423918 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:09:38.425908 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:09:39.426538 disk-uuid[565]: The operation has completed successfully. Dec 13 14:09:39.427529 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:09:39.449154 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 13 14:09:39.449247 systemd[1]: Finished disk-uuid.service. Dec 13 14:09:39.448000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.448000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.450565 systemd[1]: Starting verity-setup.service... Dec 13 14:09:39.463511 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Dec 13 14:09:39.484210 systemd[1]: Found device dev-mapper-usr.device. Dec 13 14:09:39.486252 systemd[1]: Mounting sysusr-usr.mount... Dec 13 14:09:39.488281 systemd[1]: Finished verity-setup.service. Dec 13 14:09:39.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.531904 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Dec 13 14:09:39.532343 systemd[1]: Mounted sysusr-usr.mount. Dec 13 14:09:39.532990 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Dec 13 14:09:39.533605 systemd[1]: Starting ignition-setup.service... Dec 13 14:09:39.535267 systemd[1]: Starting parse-ip-for-networkd.service... Dec 13 14:09:39.541233 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Dec 13 14:09:39.541274 kernel: BTRFS info (device vda6): using free space tree Dec 13 14:09:39.541285 kernel: BTRFS info (device vda6): has skinny extents Dec 13 14:09:39.549511 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 13 14:09:39.555233 systemd[1]: Finished ignition-setup.service. Dec 13 14:09:39.554000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.556533 systemd[1]: Starting ignition-fetch-offline.service... Dec 13 14:09:39.622000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.621530 systemd[1]: Finished parse-ip-for-networkd.service. Dec 13 14:09:39.623000 audit: BPF prog-id=9 op=LOAD Dec 13 14:09:39.624703 systemd[1]: Starting systemd-networkd.service... Dec 13 14:09:39.634240 ignition[648]: Ignition 2.14.0 Dec 13 14:09:39.634251 ignition[648]: Stage: fetch-offline Dec 13 14:09:39.634296 ignition[648]: no configs at "/usr/lib/ignition/base.d" Dec 13 14:09:39.634310 ignition[648]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:09:39.634439 ignition[648]: parsed url from cmdline: "" Dec 13 14:09:39.634442 ignition[648]: no config URL provided Dec 13 14:09:39.634447 ignition[648]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 14:09:39.634454 ignition[648]: no config at "/usr/lib/ignition/user.ign" Dec 13 14:09:39.634472 ignition[648]: op(1): [started] loading QEMU firmware config module Dec 13 14:09:39.634477 ignition[648]: op(1): executing: "modprobe" "qemu_fw_cfg" Dec 13 14:09:39.643599 ignition[648]: op(1): [finished] loading QEMU firmware config module Dec 13 14:09:39.643617 ignition[648]: QEMU firmware config was not found. Ignoring... Dec 13 14:09:39.653546 ignition[648]: parsing config with SHA512: 6125288daa40d7052653562903eb0640a2af9162e07f6ff153c8884cf8b54dd5d131a40c437955be859756ebc8accc2163c9167255827966fa1668e2c01152a9 Dec 13 14:09:39.659283 systemd-networkd[741]: lo: Link UP Dec 13 14:09:39.659297 systemd-networkd[741]: lo: Gained carrier Dec 13 14:09:39.659000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.659585 unknown[648]: fetched base config from "system" Dec 13 14:09:39.660143 ignition[648]: fetch-offline: fetch-offline passed Dec 13 14:09:39.659592 unknown[648]: fetched user config from "qemu" Dec 13 14:09:39.660197 ignition[648]: Ignition finished successfully Dec 13 14:09:39.659640 systemd-networkd[741]: Enumeration completed Dec 13 14:09:39.659715 systemd[1]: Started systemd-networkd.service. Dec 13 14:09:39.665000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.660652 systemd[1]: Reached target network.target. Dec 13 14:09:39.662318 systemd[1]: Starting iscsiuio.service... Dec 13 14:09:39.664294 systemd[1]: Finished ignition-fetch-offline.service. Dec 13 14:09:39.664976 systemd-networkd[741]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 14:09:39.665241 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Dec 13 14:09:39.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.665920 systemd[1]: Starting ignition-kargs.service... Dec 13 14:09:39.666606 systemd-networkd[741]: eth0: Link UP Dec 13 14:09:39.666610 systemd-networkd[741]: eth0: Gained carrier Dec 13 14:09:39.676653 iscsid[752]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Dec 13 14:09:39.676653 iscsid[752]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Dec 13 14:09:39.676653 iscsid[752]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Dec 13 14:09:39.676653 iscsid[752]: If using hardware iscsi like qla4xxx this message can be ignored. Dec 13 14:09:39.676653 iscsid[752]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Dec 13 14:09:39.676653 iscsid[752]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Dec 13 14:09:39.682000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.671012 systemd[1]: Started iscsiuio.service. Dec 13 14:09:39.684262 ignition[747]: Ignition 2.14.0 Dec 13 14:09:39.672618 systemd[1]: Starting iscsid.service... Dec 13 14:09:39.684268 ignition[747]: Stage: kargs Dec 13 14:09:39.689000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.678724 systemd[1]: Started iscsid.service. Dec 13 14:09:39.684363 ignition[747]: no configs at "/usr/lib/ignition/base.d" Dec 13 14:09:39.684771 systemd[1]: Starting dracut-initqueue.service... Dec 13 14:09:39.684372 ignition[747]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:09:39.689728 systemd[1]: Finished ignition-kargs.service. Dec 13 14:09:39.685044 ignition[747]: kargs: kargs passed Dec 13 14:09:39.689996 systemd-networkd[741]: eth0: DHCPv4 address 10.0.0.79/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 14:09:39.685090 ignition[747]: Ignition finished successfully Dec 13 14:09:39.691652 systemd[1]: Starting ignition-disks.service... Dec 13 14:09:39.699169 systemd[1]: Finished dracut-initqueue.service. Dec 13 14:09:39.699000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.700081 systemd[1]: Reached target remote-fs-pre.target. Dec 13 14:09:39.699658 ignition[760]: Ignition 2.14.0 Dec 13 14:09:39.701361 systemd[1]: Reached target remote-cryptsetup.target. Dec 13 14:09:39.699664 ignition[760]: Stage: disks Dec 13 14:09:39.702479 systemd[1]: Reached target remote-fs.target. Dec 13 14:09:39.699756 ignition[760]: no configs at "/usr/lib/ignition/base.d" Dec 13 14:09:39.704185 systemd[1]: Starting dracut-pre-mount.service... Dec 13 14:09:39.699765 ignition[760]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:09:39.705000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.705253 systemd[1]: Finished ignition-disks.service. Dec 13 14:09:39.700469 ignition[760]: disks: disks passed Dec 13 14:09:39.706219 systemd[1]: Reached target initrd-root-device.target. Dec 13 14:09:39.700510 ignition[760]: Ignition finished successfully Dec 13 14:09:39.712000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.707242 systemd[1]: Reached target local-fs-pre.target. Dec 13 14:09:39.708469 systemd[1]: Reached target local-fs.target. Dec 13 14:09:39.709834 systemd[1]: Reached target sysinit.target. Dec 13 14:09:39.710850 systemd[1]: Reached target basic.target. Dec 13 14:09:39.712157 systemd[1]: Finished dracut-pre-mount.service. Dec 13 14:09:39.713658 systemd[1]: Starting systemd-fsck-root.service... Dec 13 14:09:39.724752 systemd-fsck[776]: ROOT: clean, 621/553520 files, 56020/553472 blocks Dec 13 14:09:39.728352 systemd[1]: Finished systemd-fsck-root.service. Dec 13 14:09:39.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.730365 systemd[1]: Mounting sysroot.mount... Dec 13 14:09:39.735716 systemd[1]: Mounted sysroot.mount. Dec 13 14:09:39.736823 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Dec 13 14:09:39.736488 systemd[1]: Reached target initrd-root-fs.target. Dec 13 14:09:39.738543 systemd[1]: Mounting sysroot-usr.mount... Dec 13 14:09:39.739472 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Dec 13 14:09:39.739509 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 13 14:09:39.739532 systemd[1]: Reached target ignition-diskful.target. Dec 13 14:09:39.741238 systemd[1]: Mounted sysroot-usr.mount. Dec 13 14:09:39.742798 systemd[1]: Starting initrd-setup-root.service... Dec 13 14:09:39.747157 initrd-setup-root[786]: cut: /sysroot/etc/passwd: No such file or directory Dec 13 14:09:39.750802 initrd-setup-root[794]: cut: /sysroot/etc/group: No such file or directory Dec 13 14:09:39.753851 initrd-setup-root[802]: cut: /sysroot/etc/shadow: No such file or directory Dec 13 14:09:39.757755 initrd-setup-root[810]: cut: /sysroot/etc/gshadow: No such file or directory Dec 13 14:09:39.783969 systemd[1]: Finished initrd-setup-root.service. Dec 13 14:09:39.783000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.785971 systemd[1]: Starting ignition-mount.service... Dec 13 14:09:39.787141 systemd[1]: Starting sysroot-boot.service... Dec 13 14:09:39.790999 bash[828]: umount: /sysroot/usr/share/oem: not mounted. Dec 13 14:09:39.799096 ignition[829]: INFO : Ignition 2.14.0 Dec 13 14:09:39.799096 ignition[829]: INFO : Stage: mount Dec 13 14:09:39.800266 ignition[829]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 14:09:39.800266 ignition[829]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:09:39.800266 ignition[829]: INFO : mount: mount passed Dec 13 14:09:39.800266 ignition[829]: INFO : Ignition finished successfully Dec 13 14:09:39.800000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:39.800684 systemd[1]: Finished ignition-mount.service. Dec 13 14:09:39.808315 systemd[1]: Finished sysroot-boot.service. Dec 13 14:09:39.807000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:40.494503 systemd[1]: Mounting sysroot-usr-share-oem.mount... Dec 13 14:09:40.501323 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (838) Dec 13 14:09:40.501354 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Dec 13 14:09:40.501364 kernel: BTRFS info (device vda6): using free space tree Dec 13 14:09:40.502227 kernel: BTRFS info (device vda6): has skinny extents Dec 13 14:09:40.505083 systemd[1]: Mounted sysroot-usr-share-oem.mount. Dec 13 14:09:40.507307 systemd[1]: Starting ignition-files.service... Dec 13 14:09:40.520712 ignition[858]: INFO : Ignition 2.14.0 Dec 13 14:09:40.520712 ignition[858]: INFO : Stage: files Dec 13 14:09:40.521988 ignition[858]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 14:09:40.521988 ignition[858]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:09:40.521988 ignition[858]: DEBUG : files: compiled without relabeling support, skipping Dec 13 14:09:40.524503 ignition[858]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 13 14:09:40.524503 ignition[858]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 13 14:09:40.529525 ignition[858]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 13 14:09:40.530562 ignition[858]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 13 14:09:40.530562 ignition[858]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 13 14:09:40.530377 unknown[858]: wrote ssh authorized keys file for user: core Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/home/core/install.sh" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/home/core/install.sh" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 14:09:40.533449 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Dec 13 14:09:40.833921 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET result: OK Dec 13 14:09:41.462545 ignition[858]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Dec 13 14:09:41.462545 ignition[858]: INFO : files: op(7): [started] processing unit "coreos-metadata.service" Dec 13 14:09:41.465076 ignition[858]: INFO : files: op(7): op(8): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 13 14:09:41.465076 ignition[858]: INFO : files: op(7): op(8): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 13 14:09:41.465076 ignition[858]: INFO : files: op(7): [finished] processing unit "coreos-metadata.service" Dec 13 14:09:41.465076 ignition[858]: INFO : files: op(9): [started] setting preset to disabled for "coreos-metadata.service" Dec 13 14:09:41.465076 ignition[858]: INFO : files: op(9): op(a): [started] removing enablement symlink(s) for "coreos-metadata.service" Dec 13 14:09:41.502315 ignition[858]: INFO : files: op(9): op(a): [finished] removing enablement symlink(s) for "coreos-metadata.service" Dec 13 14:09:41.503454 ignition[858]: INFO : files: op(9): [finished] setting preset to disabled for "coreos-metadata.service" Dec 13 14:09:41.503454 ignition[858]: INFO : files: createResultFile: createFiles: op(b): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 13 14:09:41.503454 ignition[858]: INFO : files: createResultFile: createFiles: op(b): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 13 14:09:41.503454 ignition[858]: INFO : files: files passed Dec 13 14:09:41.503454 ignition[858]: INFO : Ignition finished successfully Dec 13 14:09:41.504000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.503821 systemd[1]: Finished ignition-files.service. Dec 13 14:09:41.505859 systemd[1]: Starting initrd-setup-root-after-ignition.service... Dec 13 14:09:41.506749 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Dec 13 14:09:41.510000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.510000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.513111 initrd-setup-root-after-ignition[883]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory Dec 13 14:09:41.507483 systemd[1]: Starting ignition-quench.service... Dec 13 14:09:41.513000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.515296 initrd-setup-root-after-ignition[885]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 14:09:41.510406 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 13 14:09:41.510491 systemd[1]: Finished ignition-quench.service. Dec 13 14:09:41.513781 systemd[1]: Finished initrd-setup-root-after-ignition.service. Dec 13 14:09:41.514798 systemd[1]: Reached target ignition-complete.target. Dec 13 14:09:41.516434 systemd[1]: Starting initrd-parse-etc.service... Dec 13 14:09:41.528425 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 13 14:09:41.528520 systemd[1]: Finished initrd-parse-etc.service. Dec 13 14:09:41.528000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.528000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.529741 systemd[1]: Reached target initrd-fs.target. Dec 13 14:09:41.530639 systemd[1]: Reached target initrd.target. Dec 13 14:09:41.531645 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Dec 13 14:09:41.532385 systemd[1]: Starting dracut-pre-pivot.service... Dec 13 14:09:41.542229 systemd[1]: Finished dracut-pre-pivot.service. Dec 13 14:09:41.541000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.543522 systemd[1]: Starting initrd-cleanup.service... Dec 13 14:09:41.551268 systemd[1]: Stopped target nss-lookup.target. Dec 13 14:09:41.552081 systemd[1]: Stopped target remote-cryptsetup.target. Dec 13 14:09:41.553219 systemd[1]: Stopped target timers.target. Dec 13 14:09:41.554218 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 13 14:09:41.554000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.554324 systemd[1]: Stopped dracut-pre-pivot.service. Dec 13 14:09:41.555378 systemd[1]: Stopped target initrd.target. Dec 13 14:09:41.556433 systemd[1]: Stopped target basic.target. Dec 13 14:09:41.557476 systemd[1]: Stopped target ignition-complete.target. Dec 13 14:09:41.558595 systemd[1]: Stopped target ignition-diskful.target. Dec 13 14:09:41.559661 systemd[1]: Stopped target initrd-root-device.target. Dec 13 14:09:41.560725 systemd[1]: Stopped target remote-fs.target. Dec 13 14:09:41.561716 systemd[1]: Stopped target remote-fs-pre.target. Dec 13 14:09:41.562788 systemd[1]: Stopped target sysinit.target. Dec 13 14:09:41.563719 systemd[1]: Stopped target local-fs.target. Dec 13 14:09:41.564671 systemd[1]: Stopped target local-fs-pre.target. Dec 13 14:09:41.565722 systemd[1]: Stopped target swap.target. Dec 13 14:09:41.566000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.566602 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 13 14:09:41.566703 systemd[1]: Stopped dracut-pre-mount.service. Dec 13 14:09:41.570000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.567733 systemd[1]: Stopped target cryptsetup.target. Dec 13 14:09:41.571000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.570031 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 13 14:09:41.570131 systemd[1]: Stopped dracut-initqueue.service. Dec 13 14:09:41.571213 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 13 14:09:41.571305 systemd[1]: Stopped ignition-fetch-offline.service. Dec 13 14:09:41.572389 systemd[1]: Stopped target paths.target. Dec 13 14:09:41.573386 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 13 14:09:41.576942 systemd[1]: Stopped systemd-ask-password-console.path. Dec 13 14:09:41.578239 systemd[1]: Stopped target slices.target. Dec 13 14:09:41.578817 systemd[1]: Stopped target sockets.target. Dec 13 14:09:41.579747 systemd[1]: iscsid.socket: Deactivated successfully. Dec 13 14:09:41.579815 systemd[1]: Closed iscsid.socket. Dec 13 14:09:41.580000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.580774 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 13 14:09:41.581000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.580870 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Dec 13 14:09:41.581900 systemd[1]: ignition-files.service: Deactivated successfully. Dec 13 14:09:41.582022 systemd[1]: Stopped ignition-files.service. Dec 13 14:09:41.583549 systemd[1]: Stopping ignition-mount.service... Dec 13 14:09:41.584735 systemd[1]: Stopping iscsiuio.service... Dec 13 14:09:41.586957 systemd[1]: Stopping sysroot-boot.service... Dec 13 14:09:41.588000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.587827 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 13 14:09:41.589000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.587971 systemd[1]: Stopped systemd-udev-trigger.service. Dec 13 14:09:41.589092 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 13 14:09:41.589175 systemd[1]: Stopped dracut-pre-trigger.service. Dec 13 14:09:41.592755 systemd[1]: iscsiuio.service: Deactivated successfully. Dec 13 14:09:41.592900 systemd[1]: Stopped iscsiuio.service. Dec 13 14:09:41.593000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.594553 ignition[898]: INFO : Ignition 2.14.0 Dec 13 14:09:41.594553 ignition[898]: INFO : Stage: umount Dec 13 14:09:41.594553 ignition[898]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 14:09:41.594553 ignition[898]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:09:41.594553 ignition[898]: INFO : umount: umount passed Dec 13 14:09:41.594553 ignition[898]: INFO : Ignition finished successfully Dec 13 14:09:41.596000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.598000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.599000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.596112 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 13 14:09:41.596416 systemd[1]: Stopped ignition-mount.service. Dec 13 14:09:41.597990 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 13 14:09:41.598514 systemd[1]: Finished initrd-cleanup.service. Dec 13 14:09:41.603000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.604000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.600995 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 13 14:09:41.606000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.601363 systemd[1]: Stopped target network.target. Dec 13 14:09:41.603029 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 13 14:09:41.603062 systemd[1]: Closed iscsiuio.socket. Dec 13 14:09:41.603578 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 13 14:09:41.603617 systemd[1]: Stopped ignition-disks.service. Dec 13 14:09:41.604723 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 13 14:09:41.604759 systemd[1]: Stopped ignition-kargs.service. Dec 13 14:09:41.605793 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 13 14:09:41.605825 systemd[1]: Stopped ignition-setup.service. Dec 13 14:09:41.607597 systemd[1]: Stopping systemd-networkd.service... Dec 13 14:09:41.608703 systemd[1]: Stopping systemd-resolved.service... Dec 13 14:09:41.616963 systemd-networkd[741]: eth0: DHCPv6 lease lost Dec 13 14:09:41.618602 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 13 14:09:41.618712 systemd[1]: Stopped systemd-networkd.service. Dec 13 14:09:41.619000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.620921 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 13 14:09:41.620000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.621022 systemd[1]: Stopped systemd-resolved.service. Dec 13 14:09:41.621825 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 13 14:09:41.621852 systemd[1]: Closed systemd-networkd.socket. Dec 13 14:09:41.625000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.626000 audit: BPF prog-id=6 op=UNLOAD Dec 13 14:09:41.623491 systemd[1]: Stopping network-cleanup.service... Dec 13 14:09:41.626000 audit: BPF prog-id=9 op=UNLOAD Dec 13 14:09:41.626000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.624206 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 13 14:09:41.624260 systemd[1]: Stopped parse-ip-for-networkd.service. Dec 13 14:09:41.628000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.625775 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 14:09:41.625814 systemd[1]: Stopped systemd-sysctl.service. Dec 13 14:09:41.627541 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 13 14:09:41.627579 systemd[1]: Stopped systemd-modules-load.service. Dec 13 14:09:41.628417 systemd[1]: Stopping systemd-udevd.service... Dec 13 14:09:41.635554 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Dec 13 14:09:41.635000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.636039 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 13 14:09:41.637000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.636119 systemd[1]: Stopped sysroot-boot.service. Dec 13 14:09:41.639000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.637328 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 13 14:09:41.637370 systemd[1]: Stopped initrd-setup-root.service. Dec 13 14:09:41.641000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.639051 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 13 14:09:41.639133 systemd[1]: Stopped network-cleanup.service. Dec 13 14:09:41.640716 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 13 14:09:41.644000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.640815 systemd[1]: Stopped systemd-udevd.service. Dec 13 14:09:41.645000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.642217 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 13 14:09:41.646000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.642246 systemd[1]: Closed systemd-udevd-control.socket. Dec 13 14:09:41.643048 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 13 14:09:41.648000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.643077 systemd[1]: Closed systemd-udevd-kernel.socket. Dec 13 14:09:41.650000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.644197 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 13 14:09:41.651000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.644235 systemd[1]: Stopped dracut-pre-udev.service. Dec 13 14:09:41.645217 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 13 14:09:41.653000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.653000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.645248 systemd[1]: Stopped dracut-cmdline.service. Dec 13 14:09:41.646364 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 14:09:41.646398 systemd[1]: Stopped dracut-cmdline-ask.service. Dec 13 14:09:41.648031 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Dec 13 14:09:41.648603 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 13 14:09:41.648650 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Dec 13 14:09:41.650345 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 13 14:09:41.650378 systemd[1]: Stopped kmod-static-nodes.service. Dec 13 14:09:41.650988 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 14:09:41.651022 systemd[1]: Stopped systemd-vconsole-setup.service. Dec 13 14:09:41.652729 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Dec 13 14:09:41.653204 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 13 14:09:41.653275 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Dec 13 14:09:41.654005 systemd[1]: Reached target initrd-switch-root.target. Dec 13 14:09:41.655712 systemd[1]: Starting initrd-switch-root.service... Dec 13 14:09:41.661878 systemd[1]: Switching root. Dec 13 14:09:41.677677 iscsid[752]: iscsid shutting down. Dec 13 14:09:41.678236 systemd-journald[289]: Received SIGTERM from PID 1 (n/a). Dec 13 14:09:41.678277 systemd-journald[289]: Journal stopped Dec 13 14:09:43.662315 kernel: SELinux: Class mctp_socket not defined in policy. Dec 13 14:09:43.662369 kernel: SELinux: Class anon_inode not defined in policy. Dec 13 14:09:43.662380 kernel: SELinux: the above unknown classes and permissions will be allowed Dec 13 14:09:43.662393 kernel: SELinux: policy capability network_peer_controls=1 Dec 13 14:09:43.662407 kernel: SELinux: policy capability open_perms=1 Dec 13 14:09:43.662421 kernel: SELinux: policy capability extended_socket_class=1 Dec 13 14:09:43.662430 kernel: SELinux: policy capability always_check_network=0 Dec 13 14:09:43.662444 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 13 14:09:43.662454 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 13 14:09:43.662464 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 13 14:09:43.662473 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 13 14:09:43.662484 systemd[1]: Successfully loaded SELinux policy in 42.897ms. Dec 13 14:09:43.662503 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 6.929ms. Dec 13 14:09:43.662515 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Dec 13 14:09:43.662525 systemd[1]: Detected virtualization kvm. Dec 13 14:09:43.662536 systemd[1]: Detected architecture arm64. Dec 13 14:09:43.662547 systemd[1]: Detected first boot. Dec 13 14:09:43.662558 systemd[1]: Initializing machine ID from VM UUID. Dec 13 14:09:43.662569 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Dec 13 14:09:43.662578 systemd[1]: Populated /etc with preset unit settings. Dec 13 14:09:43.662589 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Dec 13 14:09:43.662600 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Dec 13 14:09:43.662612 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 14:09:43.662623 kernel: kauditd_printk_skb: 80 callbacks suppressed Dec 13 14:09:43.662634 kernel: audit: type=1334 audit(1734098983.545:84): prog-id=12 op=LOAD Dec 13 14:09:43.662644 kernel: audit: type=1334 audit(1734098983.545:85): prog-id=3 op=UNLOAD Dec 13 14:09:43.662654 kernel: audit: type=1334 audit(1734098983.545:86): prog-id=13 op=LOAD Dec 13 14:09:43.662664 kernel: audit: type=1334 audit(1734098983.546:87): prog-id=14 op=LOAD Dec 13 14:09:43.662673 kernel: audit: type=1334 audit(1734098983.546:88): prog-id=4 op=UNLOAD Dec 13 14:09:43.662682 kernel: audit: type=1334 audit(1734098983.546:89): prog-id=5 op=UNLOAD Dec 13 14:09:43.662691 kernel: audit: type=1334 audit(1734098983.547:90): prog-id=15 op=LOAD Dec 13 14:09:43.662701 kernel: audit: type=1334 audit(1734098983.547:91): prog-id=12 op=UNLOAD Dec 13 14:09:43.662712 systemd[1]: iscsid.service: Deactivated successfully. Dec 13 14:09:43.662723 kernel: audit: type=1334 audit(1734098983.547:92): prog-id=16 op=LOAD Dec 13 14:09:43.662732 kernel: audit: type=1334 audit(1734098983.548:93): prog-id=17 op=LOAD Dec 13 14:09:43.662743 systemd[1]: Stopped iscsid.service. Dec 13 14:09:43.662753 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 13 14:09:43.662764 systemd[1]: Stopped initrd-switch-root.service. Dec 13 14:09:43.662774 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 13 14:09:43.662785 systemd[1]: Created slice system-addon\x2dconfig.slice. Dec 13 14:09:43.662796 systemd[1]: Created slice system-addon\x2drun.slice. Dec 13 14:09:43.662807 systemd[1]: Created slice system-getty.slice. Dec 13 14:09:43.662818 systemd[1]: Created slice system-modprobe.slice. Dec 13 14:09:43.662828 systemd[1]: Created slice system-serial\x2dgetty.slice. Dec 13 14:09:43.662839 systemd[1]: Created slice system-system\x2dcloudinit.slice. Dec 13 14:09:43.662850 systemd[1]: Created slice system-systemd\x2dfsck.slice. Dec 13 14:09:43.662861 systemd[1]: Created slice user.slice. Dec 13 14:09:43.662871 systemd[1]: Started systemd-ask-password-console.path. Dec 13 14:09:43.662898 systemd[1]: Started systemd-ask-password-wall.path. Dec 13 14:09:43.662912 systemd[1]: Set up automount boot.automount. Dec 13 14:09:43.662935 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Dec 13 14:09:43.662948 systemd[1]: Stopped target initrd-switch-root.target. Dec 13 14:09:43.662960 systemd[1]: Stopped target initrd-fs.target. Dec 13 14:09:43.662970 systemd[1]: Stopped target initrd-root-fs.target. Dec 13 14:09:43.662980 systemd[1]: Reached target integritysetup.target. Dec 13 14:09:43.662990 systemd[1]: Reached target remote-cryptsetup.target. Dec 13 14:09:43.663002 systemd[1]: Reached target remote-fs.target. Dec 13 14:09:43.663012 systemd[1]: Reached target slices.target. Dec 13 14:09:43.663022 systemd[1]: Reached target swap.target. Dec 13 14:09:43.663032 systemd[1]: Reached target torcx.target. Dec 13 14:09:43.663042 systemd[1]: Reached target veritysetup.target. Dec 13 14:09:43.663054 systemd[1]: Listening on systemd-coredump.socket. Dec 13 14:09:43.663064 systemd[1]: Listening on systemd-initctl.socket. Dec 13 14:09:43.663074 systemd[1]: Listening on systemd-networkd.socket. Dec 13 14:09:43.663084 systemd[1]: Listening on systemd-udevd-control.socket. Dec 13 14:09:43.663096 systemd[1]: Listening on systemd-udevd-kernel.socket. Dec 13 14:09:43.663108 systemd[1]: Listening on systemd-userdbd.socket. Dec 13 14:09:43.663118 systemd[1]: Mounting dev-hugepages.mount... Dec 13 14:09:43.663129 systemd[1]: Mounting dev-mqueue.mount... Dec 13 14:09:43.663139 systemd[1]: Mounting media.mount... Dec 13 14:09:43.663149 systemd[1]: Mounting sys-kernel-debug.mount... Dec 13 14:09:43.663159 systemd[1]: Mounting sys-kernel-tracing.mount... Dec 13 14:09:43.663169 systemd[1]: Mounting tmp.mount... Dec 13 14:09:43.663180 systemd[1]: Starting flatcar-tmpfiles.service... Dec 13 14:09:43.663190 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:09:43.663202 systemd[1]: Starting kmod-static-nodes.service... Dec 13 14:09:43.663212 systemd[1]: Starting modprobe@configfs.service... Dec 13 14:09:43.663223 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:09:43.663233 systemd[1]: Starting modprobe@drm.service... Dec 13 14:09:43.663243 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:09:43.663253 systemd[1]: Starting modprobe@fuse.service... Dec 13 14:09:43.663263 systemd[1]: Starting modprobe@loop.service... Dec 13 14:09:43.663274 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 13 14:09:43.663286 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 13 14:09:43.663297 systemd[1]: Stopped systemd-fsck-root.service. Dec 13 14:09:43.663307 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 13 14:09:43.663318 systemd[1]: Stopped systemd-fsck-usr.service. Dec 13 14:09:43.663329 systemd[1]: Stopped systemd-journald.service. Dec 13 14:09:43.663339 kernel: fuse: init (API version 7.34) Dec 13 14:09:43.663348 systemd[1]: Starting systemd-journald.service... Dec 13 14:09:43.663358 systemd[1]: Starting systemd-modules-load.service... Dec 13 14:09:43.663368 kernel: loop: module loaded Dec 13 14:09:43.663378 systemd[1]: Starting systemd-network-generator.service... Dec 13 14:09:43.663390 systemd[1]: Starting systemd-remount-fs.service... Dec 13 14:09:43.663400 systemd[1]: Starting systemd-udev-trigger.service... Dec 13 14:09:43.663411 systemd[1]: verity-setup.service: Deactivated successfully. Dec 13 14:09:43.663421 systemd[1]: Stopped verity-setup.service. Dec 13 14:09:43.663431 systemd[1]: Mounted dev-hugepages.mount. Dec 13 14:09:43.663441 systemd[1]: Mounted dev-mqueue.mount. Dec 13 14:09:43.663451 systemd[1]: Mounted media.mount. Dec 13 14:09:43.663461 systemd[1]: Mounted sys-kernel-debug.mount. Dec 13 14:09:43.663471 systemd[1]: Mounted sys-kernel-tracing.mount. Dec 13 14:09:43.663483 systemd[1]: Mounted tmp.mount. Dec 13 14:09:43.663493 systemd[1]: Finished kmod-static-nodes.service. Dec 13 14:09:43.663523 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 13 14:09:43.663534 systemd[1]: Finished modprobe@configfs.service. Dec 13 14:09:43.663547 systemd-journald[992]: Journal started Dec 13 14:09:43.663591 systemd-journald[992]: Runtime Journal (/run/log/journal/63af08d6234f452d88df29a5cb10701d) is 6.0M, max 48.7M, 42.6M free. Dec 13 14:09:41.756000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 13 14:09:41.826000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Dec 13 14:09:41.827000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Dec 13 14:09:41.827000 audit: BPF prog-id=10 op=LOAD Dec 13 14:09:41.827000 audit: BPF prog-id=10 op=UNLOAD Dec 13 14:09:41.827000 audit: BPF prog-id=11 op=LOAD Dec 13 14:09:41.827000 audit: BPF prog-id=11 op=UNLOAD Dec 13 14:09:41.882000 audit[931]: AVC avc: denied { associate } for pid=931 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Dec 13 14:09:41.882000 audit[931]: SYSCALL arch=c00000b7 syscall=5 success=yes exit=0 a0=40001c58ac a1=40000c8de0 a2=40000cf0c0 a3=32 items=0 ppid=914 pid=931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:09:41.882000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Dec 13 14:09:41.883000 audit[931]: AVC avc: denied { associate } for pid=931 comm="torcx-generator" name="lib" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Dec 13 14:09:41.883000 audit[931]: SYSCALL arch=c00000b7 syscall=34 success=yes exit=0 a0=ffffffffffffff9c a1=40001c5985 a2=1ed a3=0 items=2 ppid=914 pid=931 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:09:41.883000 audit: CWD cwd="/" Dec 13 14:09:41.883000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:09:41.883000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:09:41.883000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Dec 13 14:09:43.545000 audit: BPF prog-id=12 op=LOAD Dec 13 14:09:43.545000 audit: BPF prog-id=3 op=UNLOAD Dec 13 14:09:43.545000 audit: BPF prog-id=13 op=LOAD Dec 13 14:09:43.546000 audit: BPF prog-id=14 op=LOAD Dec 13 14:09:43.546000 audit: BPF prog-id=4 op=UNLOAD Dec 13 14:09:43.546000 audit: BPF prog-id=5 op=UNLOAD Dec 13 14:09:43.547000 audit: BPF prog-id=15 op=LOAD Dec 13 14:09:43.547000 audit: BPF prog-id=12 op=UNLOAD Dec 13 14:09:43.547000 audit: BPF prog-id=16 op=LOAD Dec 13 14:09:43.548000 audit: BPF prog-id=17 op=LOAD Dec 13 14:09:43.548000 audit: BPF prog-id=13 op=UNLOAD Dec 13 14:09:43.548000 audit: BPF prog-id=14 op=UNLOAD Dec 13 14:09:43.549000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.553000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.555000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.555000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.565000 audit: BPF prog-id=15 op=UNLOAD Dec 13 14:09:43.629000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.631000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.632000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.632000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.635000 audit: BPF prog-id=18 op=LOAD Dec 13 14:09:43.635000 audit: BPF prog-id=19 op=LOAD Dec 13 14:09:43.635000 audit: BPF prog-id=20 op=LOAD Dec 13 14:09:43.635000 audit: BPF prog-id=16 op=UNLOAD Dec 13 14:09:43.635000 audit: BPF prog-id=17 op=UNLOAD Dec 13 14:09:43.648000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.661000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.661000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Dec 13 14:09:43.661000 audit[992]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=6 a1=ffffdfd60230 a2=4000 a3=1 items=0 ppid=1 pid=992 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:09:43.661000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Dec 13 14:09:43.662000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.662000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.544764 systemd[1]: Queued start job for default target multi-user.target. Dec 13 14:09:41.881674 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.6 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.6 /var/lib/torcx/store]" Dec 13 14:09:43.544776 systemd[1]: Unnecessary job was removed for dev-vda6.device. Dec 13 14:09:43.664990 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:09:43.665008 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:09:41.881966 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Dec 13 14:09:43.550076 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 13 14:09:41.881985 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Dec 13 14:09:41.882016 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Dec 13 14:09:41.882025 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="skipped missing lower profile" missing profile=oem Dec 13 14:09:41.882055 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Dec 13 14:09:41.882067 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Dec 13 14:09:41.882263 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Dec 13 14:09:41.882296 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Dec 13 14:09:43.664000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.664000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:41.882307 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Dec 13 14:09:41.882774 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Dec 13 14:09:41.882810 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Dec 13 14:09:41.882827 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.6: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.6 Dec 13 14:09:41.882841 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Dec 13 14:09:41.882857 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.6: no such file or directory" path=/var/lib/torcx/store/3510.3.6 Dec 13 14:09:41.882870 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:41Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Dec 13 14:09:43.666297 systemd[1]: Started systemd-journald.service. Dec 13 14:09:43.308735 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:43Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:09:43.309017 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:43Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:09:43.309125 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:43Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:09:43.309350 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:43Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:09:43.309402 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:43Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Dec 13 14:09:43.309456 /usr/lib/systemd/system-generators/torcx-generator[931]: time="2024-12-13T14:09:43Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Dec 13 14:09:43.666000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.667468 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 14:09:43.667617 systemd[1]: Finished modprobe@drm.service. Dec 13 14:09:43.668000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.668000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.668444 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:09:43.668586 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:09:43.669000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.669000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.669470 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 13 14:09:43.669620 systemd[1]: Finished modprobe@fuse.service. Dec 13 14:09:43.670000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.670000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.670573 systemd[1]: Finished flatcar-tmpfiles.service. Dec 13 14:09:43.671000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.671405 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:09:43.671550 systemd[1]: Finished modprobe@loop.service. Dec 13 14:09:43.672000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.672000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.672500 systemd[1]: Finished systemd-modules-load.service. Dec 13 14:09:43.673000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.673720 systemd[1]: Finished systemd-network-generator.service. Dec 13 14:09:43.674000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.674766 systemd[1]: Finished systemd-remount-fs.service. Dec 13 14:09:43.675000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.675898 systemd[1]: Reached target network-pre.target. Dec 13 14:09:43.677767 systemd[1]: Mounting sys-fs-fuse-connections.mount... Dec 13 14:09:43.679551 systemd[1]: Mounting sys-kernel-config.mount... Dec 13 14:09:43.680162 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 13 14:09:43.681624 systemd[1]: Starting systemd-hwdb-update.service... Dec 13 14:09:43.683484 systemd[1]: Starting systemd-journal-flush.service... Dec 13 14:09:43.684264 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:09:43.685307 systemd[1]: Starting systemd-random-seed.service... Dec 13 14:09:43.686067 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:09:43.687090 systemd[1]: Starting systemd-sysctl.service... Dec 13 14:09:43.688981 systemd[1]: Starting systemd-sysusers.service... Dec 13 14:09:43.691218 systemd-journald[992]: Time spent on flushing to /var/log/journal/63af08d6234f452d88df29a5cb10701d is 19.178ms for 982 entries. Dec 13 14:09:43.691218 systemd-journald[992]: System Journal (/var/log/journal/63af08d6234f452d88df29a5cb10701d) is 8.0M, max 195.6M, 187.6M free. Dec 13 14:09:43.722374 systemd-journald[992]: Received client request to flush runtime journal. Dec 13 14:09:43.699000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.700000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.705000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.714000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.692517 systemd[1]: Mounted sys-fs-fuse-connections.mount. Dec 13 14:09:43.693443 systemd[1]: Mounted sys-kernel-config.mount. Dec 13 14:09:43.723558 udevadm[1032]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Dec 13 14:09:43.698483 systemd[1]: Finished systemd-udev-trigger.service. Dec 13 14:09:43.699465 systemd[1]: Finished systemd-random-seed.service. Dec 13 14:09:43.700177 systemd[1]: Reached target first-boot-complete.target. Dec 13 14:09:43.701738 systemd[1]: Starting systemd-udev-settle.service... Dec 13 14:09:43.706208 systemd[1]: Finished systemd-sysctl.service. Dec 13 14:09:43.713761 systemd[1]: Finished systemd-sysusers.service. Dec 13 14:09:43.723000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:43.715435 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Dec 13 14:09:43.723287 systemd[1]: Finished systemd-journal-flush.service. Dec 13 14:09:43.732206 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Dec 13 14:09:43.732000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.046932 systemd[1]: Finished systemd-hwdb-update.service. Dec 13 14:09:44.046000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.047000 audit: BPF prog-id=21 op=LOAD Dec 13 14:09:44.047000 audit: BPF prog-id=22 op=LOAD Dec 13 14:09:44.047000 audit: BPF prog-id=7 op=UNLOAD Dec 13 14:09:44.047000 audit: BPF prog-id=8 op=UNLOAD Dec 13 14:09:44.048860 systemd[1]: Starting systemd-udevd.service... Dec 13 14:09:44.075353 systemd-udevd[1036]: Using default interface naming scheme 'v252'. Dec 13 14:09:44.086656 systemd[1]: Started systemd-udevd.service. Dec 13 14:09:44.086000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.087000 audit: BPF prog-id=23 op=LOAD Dec 13 14:09:44.089457 systemd[1]: Starting systemd-networkd.service... Dec 13 14:09:44.096000 audit: BPF prog-id=24 op=LOAD Dec 13 14:09:44.096000 audit: BPF prog-id=25 op=LOAD Dec 13 14:09:44.096000 audit: BPF prog-id=26 op=LOAD Dec 13 14:09:44.097896 systemd[1]: Starting systemd-userdbd.service... Dec 13 14:09:44.105304 systemd[1]: Condition check resulted in dev-ttyAMA0.device being skipped. Dec 13 14:09:44.128550 systemd[1]: Started systemd-userdbd.service. Dec 13 14:09:44.128000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.154982 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Dec 13 14:09:44.190281 systemd-networkd[1044]: lo: Link UP Dec 13 14:09:44.190539 systemd-networkd[1044]: lo: Gained carrier Dec 13 14:09:44.190974 systemd-networkd[1044]: Enumeration completed Dec 13 14:09:44.191183 systemd-networkd[1044]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 14:09:44.191194 systemd[1]: Started systemd-networkd.service. Dec 13 14:09:44.190000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.195698 systemd[1]: Finished systemd-udev-settle.service. Dec 13 14:09:44.196107 systemd-networkd[1044]: eth0: Link UP Dec 13 14:09:44.196184 systemd-networkd[1044]: eth0: Gained carrier Dec 13 14:09:44.195000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.197440 systemd[1]: Starting lvm2-activation-early.service... Dec 13 14:09:44.207245 lvm[1069]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 14:09:44.222017 systemd-networkd[1044]: eth0: DHCPv4 address 10.0.0.79/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 14:09:44.239685 systemd[1]: Finished lvm2-activation-early.service. Dec 13 14:09:44.239000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.240484 systemd[1]: Reached target cryptsetup.target. Dec 13 14:09:44.242157 systemd[1]: Starting lvm2-activation.service... Dec 13 14:09:44.245509 lvm[1070]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 14:09:44.280669 systemd[1]: Finished lvm2-activation.service. Dec 13 14:09:44.280000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.281414 systemd[1]: Reached target local-fs-pre.target. Dec 13 14:09:44.282066 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 13 14:09:44.282097 systemd[1]: Reached target local-fs.target. Dec 13 14:09:44.282643 systemd[1]: Reached target machines.target. Dec 13 14:09:44.284277 systemd[1]: Starting ldconfig.service... Dec 13 14:09:44.285164 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.285215 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.286301 systemd[1]: Starting systemd-boot-update.service... Dec 13 14:09:44.287971 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Dec 13 14:09:44.289789 systemd[1]: Starting systemd-machine-id-commit.service... Dec 13 14:09:44.291526 systemd[1]: Starting systemd-sysext.service... Dec 13 14:09:44.292806 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1072 (bootctl) Dec 13 14:09:44.294122 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Dec 13 14:09:44.299233 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Dec 13 14:09:44.299000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.301577 systemd[1]: Unmounting usr-share-oem.mount... Dec 13 14:09:44.305298 systemd[1]: usr-share-oem.mount: Deactivated successfully. Dec 13 14:09:44.305477 systemd[1]: Unmounted usr-share-oem.mount. Dec 13 14:09:44.318915 kernel: loop0: detected capacity change from 0 to 194096 Dec 13 14:09:44.358046 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 13 14:09:44.358613 systemd[1]: Finished systemd-machine-id-commit.service. Dec 13 14:09:44.361000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.368907 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 13 14:09:44.382383 systemd-fsck[1083]: fsck.fat 4.2 (2021-01-31) Dec 13 14:09:44.382383 systemd-fsck[1083]: /dev/vda1: 236 files, 117175/258078 clusters Dec 13 14:09:44.388955 kernel: loop1: detected capacity change from 0 to 194096 Dec 13 14:09:44.385000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.384167 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Dec 13 14:09:44.390044 systemd[1]: Mounting boot.mount... Dec 13 14:09:44.394695 (sd-sysext)[1086]: Using extensions 'kubernetes'. Dec 13 14:09:44.395062 (sd-sysext)[1086]: Merged extensions into '/usr'. Dec 13 14:09:44.402634 systemd[1]: Mounted boot.mount. Dec 13 14:09:44.413413 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.414756 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:09:44.416628 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:09:44.418391 systemd[1]: Starting modprobe@loop.service... Dec 13 14:09:44.419078 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.419210 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.420048 systemd[1]: Finished systemd-boot-update.service. Dec 13 14:09:44.420000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.421244 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:09:44.421360 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:09:44.421000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.421000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.422445 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:09:44.422552 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:09:44.422000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.422000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.423693 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:09:44.423792 systemd[1]: Finished modprobe@loop.service. Dec 13 14:09:44.423000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.423000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.425058 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:09:44.425163 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.474582 ldconfig[1071]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 13 14:09:44.478220 systemd[1]: Finished ldconfig.service. Dec 13 14:09:44.477000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.651564 systemd[1]: Mounting usr-share-oem.mount... Dec 13 14:09:44.656369 systemd[1]: Mounted usr-share-oem.mount. Dec 13 14:09:44.657966 systemd[1]: Finished systemd-sysext.service. Dec 13 14:09:44.657000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.659646 systemd[1]: Starting ensure-sysext.service... Dec 13 14:09:44.661161 systemd[1]: Starting systemd-tmpfiles-setup.service... Dec 13 14:09:44.665393 systemd[1]: Reloading. Dec 13 14:09:44.670490 systemd-tmpfiles[1094]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Dec 13 14:09:44.672434 systemd-tmpfiles[1094]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 13 14:09:44.673867 systemd-tmpfiles[1094]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 13 14:09:44.705864 /usr/lib/systemd/system-generators/torcx-generator[1114]: time="2024-12-13T14:09:44Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.6 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.6 /var/lib/torcx/store]" Dec 13 14:09:44.706328 /usr/lib/systemd/system-generators/torcx-generator[1114]: time="2024-12-13T14:09:44Z" level=info msg="torcx already run" Dec 13 14:09:44.756091 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Dec 13 14:09:44.756123 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Dec 13 14:09:44.771506 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 14:09:44.811000 audit: BPF prog-id=27 op=LOAD Dec 13 14:09:44.811000 audit: BPF prog-id=18 op=UNLOAD Dec 13 14:09:44.811000 audit: BPF prog-id=28 op=LOAD Dec 13 14:09:44.811000 audit: BPF prog-id=29 op=LOAD Dec 13 14:09:44.811000 audit: BPF prog-id=19 op=UNLOAD Dec 13 14:09:44.811000 audit: BPF prog-id=20 op=UNLOAD Dec 13 14:09:44.811000 audit: BPF prog-id=30 op=LOAD Dec 13 14:09:44.811000 audit: BPF prog-id=24 op=UNLOAD Dec 13 14:09:44.811000 audit: BPF prog-id=31 op=LOAD Dec 13 14:09:44.811000 audit: BPF prog-id=32 op=LOAD Dec 13 14:09:44.811000 audit: BPF prog-id=25 op=UNLOAD Dec 13 14:09:44.811000 audit: BPF prog-id=26 op=UNLOAD Dec 13 14:09:44.812000 audit: BPF prog-id=33 op=LOAD Dec 13 14:09:44.812000 audit: BPF prog-id=23 op=UNLOAD Dec 13 14:09:44.814000 audit: BPF prog-id=34 op=LOAD Dec 13 14:09:44.814000 audit: BPF prog-id=35 op=LOAD Dec 13 14:09:44.814000 audit: BPF prog-id=21 op=UNLOAD Dec 13 14:09:44.814000 audit: BPF prog-id=22 op=UNLOAD Dec 13 14:09:44.816978 systemd[1]: Finished systemd-tmpfiles-setup.service. Dec 13 14:09:44.816000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.821024 systemd[1]: Starting audit-rules.service... Dec 13 14:09:44.822721 systemd[1]: Starting clean-ca-certificates.service... Dec 13 14:09:44.824741 systemd[1]: Starting systemd-journal-catalog-update.service... Dec 13 14:09:44.825000 audit: BPF prog-id=36 op=LOAD Dec 13 14:09:44.830470 systemd[1]: Starting systemd-resolved.service... Dec 13 14:09:44.832000 audit: BPF prog-id=37 op=LOAD Dec 13 14:09:44.834756 systemd[1]: Starting systemd-timesyncd.service... Dec 13 14:09:44.836552 systemd[1]: Starting systemd-update-utmp.service... Dec 13 14:09:44.841316 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.842678 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:09:44.844452 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:09:44.844000 audit[1159]: SYSTEM_BOOT pid=1159 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.846310 systemd[1]: Starting modprobe@loop.service... Dec 13 14:09:44.846986 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.847120 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.848042 systemd[1]: Finished clean-ca-certificates.service. Dec 13 14:09:44.847000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.849069 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:09:44.849185 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:09:44.849000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.849000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.850296 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:09:44.850401 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:09:44.850000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.850000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.851411 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:09:44.851515 systemd[1]: Finished modprobe@loop.service. Dec 13 14:09:44.852000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.852000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.855150 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.856420 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:09:44.858153 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:09:44.860004 systemd[1]: Starting modprobe@loop.service... Dec 13 14:09:44.860707 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.860999 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.861214 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 14:09:44.862467 systemd[1]: Finished systemd-journal-catalog-update.service. Dec 13 14:09:44.863000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.863735 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:09:44.863845 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:09:44.865100 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:09:44.865210 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:09:44.864000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.864000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.865000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.865000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.866304 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:09:44.866412 systemd[1]: Finished modprobe@loop.service. Dec 13 14:09:44.866000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.866000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.867618 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:09:44.867750 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.870239 systemd[1]: Starting systemd-update-done.service... Dec 13 14:09:44.871640 systemd[1]: Finished systemd-update-utmp.service. Dec 13 14:09:44.872000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.875448 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.876607 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:09:44.878258 systemd[1]: Starting modprobe@drm.service... Dec 13 14:09:44.880061 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:09:44.881911 systemd[1]: Starting modprobe@loop.service... Dec 13 14:09:44.882591 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.882726 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.884063 systemd[1]: Starting systemd-networkd-wait-online.service... Dec 13 14:09:44.885231 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 14:09:44.886263 systemd[1]: Finished systemd-update-done.service. Dec 13 14:09:44.886000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.887337 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:09:44.887450 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:09:44.887000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.887000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.888475 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 14:09:44.888651 systemd[1]: Finished modprobe@drm.service. Dec 13 14:09:44.889666 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:09:44.889771 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:09:44.889000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.889000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.890000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.890000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.890000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.890000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:09:44.890796 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:09:44.890920 systemd[1]: Finished modprobe@loop.service. Dec 13 14:09:44.892015 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:09:44.892110 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.893672 systemd[1]: Started systemd-timesyncd.service. Dec 13 14:09:44.894116 augenrules[1184]: No rules Dec 13 14:09:44.892000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Dec 13 14:09:44.892000 audit[1184]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffc806d120 a2=420 a3=0 items=0 ppid=1153 pid=1184 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:09:44.892000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Dec 13 14:09:44.894830 systemd-timesyncd[1158]: Contacted time server 10.0.0.1:123 (10.0.0.1). Dec 13 14:09:44.895101 systemd[1]: Finished ensure-sysext.service. Dec 13 14:09:44.895226 systemd-timesyncd[1158]: Initial clock synchronization to Fri 2024-12-13 14:09:44.502462 UTC. Dec 13 14:09:44.895974 systemd[1]: Finished audit-rules.service. Dec 13 14:09:44.896705 systemd[1]: Reached target time-set.target. Dec 13 14:09:44.901734 systemd-resolved[1157]: Positive Trust Anchors: Dec 13 14:09:44.901744 systemd-resolved[1157]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 14:09:44.901773 systemd-resolved[1157]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Dec 13 14:09:44.910399 systemd-resolved[1157]: Defaulting to hostname 'linux'. Dec 13 14:09:44.911705 systemd[1]: Started systemd-resolved.service. Dec 13 14:09:44.912372 systemd[1]: Reached target network.target. Dec 13 14:09:44.912938 systemd[1]: Reached target nss-lookup.target. Dec 13 14:09:44.913495 systemd[1]: Reached target sysinit.target. Dec 13 14:09:44.914116 systemd[1]: Started motdgen.path. Dec 13 14:09:44.914623 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Dec 13 14:09:44.915598 systemd[1]: Started logrotate.timer. Dec 13 14:09:44.916245 systemd[1]: Started mdadm.timer. Dec 13 14:09:44.916729 systemd[1]: Started systemd-tmpfiles-clean.timer. Dec 13 14:09:44.917362 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 13 14:09:44.917393 systemd[1]: Reached target paths.target. Dec 13 14:09:44.917914 systemd[1]: Reached target timers.target. Dec 13 14:09:44.918827 systemd[1]: Listening on dbus.socket. Dec 13 14:09:44.920390 systemd[1]: Starting docker.socket... Dec 13 14:09:44.923363 systemd[1]: Listening on sshd.socket. Dec 13 14:09:44.924021 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.924427 systemd[1]: Listening on docker.socket. Dec 13 14:09:44.925072 systemd[1]: Reached target sockets.target. Dec 13 14:09:44.925616 systemd[1]: Reached target basic.target. Dec 13 14:09:44.926199 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.926227 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Dec 13 14:09:44.927186 systemd[1]: Starting containerd.service... Dec 13 14:09:44.928686 systemd[1]: Starting dbus.service... Dec 13 14:09:44.930337 systemd[1]: Starting enable-oem-cloudinit.service... Dec 13 14:09:44.932186 systemd[1]: Starting extend-filesystems.service... Dec 13 14:09:44.932933 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Dec 13 14:09:44.934254 systemd[1]: Starting motdgen.service... Dec 13 14:09:44.936020 systemd[1]: Starting ssh-key-proc-cmdline.service... Dec 13 14:09:44.938278 systemd[1]: Starting sshd-keygen.service... Dec 13 14:09:44.940905 systemd[1]: Starting systemd-logind.service... Dec 13 14:09:44.941524 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:09:44.941592 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 13 14:09:44.942190 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 13 14:09:44.942837 systemd[1]: Starting update-engine.service... Dec 13 14:09:44.944596 systemd[1]: Starting update-ssh-keys-after-ignition.service... Dec 13 14:09:44.948865 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 13 14:09:44.950876 jq[1208]: true Dec 13 14:09:44.951434 jq[1195]: false Dec 13 14:09:44.949064 systemd[1]: Finished ssh-key-proc-cmdline.service. Dec 13 14:09:44.954228 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 13 14:09:44.954437 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Dec 13 14:09:44.964608 extend-filesystems[1196]: Found loop1 Dec 13 14:09:44.965577 extend-filesystems[1196]: Found vda Dec 13 14:09:44.966232 extend-filesystems[1196]: Found vda1 Dec 13 14:09:44.966865 extend-filesystems[1196]: Found vda2 Dec 13 14:09:44.967737 extend-filesystems[1196]: Found vda3 Dec 13 14:09:44.967737 extend-filesystems[1196]: Found usr Dec 13 14:09:44.967737 extend-filesystems[1196]: Found vda4 Dec 13 14:09:44.967737 extend-filesystems[1196]: Found vda6 Dec 13 14:09:44.967737 extend-filesystems[1196]: Found vda7 Dec 13 14:09:44.967737 extend-filesystems[1196]: Found vda9 Dec 13 14:09:44.967737 extend-filesystems[1196]: Checking size of /dev/vda9 Dec 13 14:09:44.971813 jq[1213]: true Dec 13 14:09:44.971403 systemd[1]: motdgen.service: Deactivated successfully. Dec 13 14:09:44.971618 systemd[1]: Finished motdgen.service. Dec 13 14:09:44.981548 systemd[1]: Started dbus.service. Dec 13 14:09:44.981032 dbus-daemon[1194]: [system] SELinux support is enabled Dec 13 14:09:44.983740 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 13 14:09:44.983766 systemd[1]: Reached target system-config.target. Dec 13 14:09:44.984551 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 13 14:09:44.984573 systemd[1]: Reached target user-config.target. Dec 13 14:09:44.985506 extend-filesystems[1196]: Resized partition /dev/vda9 Dec 13 14:09:44.999030 extend-filesystems[1235]: resize2fs 1.46.5 (30-Dec-2021) Dec 13 14:09:45.003108 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Dec 13 14:09:45.023759 systemd-logind[1203]: Watching system buttons on /dev/input/event0 (Power Button) Dec 13 14:09:45.024350 systemd-logind[1203]: New seat seat0. Dec 13 14:09:45.027362 systemd[1]: Started systemd-logind.service. Dec 13 14:09:45.034504 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Dec 13 14:09:45.035467 update_engine[1204]: I1213 14:09:45.034400 1204 main.cc:92] Flatcar Update Engine starting Dec 13 14:09:45.056561 update_engine[1204]: I1213 14:09:45.040100 1204 update_check_scheduler.cc:74] Next update check in 8m6s Dec 13 14:09:45.040104 systemd[1]: Started update-engine.service. Dec 13 14:09:45.042417 systemd[1]: Started locksmithd.service. Dec 13 14:09:45.056929 extend-filesystems[1235]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Dec 13 14:09:45.056929 extend-filesystems[1235]: old_desc_blocks = 1, new_desc_blocks = 1 Dec 13 14:09:45.056929 extend-filesystems[1235]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Dec 13 14:09:45.060046 extend-filesystems[1196]: Resized filesystem in /dev/vda9 Dec 13 14:09:45.060705 bash[1243]: Updated "/home/core/.ssh/authorized_keys" Dec 13 14:09:45.060794 env[1212]: time="2024-12-13T14:09:45.057385472Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Dec 13 14:09:45.060312 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 13 14:09:45.060474 systemd[1]: Finished extend-filesystems.service. Dec 13 14:09:45.061674 systemd[1]: Finished update-ssh-keys-after-ignition.service. Dec 13 14:09:45.075633 env[1212]: time="2024-12-13T14:09:45.075589673Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 13 14:09:45.075971 env[1212]: time="2024-12-13T14:09:45.075945142Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:09:45.077152 env[1212]: time="2024-12-13T14:09:45.077100904Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.173-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 13 14:09:45.077152 env[1212]: time="2024-12-13T14:09:45.077140997Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:09:45.077440 env[1212]: time="2024-12-13T14:09:45.077411641Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 14:09:45.077440 env[1212]: time="2024-12-13T14:09:45.077433931Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 13 14:09:45.077503 env[1212]: time="2024-12-13T14:09:45.077447968Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Dec 13 14:09:45.077503 env[1212]: time="2024-12-13T14:09:45.077457325Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 13 14:09:45.077564 env[1212]: time="2024-12-13T14:09:45.077546754Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:09:45.078080 env[1212]: time="2024-12-13T14:09:45.078060311Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:09:45.078202 env[1212]: time="2024-12-13T14:09:45.078183176Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 14:09:45.078236 env[1212]: time="2024-12-13T14:09:45.078202119Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 13 14:09:45.078270 env[1212]: time="2024-12-13T14:09:45.078254726Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Dec 13 14:09:45.078297 env[1212]: time="2024-12-13T14:09:45.078271539Z" level=info msg="metadata content store policy set" policy=shared Dec 13 14:09:45.081550 env[1212]: time="2024-12-13T14:09:45.081521397Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 13 14:09:45.081603 env[1212]: time="2024-12-13T14:09:45.081555403Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 13 14:09:45.081603 env[1212]: time="2024-12-13T14:09:45.081567690Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 13 14:09:45.081640 env[1212]: time="2024-12-13T14:09:45.081603256Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.081640 env[1212]: time="2024-12-13T14:09:45.081617482Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.081640 env[1212]: time="2024-12-13T14:09:45.081629921Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.081711 env[1212]: time="2024-12-13T14:09:45.081641675Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.082056 env[1212]: time="2024-12-13T14:09:45.082029135Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.082094 env[1212]: time="2024-12-13T14:09:45.082057055Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.082094 env[1212]: time="2024-12-13T14:09:45.082070673Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.082094 env[1212]: time="2024-12-13T14:09:45.082082046Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.082156 env[1212]: time="2024-12-13T14:09:45.082094409Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 13 14:09:45.082220 env[1212]: time="2024-12-13T14:09:45.082198064Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 13 14:09:45.082292 env[1212]: time="2024-12-13T14:09:45.082275244Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 13 14:09:45.082531 env[1212]: time="2024-12-13T14:09:45.082510931Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 13 14:09:45.082571 env[1212]: time="2024-12-13T14:09:45.082546078Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082571 env[1212]: time="2024-12-13T14:09:45.082559620Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 13 14:09:45.082678 env[1212]: time="2024-12-13T14:09:45.082662020Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082715 env[1212]: time="2024-12-13T14:09:45.082683930Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082715 env[1212]: time="2024-12-13T14:09:45.082696482Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082779 env[1212]: time="2024-12-13T14:09:45.082762213Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082810 env[1212]: time="2024-12-13T14:09:45.082781384Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082810 env[1212]: time="2024-12-13T14:09:45.082794051Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082810 env[1212]: time="2024-12-13T14:09:45.082805577Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082875 env[1212]: time="2024-12-13T14:09:45.082817331Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.082875 env[1212]: time="2024-12-13T14:09:45.082829617Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 13 14:09:45.082980 env[1212]: time="2024-12-13T14:09:45.082959328Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.083008 env[1212]: time="2024-12-13T14:09:45.082982037Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.083008 env[1212]: time="2024-12-13T14:09:45.082995313Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.083008 env[1212]: time="2024-12-13T14:09:45.083005850Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 13 14:09:45.083059 env[1212]: time="2024-12-13T14:09:45.083018364Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Dec 13 14:09:45.083059 env[1212]: time="2024-12-13T14:09:45.083029015Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 13 14:09:45.083059 env[1212]: time="2024-12-13T14:09:45.083044078Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Dec 13 14:09:45.083118 env[1212]: time="2024-12-13T14:09:45.083074357Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 13 14:09:45.083423 env[1212]: time="2024-12-13T14:09:45.083352038Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 13 14:09:45.083423 env[1212]: time="2024-12-13T14:09:45.083409134Z" level=info msg="Connect containerd service" Dec 13 14:09:45.084052 env[1212]: time="2024-12-13T14:09:45.083435380Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 13 14:09:45.084243 env[1212]: time="2024-12-13T14:09:45.084192271Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 14:09:45.084998 env[1212]: time="2024-12-13T14:09:45.084942542Z" level=info msg="Start subscribing containerd event" Dec 13 14:09:45.085655 env[1212]: time="2024-12-13T14:09:45.085122388Z" level=info msg="Start recovering state" Dec 13 14:09:45.085655 env[1212]: time="2024-12-13T14:09:45.085187434Z" level=info msg="Start event monitor" Dec 13 14:09:45.085655 env[1212]: time="2024-12-13T14:09:45.085206491Z" level=info msg="Start snapshots syncer" Dec 13 14:09:45.085655 env[1212]: time="2024-12-13T14:09:45.085219120Z" level=info msg="Start cni network conf syncer for default" Dec 13 14:09:45.085655 env[1212]: time="2024-12-13T14:09:45.085226309Z" level=info msg="Start streaming server" Dec 13 14:09:45.085784 env[1212]: time="2024-12-13T14:09:45.085569455Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 13 14:09:45.085866 env[1212]: time="2024-12-13T14:09:45.085836865Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 13 14:09:45.086826 locksmithd[1244]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 13 14:09:45.087988 env[1212]: time="2024-12-13T14:09:45.087963027Z" level=info msg="containerd successfully booted in 0.035774s" Dec 13 14:09:45.088013 systemd[1]: Started containerd.service. Dec 13 14:09:45.941102 systemd-networkd[1044]: eth0: Gained IPv6LL Dec 13 14:09:45.942918 systemd[1]: Finished systemd-networkd-wait-online.service. Dec 13 14:09:45.943849 systemd[1]: Reached target network-online.target. Dec 13 14:09:45.945858 systemd[1]: Starting kubelet.service... Dec 13 14:09:46.437866 systemd[1]: Started kubelet.service. Dec 13 14:09:46.943145 kubelet[1258]: E1213 14:09:46.943093 1258 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 14:09:46.944986 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 14:09:46.945126 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 14:09:47.631431 sshd_keygen[1211]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 13 14:09:47.647942 systemd[1]: Finished sshd-keygen.service. Dec 13 14:09:47.649944 systemd[1]: Starting issuegen.service... Dec 13 14:09:47.654234 systemd[1]: issuegen.service: Deactivated successfully. Dec 13 14:09:47.654386 systemd[1]: Finished issuegen.service. Dec 13 14:09:47.656199 systemd[1]: Starting systemd-user-sessions.service... Dec 13 14:09:47.663854 systemd[1]: Finished systemd-user-sessions.service. Dec 13 14:09:47.665705 systemd[1]: Started getty@tty1.service. Dec 13 14:09:47.667468 systemd[1]: Started serial-getty@ttyAMA0.service. Dec 13 14:09:47.668292 systemd[1]: Reached target getty.target. Dec 13 14:09:47.669033 systemd[1]: Reached target multi-user.target. Dec 13 14:09:47.670728 systemd[1]: Starting systemd-update-utmp-runlevel.service... Dec 13 14:09:47.676826 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Dec 13 14:09:47.676998 systemd[1]: Finished systemd-update-utmp-runlevel.service. Dec 13 14:09:47.677785 systemd[1]: Startup finished in 553ms (kernel) + 4.140s (initrd) + 5.969s (userspace) = 10.663s. Dec 13 14:09:49.602845 systemd[1]: Created slice system-sshd.slice. Dec 13 14:09:49.603951 systemd[1]: Started sshd@0-10.0.0.79:22-10.0.0.1:53442.service. Dec 13 14:09:49.652761 sshd[1281]: Accepted publickey for core from 10.0.0.1 port 53442 ssh2: RSA SHA256:/HJyHm5Z3TKV0xVrRefgtheJNUHxRnoHBht1EzpqsE0 Dec 13 14:09:49.654861 sshd[1281]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:09:49.662362 systemd[1]: Created slice user-500.slice. Dec 13 14:09:49.663432 systemd[1]: Starting user-runtime-dir@500.service... Dec 13 14:09:49.664932 systemd-logind[1203]: New session 1 of user core. Dec 13 14:09:49.671008 systemd[1]: Finished user-runtime-dir@500.service. Dec 13 14:09:49.672250 systemd[1]: Starting user@500.service... Dec 13 14:09:49.674929 (systemd)[1284]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:09:49.730157 systemd[1284]: Queued start job for default target default.target. Dec 13 14:09:49.730579 systemd[1284]: Reached target paths.target. Dec 13 14:09:49.730598 systemd[1284]: Reached target sockets.target. Dec 13 14:09:49.730610 systemd[1284]: Reached target timers.target. Dec 13 14:09:49.730620 systemd[1284]: Reached target basic.target. Dec 13 14:09:49.730668 systemd[1284]: Reached target default.target. Dec 13 14:09:49.730694 systemd[1284]: Startup finished in 50ms. Dec 13 14:09:49.730860 systemd[1]: Started user@500.service. Dec 13 14:09:49.731758 systemd[1]: Started session-1.scope. Dec 13 14:09:49.781313 systemd[1]: Started sshd@1-10.0.0.79:22-10.0.0.1:53448.service. Dec 13 14:09:49.812402 sshd[1293]: Accepted publickey for core from 10.0.0.1 port 53448 ssh2: RSA SHA256:/HJyHm5Z3TKV0xVrRefgtheJNUHxRnoHBht1EzpqsE0 Dec 13 14:09:49.813573 sshd[1293]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:09:49.817079 systemd-logind[1203]: New session 2 of user core. Dec 13 14:09:49.818172 systemd[1]: Started session-2.scope. Dec 13 14:09:49.873194 sshd[1293]: pam_unix(sshd:session): session closed for user core Dec 13 14:09:49.876193 systemd[1]: Started sshd@2-10.0.0.79:22-10.0.0.1:53458.service. Dec 13 14:09:49.876619 systemd[1]: sshd@1-10.0.0.79:22-10.0.0.1:53448.service: Deactivated successfully. Dec 13 14:09:49.877304 systemd[1]: session-2.scope: Deactivated successfully. Dec 13 14:09:49.877834 systemd-logind[1203]: Session 2 logged out. Waiting for processes to exit. Dec 13 14:09:49.878667 systemd-logind[1203]: Removed session 2. Dec 13 14:09:49.909426 sshd[1298]: Accepted publickey for core from 10.0.0.1 port 53458 ssh2: RSA SHA256:/HJyHm5Z3TKV0xVrRefgtheJNUHxRnoHBht1EzpqsE0 Dec 13 14:09:49.910684 sshd[1298]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:09:49.914727 systemd[1]: Started session-3.scope. Dec 13 14:09:49.915055 systemd-logind[1203]: New session 3 of user core. Dec 13 14:09:49.964683 sshd[1298]: pam_unix(sshd:session): session closed for user core Dec 13 14:09:49.967403 systemd[1]: sshd@2-10.0.0.79:22-10.0.0.1:53458.service: Deactivated successfully. Dec 13 14:09:49.967981 systemd[1]: session-3.scope: Deactivated successfully. Dec 13 14:09:49.968434 systemd-logind[1203]: Session 3 logged out. Waiting for processes to exit. Dec 13 14:09:49.969488 systemd[1]: Started sshd@3-10.0.0.79:22-10.0.0.1:53472.service. Dec 13 14:09:49.970119 systemd-logind[1203]: Removed session 3. Dec 13 14:09:50.001451 sshd[1305]: Accepted publickey for core from 10.0.0.1 port 53472 ssh2: RSA SHA256:/HJyHm5Z3TKV0xVrRefgtheJNUHxRnoHBht1EzpqsE0 Dec 13 14:09:50.002765 sshd[1305]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:09:50.010690 systemd-logind[1203]: New session 4 of user core. Dec 13 14:09:50.011493 systemd[1]: Started session-4.scope. Dec 13 14:09:50.065620 sshd[1305]: pam_unix(sshd:session): session closed for user core Dec 13 14:09:50.068821 systemd[1]: Started sshd@4-10.0.0.79:22-10.0.0.1:53482.service. Dec 13 14:09:50.069362 systemd[1]: sshd@3-10.0.0.79:22-10.0.0.1:53472.service: Deactivated successfully. Dec 13 14:09:50.070101 systemd[1]: session-4.scope: Deactivated successfully. Dec 13 14:09:50.070588 systemd-logind[1203]: Session 4 logged out. Waiting for processes to exit. Dec 13 14:09:50.071319 systemd-logind[1203]: Removed session 4. Dec 13 14:09:50.100497 sshd[1311]: Accepted publickey for core from 10.0.0.1 port 53482 ssh2: RSA SHA256:/HJyHm5Z3TKV0xVrRefgtheJNUHxRnoHBht1EzpqsE0 Dec 13 14:09:50.101754 sshd[1311]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:09:50.105271 systemd-logind[1203]: New session 5 of user core. Dec 13 14:09:50.106045 systemd[1]: Started session-5.scope. Dec 13 14:09:50.165658 sudo[1315]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 13 14:09:50.165899 sudo[1315]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Dec 13 14:09:50.180256 systemd[1]: Starting coreos-metadata.service... Dec 13 14:09:50.187016 systemd[1]: coreos-metadata.service: Deactivated successfully. Dec 13 14:09:50.187182 systemd[1]: Finished coreos-metadata.service. Dec 13 14:09:50.686760 systemd[1]: Stopped kubelet.service. Dec 13 14:09:50.688702 systemd[1]: Starting kubelet.service... Dec 13 14:09:50.706830 systemd[1]: Reloading. Dec 13 14:09:50.756715 /usr/lib/systemd/system-generators/torcx-generator[1382]: time="2024-12-13T14:09:50Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.6 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.6 /var/lib/torcx/store]" Dec 13 14:09:50.757045 /usr/lib/systemd/system-generators/torcx-generator[1382]: time="2024-12-13T14:09:50Z" level=info msg="torcx already run" Dec 13 14:09:50.826688 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Dec 13 14:09:50.826831 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Dec 13 14:09:50.841765 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 14:09:50.904266 systemd[1]: Started kubelet.service. Dec 13 14:09:50.906073 systemd[1]: Stopping kubelet.service... Dec 13 14:09:50.906494 systemd[1]: kubelet.service: Deactivated successfully. Dec 13 14:09:50.906798 systemd[1]: Stopped kubelet.service. Dec 13 14:09:50.908563 systemd[1]: Starting kubelet.service... Dec 13 14:09:50.985544 systemd[1]: Started kubelet.service. Dec 13 14:09:51.020488 kubelet[1425]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 14:09:51.020488 kubelet[1425]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 14:09:51.020488 kubelet[1425]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 14:09:51.022750 kubelet[1425]: I1213 14:09:51.022705 1425 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 14:09:51.739926 kubelet[1425]: I1213 14:09:51.739876 1425 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Dec 13 14:09:51.739926 kubelet[1425]: I1213 14:09:51.739914 1425 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 14:09:51.740120 kubelet[1425]: I1213 14:09:51.740104 1425 server.go:927] "Client rotation is on, will bootstrap in background" Dec 13 14:09:51.758606 kubelet[1425]: I1213 14:09:51.758163 1425 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 14:09:51.765962 kubelet[1425]: I1213 14:09:51.765932 1425 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 14:09:51.766372 kubelet[1425]: I1213 14:09:51.766345 1425 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 14:09:51.766522 kubelet[1425]: I1213 14:09:51.766371 1425 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"10.0.0.79","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Dec 13 14:09:51.766614 kubelet[1425]: I1213 14:09:51.766590 1425 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 14:09:51.766614 kubelet[1425]: I1213 14:09:51.766600 1425 container_manager_linux.go:301] "Creating device plugin manager" Dec 13 14:09:51.766784 kubelet[1425]: I1213 14:09:51.766771 1425 state_mem.go:36] "Initialized new in-memory state store" Dec 13 14:09:51.767850 kubelet[1425]: I1213 14:09:51.767829 1425 kubelet.go:400] "Attempting to sync node with API server" Dec 13 14:09:51.767850 kubelet[1425]: I1213 14:09:51.767851 1425 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 14:09:51.768179 kubelet[1425]: I1213 14:09:51.768157 1425 kubelet.go:312] "Adding apiserver pod source" Dec 13 14:09:51.768357 kubelet[1425]: I1213 14:09:51.768320 1425 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 14:09:51.768522 kubelet[1425]: E1213 14:09:51.768499 1425 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:51.768563 kubelet[1425]: E1213 14:09:51.768547 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:51.771247 kubelet[1425]: I1213 14:09:51.771206 1425 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Dec 13 14:09:51.771612 kubelet[1425]: I1213 14:09:51.771595 1425 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 14:09:51.771671 kubelet[1425]: W1213 14:09:51.771637 1425 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 13 14:09:51.772551 kubelet[1425]: I1213 14:09:51.772527 1425 server.go:1264] "Started kubelet" Dec 13 14:09:51.773440 kubelet[1425]: I1213 14:09:51.773404 1425 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 14:09:51.793261 kubelet[1425]: I1213 14:09:51.793218 1425 server.go:455] "Adding debug handlers to kubelet server" Dec 13 14:09:51.793383 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Dec 13 14:09:51.793503 kubelet[1425]: I1213 14:09:51.793481 1425 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 14:09:51.794739 kubelet[1425]: I1213 14:09:51.794718 1425 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Dec 13 14:09:51.795467 kubelet[1425]: I1213 14:09:51.795447 1425 volume_manager.go:291] "Starting Kubelet Volume Manager" Dec 13 14:09:51.795833 kubelet[1425]: I1213 14:09:51.793288 1425 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 14:09:51.797587 kubelet[1425]: I1213 14:09:51.797554 1425 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 14:09:51.797906 kubelet[1425]: W1213 14:09:51.797835 1425 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "10.0.0.79" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 13 14:09:51.797906 kubelet[1425]: E1213 14:09:51.797879 1425 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes "10.0.0.79" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 13 14:09:51.798050 kubelet[1425]: W1213 14:09:51.797979 1425 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 13 14:09:51.798050 kubelet[1425]: E1213 14:09:51.797992 1425 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 13 14:09:51.799878 kubelet[1425]: I1213 14:09:51.799428 1425 reconciler.go:26] "Reconciler: start to sync state" Dec 13 14:09:51.799878 kubelet[1425]: I1213 14:09:51.799758 1425 factory.go:221] Registration of the systemd container factory successfully Dec 13 14:09:51.799878 kubelet[1425]: I1213 14:09:51.799850 1425 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 14:09:51.804436 kubelet[1425]: E1213 14:09:51.804405 1425 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 14:09:51.804524 kubelet[1425]: I1213 14:09:51.804491 1425 factory.go:221] Registration of the containerd container factory successfully Dec 13 14:09:51.814674 kubelet[1425]: I1213 14:09:51.814120 1425 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 14:09:51.815697 kubelet[1425]: I1213 14:09:51.815667 1425 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 14:09:51.815697 kubelet[1425]: I1213 14:09:51.815701 1425 state_mem.go:36] "Initialized new in-memory state store" Dec 13 14:09:51.815820 kubelet[1425]: E1213 14:09:51.815729 1425 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"10.0.0.79\" not found" node="10.0.0.79" Dec 13 14:09:51.896033 kubelet[1425]: I1213 14:09:51.896005 1425 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.79" Dec 13 14:09:51.954072 kubelet[1425]: I1213 14:09:51.954039 1425 policy_none.go:49] "None policy: Start" Dec 13 14:09:51.954863 kubelet[1425]: I1213 14:09:51.954846 1425 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 14:09:51.954863 kubelet[1425]: I1213 14:09:51.954871 1425 state_mem.go:35] "Initializing new in-memory state store" Dec 13 14:09:51.956224 kubelet[1425]: I1213 14:09:51.956201 1425 kubelet_node_status.go:76] "Successfully registered node" node="10.0.0.79" Dec 13 14:09:51.978907 systemd[1]: Created slice kubepods.slice. Dec 13 14:09:51.982601 systemd[1]: Created slice kubepods-burstable.slice. Dec 13 14:09:51.983599 kubelet[1425]: E1213 14:09:51.983571 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:51.985253 systemd[1]: Created slice kubepods-besteffort.slice. Dec 13 14:09:51.994767 kubelet[1425]: I1213 14:09:51.994669 1425 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 14:09:51.995797 kubelet[1425]: I1213 14:09:51.995763 1425 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 14:09:51.996448 kubelet[1425]: I1213 14:09:51.996430 1425 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 14:09:51.996614 kubelet[1425]: I1213 14:09:51.996585 1425 kubelet.go:2337] "Starting kubelet main sync loop" Dec 13 14:09:51.997083 kubelet[1425]: E1213 14:09:51.997055 1425 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 14:09:51.997822 kubelet[1425]: I1213 14:09:51.997798 1425 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 14:09:51.998081 kubelet[1425]: I1213 14:09:51.998044 1425 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 13 14:09:51.998240 kubelet[1425]: I1213 14:09:51.998228 1425 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 14:09:52.000842 kubelet[1425]: E1213 14:09:52.000826 1425 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"10.0.0.79\" not found" Dec 13 14:09:52.018739 sudo[1315]: pam_unix(sudo:session): session closed for user root Dec 13 14:09:52.020461 sshd[1311]: pam_unix(sshd:session): session closed for user core Dec 13 14:09:52.022602 systemd[1]: session-5.scope: Deactivated successfully. Dec 13 14:09:52.023202 systemd-logind[1203]: Session 5 logged out. Waiting for processes to exit. Dec 13 14:09:52.023307 systemd[1]: sshd@4-10.0.0.79:22-10.0.0.1:53482.service: Deactivated successfully. Dec 13 14:09:52.024226 systemd-logind[1203]: Removed session 5. Dec 13 14:09:52.084544 kubelet[1425]: E1213 14:09:52.084491 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:52.185054 kubelet[1425]: E1213 14:09:52.185002 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:52.285774 kubelet[1425]: E1213 14:09:52.285675 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:52.386121 kubelet[1425]: E1213 14:09:52.386081 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:52.486529 kubelet[1425]: E1213 14:09:52.486493 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:52.587047 kubelet[1425]: E1213 14:09:52.586942 1425 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"10.0.0.79\" not found" Dec 13 14:09:52.688518 kubelet[1425]: I1213 14:09:52.688487 1425 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Dec 13 14:09:52.688956 env[1212]: time="2024-12-13T14:09:52.688768153Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 13 14:09:52.689219 kubelet[1425]: I1213 14:09:52.688938 1425 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Dec 13 14:09:52.741973 kubelet[1425]: I1213 14:09:52.741902 1425 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 13 14:09:52.742095 kubelet[1425]: W1213 14:09:52.742068 1425 reflector.go:470] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 13 14:09:52.742121 kubelet[1425]: W1213 14:09:52.742099 1425 reflector.go:470] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 13 14:09:52.742167 kubelet[1425]: W1213 14:09:52.742120 1425 reflector.go:470] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 13 14:09:52.769299 kubelet[1425]: I1213 14:09:52.769184 1425 apiserver.go:52] "Watching apiserver" Dec 13 14:09:52.769299 kubelet[1425]: E1213 14:09:52.769179 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:52.784728 kubelet[1425]: I1213 14:09:52.784658 1425 topology_manager.go:215] "Topology Admit Handler" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" podNamespace="kube-system" podName="cilium-dckzx" Dec 13 14:09:52.784846 kubelet[1425]: I1213 14:09:52.784821 1425 topology_manager.go:215] "Topology Admit Handler" podUID="3cbd2446-4723-4796-b3bf-6e8c7aa2571f" podNamespace="kube-system" podName="kube-proxy-fdhrn" Dec 13 14:09:52.789343 systemd[1]: Created slice kubepods-besteffort-pod3cbd2446_4723_4796_b3bf_6e8c7aa2571f.slice. Dec 13 14:09:52.795544 kubelet[1425]: I1213 14:09:52.795500 1425 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Dec 13 14:09:52.800200 systemd[1]: Created slice kubepods-burstable-pod15a016f6_3ef3_4719_9ea1_8422641ba5a5.slice. Dec 13 14:09:52.804160 kubelet[1425]: I1213 14:09:52.804000 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3cbd2446-4723-4796-b3bf-6e8c7aa2571f-kube-proxy\") pod \"kube-proxy-fdhrn\" (UID: \"3cbd2446-4723-4796-b3bf-6e8c7aa2571f\") " pod="kube-system/kube-proxy-fdhrn" Dec 13 14:09:52.804160 kubelet[1425]: I1213 14:09:52.804049 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfwnk\" (UniqueName: \"kubernetes.io/projected/3cbd2446-4723-4796-b3bf-6e8c7aa2571f-kube-api-access-tfwnk\") pod \"kube-proxy-fdhrn\" (UID: \"3cbd2446-4723-4796-b3bf-6e8c7aa2571f\") " pod="kube-system/kube-proxy-fdhrn" Dec 13 14:09:52.804160 kubelet[1425]: I1213 14:09:52.804073 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-run\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804160 kubelet[1425]: I1213 14:09:52.804092 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-lib-modules\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804160 kubelet[1425]: I1213 14:09:52.804112 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shzrw\" (UniqueName: \"kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-kube-api-access-shzrw\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804621 kubelet[1425]: I1213 14:09:52.804133 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/15a016f6-3ef3-4719-9ea1-8422641ba5a5-clustermesh-secrets\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804621 kubelet[1425]: I1213 14:09:52.804151 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-net\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804621 kubelet[1425]: I1213 14:09:52.804170 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-kernel\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804621 kubelet[1425]: I1213 14:09:52.804202 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3cbd2446-4723-4796-b3bf-6e8c7aa2571f-xtables-lock\") pod \"kube-proxy-fdhrn\" (UID: \"3cbd2446-4723-4796-b3bf-6e8c7aa2571f\") " pod="kube-system/kube-proxy-fdhrn" Dec 13 14:09:52.804621 kubelet[1425]: I1213 14:09:52.804233 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3cbd2446-4723-4796-b3bf-6e8c7aa2571f-lib-modules\") pod \"kube-proxy-fdhrn\" (UID: \"3cbd2446-4723-4796-b3bf-6e8c7aa2571f\") " pod="kube-system/kube-proxy-fdhrn" Dec 13 14:09:52.804728 kubelet[1425]: I1213 14:09:52.804256 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-bpf-maps\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804728 kubelet[1425]: I1213 14:09:52.804276 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hostproc\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804728 kubelet[1425]: I1213 14:09:52.804293 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-etc-cni-netd\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804728 kubelet[1425]: I1213 14:09:52.804310 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-cgroup\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804728 kubelet[1425]: I1213 14:09:52.804328 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-config-path\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804728 kubelet[1425]: I1213 14:09:52.804348 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hubble-tls\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804851 kubelet[1425]: I1213 14:09:52.804379 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cni-path\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:52.804851 kubelet[1425]: I1213 14:09:52.804431 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-xtables-lock\") pod \"cilium-dckzx\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " pod="kube-system/cilium-dckzx" Dec 13 14:09:53.098140 kubelet[1425]: E1213 14:09:53.098090 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:09:53.098802 env[1212]: time="2024-12-13T14:09:53.098736814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-fdhrn,Uid:3cbd2446-4723-4796-b3bf-6e8c7aa2571f,Namespace:kube-system,Attempt:0,}" Dec 13 14:09:53.110074 kubelet[1425]: E1213 14:09:53.110042 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:09:53.110523 env[1212]: time="2024-12-13T14:09:53.110446095Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-dckzx,Uid:15a016f6-3ef3-4719-9ea1-8422641ba5a5,Namespace:kube-system,Attempt:0,}" Dec 13 14:09:53.624373 env[1212]: time="2024-12-13T14:09:53.624330977Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.626585 env[1212]: time="2024-12-13T14:09:53.626543340Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.627670 env[1212]: time="2024-12-13T14:09:53.627633104Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.629339 env[1212]: time="2024-12-13T14:09:53.629308156Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.631921 env[1212]: time="2024-12-13T14:09:53.631865290Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.636320 env[1212]: time="2024-12-13T14:09:53.636284118Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.637565 env[1212]: time="2024-12-13T14:09:53.637534371Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.638550 env[1212]: time="2024-12-13T14:09:53.638507847Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:09:53.664889 env[1212]: time="2024-12-13T14:09:53.664796736Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:09:53.664889 env[1212]: time="2024-12-13T14:09:53.664840270Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:09:53.664889 env[1212]: time="2024-12-13T14:09:53.664853051Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:09:53.665058 env[1212]: time="2024-12-13T14:09:53.664902602Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:09:53.665058 env[1212]: time="2024-12-13T14:09:53.664968433Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:09:53.665058 env[1212]: time="2024-12-13T14:09:53.664998596Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:09:53.665417 env[1212]: time="2024-12-13T14:09:53.665373649Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909 pid=1487 runtime=io.containerd.runc.v2 Dec 13 14:09:53.665511 env[1212]: time="2024-12-13T14:09:53.665456508Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d39b2b18ee648318f3993717090ac4e9bec18bbe94edfbacc251b5b96680349c pid=1486 runtime=io.containerd.runc.v2 Dec 13 14:09:53.705555 systemd[1]: Started cri-containerd-6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909.scope. Dec 13 14:09:53.706938 systemd[1]: Started cri-containerd-d39b2b18ee648318f3993717090ac4e9bec18bbe94edfbacc251b5b96680349c.scope. Dec 13 14:09:53.765716 env[1212]: time="2024-12-13T14:09:53.765666936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-dckzx,Uid:15a016f6-3ef3-4719-9ea1-8422641ba5a5,Namespace:kube-system,Attempt:0,} returns sandbox id \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\"" Dec 13 14:09:53.766215 env[1212]: time="2024-12-13T14:09:53.765720773Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-fdhrn,Uid:3cbd2446-4723-4796-b3bf-6e8c7aa2571f,Namespace:kube-system,Attempt:0,} returns sandbox id \"d39b2b18ee648318f3993717090ac4e9bec18bbe94edfbacc251b5b96680349c\"" Dec 13 14:09:53.772499 kubelet[1425]: E1213 14:09:53.770136 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:53.772499 kubelet[1425]: E1213 14:09:53.771783 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:09:53.772746 kubelet[1425]: E1213 14:09:53.772723 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:09:53.774004 env[1212]: time="2024-12-13T14:09:53.773971960Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Dec 13 14:09:53.912406 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount648183733.mount: Deactivated successfully. Dec 13 14:09:54.770658 kubelet[1425]: E1213 14:09:54.770613 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:55.771717 kubelet[1425]: E1213 14:09:55.771678 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:56.772708 kubelet[1425]: E1213 14:09:56.772654 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:57.773554 kubelet[1425]: E1213 14:09:57.773497 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:58.008153 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2609102307.mount: Deactivated successfully. Dec 13 14:09:58.774242 kubelet[1425]: E1213 14:09:58.774193 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:09:59.774737 kubelet[1425]: E1213 14:09:59.774673 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:00.110922 env[1212]: time="2024-12-13T14:10:00.110805614Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:00.112536 env[1212]: time="2024-12-13T14:10:00.112497465Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:00.114090 env[1212]: time="2024-12-13T14:10:00.114065540Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:00.115402 env[1212]: time="2024-12-13T14:10:00.115350064Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Dec 13 14:10:00.116344 env[1212]: time="2024-12-13T14:10:00.116316984Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.8\"" Dec 13 14:10:00.117806 env[1212]: time="2024-12-13T14:10:00.117770741Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 13 14:10:00.126758 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3856836044.mount: Deactivated successfully. Dec 13 14:10:00.130842 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount890711865.mount: Deactivated successfully. Dec 13 14:10:00.135506 env[1212]: time="2024-12-13T14:10:00.135461425Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\"" Dec 13 14:10:00.136151 env[1212]: time="2024-12-13T14:10:00.136118608Z" level=info msg="StartContainer for \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\"" Dec 13 14:10:00.152337 systemd[1]: Started cri-containerd-ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98.scope. Dec 13 14:10:00.190506 env[1212]: time="2024-12-13T14:10:00.190444033Z" level=info msg="StartContainer for \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\" returns successfully" Dec 13 14:10:00.223278 systemd[1]: cri-containerd-ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98.scope: Deactivated successfully. Dec 13 14:10:00.394050 env[1212]: time="2024-12-13T14:10:00.393938501Z" level=info msg="shim disconnected" id=ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98 Dec 13 14:10:00.394050 env[1212]: time="2024-12-13T14:10:00.393981097Z" level=warning msg="cleaning up after shim disconnected" id=ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98 namespace=k8s.io Dec 13 14:10:00.394050 env[1212]: time="2024-12-13T14:10:00.393990872Z" level=info msg="cleaning up dead shim" Dec 13 14:10:00.401680 env[1212]: time="2024-12-13T14:10:00.401631380Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:00Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1603 runtime=io.containerd.runc.v2\n" Dec 13 14:10:00.775739 kubelet[1425]: E1213 14:10:00.775627 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:01.016509 kubelet[1425]: E1213 14:10:01.016478 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:01.018506 env[1212]: time="2024-12-13T14:10:01.018458799Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 13 14:10:01.059076 env[1212]: time="2024-12-13T14:10:01.058971834Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\"" Dec 13 14:10:01.059800 env[1212]: time="2024-12-13T14:10:01.059772333Z" level=info msg="StartContainer for \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\"" Dec 13 14:10:01.074380 systemd[1]: Started cri-containerd-ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748.scope. Dec 13 14:10:01.113908 env[1212]: time="2024-12-13T14:10:01.111093190Z" level=info msg="StartContainer for \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\" returns successfully" Dec 13 14:10:01.126395 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98-rootfs.mount: Deactivated successfully. Dec 13 14:10:01.129519 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 14:10:01.130028 systemd[1]: Stopped systemd-sysctl.service. Dec 13 14:10:01.130345 systemd[1]: Stopping systemd-sysctl.service... Dec 13 14:10:01.131768 systemd[1]: Starting systemd-sysctl.service... Dec 13 14:10:01.133790 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Dec 13 14:10:01.137039 systemd[1]: cri-containerd-ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748.scope: Deactivated successfully. Dec 13 14:10:01.140628 systemd[1]: Finished systemd-sysctl.service. Dec 13 14:10:01.151859 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748-rootfs.mount: Deactivated successfully. Dec 13 14:10:01.169516 env[1212]: time="2024-12-13T14:10:01.169467253Z" level=info msg="shim disconnected" id=ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748 Dec 13 14:10:01.169516 env[1212]: time="2024-12-13T14:10:01.169514776Z" level=warning msg="cleaning up after shim disconnected" id=ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748 namespace=k8s.io Dec 13 14:10:01.169708 env[1212]: time="2024-12-13T14:10:01.169525116Z" level=info msg="cleaning up dead shim" Dec 13 14:10:01.175552 env[1212]: time="2024-12-13T14:10:01.175509626Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:01Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1668 runtime=io.containerd.runc.v2\n" Dec 13 14:10:01.386249 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2400254630.mount: Deactivated successfully. Dec 13 14:10:01.776595 kubelet[1425]: E1213 14:10:01.776288 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:01.793041 env[1212]: time="2024-12-13T14:10:01.793002268Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.30.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:01.794066 env[1212]: time="2024-12-13T14:10:01.794037201Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:4612aebc0675831aedbbde7cd56b85db91f1fdcf05ef923072961538ec497adb,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:01.795756 env[1212]: time="2024-12-13T14:10:01.795729387Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.30.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:01.797315 env[1212]: time="2024-12-13T14:10:01.797289859Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:f6d6be9417e22af78905000ac4fd134896bacd2188ea63c7cac8edd7a5d7e9b5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:01.797828 env[1212]: time="2024-12-13T14:10:01.797797543Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.8\" returns image reference \"sha256:4612aebc0675831aedbbde7cd56b85db91f1fdcf05ef923072961538ec497adb\"" Dec 13 14:10:01.800474 env[1212]: time="2024-12-13T14:10:01.800438364Z" level=info msg="CreateContainer within sandbox \"d39b2b18ee648318f3993717090ac4e9bec18bbe94edfbacc251b5b96680349c\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 13 14:10:01.810520 env[1212]: time="2024-12-13T14:10:01.810482990Z" level=info msg="CreateContainer within sandbox \"d39b2b18ee648318f3993717090ac4e9bec18bbe94edfbacc251b5b96680349c\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"91fe41ecacbe6480fc77f8494763cb4df4f4ec8733ef5da3a8f793e37400a499\"" Dec 13 14:10:01.811347 env[1212]: time="2024-12-13T14:10:01.811312679Z" level=info msg="StartContainer for \"91fe41ecacbe6480fc77f8494763cb4df4f4ec8733ef5da3a8f793e37400a499\"" Dec 13 14:10:01.824756 systemd[1]: Started cri-containerd-91fe41ecacbe6480fc77f8494763cb4df4f4ec8733ef5da3a8f793e37400a499.scope. Dec 13 14:10:01.871639 env[1212]: time="2024-12-13T14:10:01.871582184Z" level=info msg="StartContainer for \"91fe41ecacbe6480fc77f8494763cb4df4f4ec8733ef5da3a8f793e37400a499\" returns successfully" Dec 13 14:10:02.019353 kubelet[1425]: E1213 14:10:02.019312 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:02.020969 kubelet[1425]: E1213 14:10:02.020941 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:02.021576 env[1212]: time="2024-12-13T14:10:02.021541178Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 13 14:10:02.035990 env[1212]: time="2024-12-13T14:10:02.035310906Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\"" Dec 13 14:10:02.035990 env[1212]: time="2024-12-13T14:10:02.035964139Z" level=info msg="StartContainer for \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\"" Dec 13 14:10:02.051458 systemd[1]: Started cri-containerd-faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312.scope. Dec 13 14:10:02.086368 env[1212]: time="2024-12-13T14:10:02.086310626Z" level=info msg="StartContainer for \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\" returns successfully" Dec 13 14:10:02.101166 systemd[1]: cri-containerd-faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312.scope: Deactivated successfully. Dec 13 14:10:02.221683 env[1212]: time="2024-12-13T14:10:02.221633949Z" level=info msg="shim disconnected" id=faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312 Dec 13 14:10:02.221683 env[1212]: time="2024-12-13T14:10:02.221680114Z" level=warning msg="cleaning up after shim disconnected" id=faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312 namespace=k8s.io Dec 13 14:10:02.221683 env[1212]: time="2024-12-13T14:10:02.221692610Z" level=info msg="cleaning up dead shim" Dec 13 14:10:02.232164 env[1212]: time="2024-12-13T14:10:02.228775824Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:02Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1837 runtime=io.containerd.runc.v2\n" Dec 13 14:10:02.776964 kubelet[1425]: E1213 14:10:02.776926 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:03.025190 kubelet[1425]: E1213 14:10:03.025159 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:03.025650 kubelet[1425]: E1213 14:10:03.025629 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:03.027948 env[1212]: time="2024-12-13T14:10:03.027647066Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 13 14:10:03.040081 kubelet[1425]: I1213 14:10:03.039744 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-fdhrn" podStartSLOduration=4.014323374 podStartE2EDuration="12.039727757s" podCreationTimestamp="2024-12-13 14:09:51 +0000 UTC" firstStartedPulling="2024-12-13 14:09:53.773639105 +0000 UTC m=+2.784084054" lastFinishedPulling="2024-12-13 14:10:01.799043487 +0000 UTC m=+10.809488437" observedRunningTime="2024-12-13 14:10:02.040765969 +0000 UTC m=+11.051210919" watchObservedRunningTime="2024-12-13 14:10:03.039727757 +0000 UTC m=+12.050172706" Dec 13 14:10:03.039792 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2192014738.mount: Deactivated successfully. Dec 13 14:10:03.043951 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1105477490.mount: Deactivated successfully. Dec 13 14:10:03.047798 env[1212]: time="2024-12-13T14:10:03.047734069Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\"" Dec 13 14:10:03.048547 env[1212]: time="2024-12-13T14:10:03.048508978Z" level=info msg="StartContainer for \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\"" Dec 13 14:10:03.062576 systemd[1]: Started cri-containerd-0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a.scope. Dec 13 14:10:03.090732 systemd[1]: cri-containerd-0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a.scope: Deactivated successfully. Dec 13 14:10:03.091350 env[1212]: time="2024-12-13T14:10:03.091255132Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod15a016f6_3ef3_4719_9ea1_8422641ba5a5.slice/cri-containerd-0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a.scope/memory.events\": no such file or directory" Dec 13 14:10:03.093446 env[1212]: time="2024-12-13T14:10:03.093399903Z" level=info msg="StartContainer for \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\" returns successfully" Dec 13 14:10:03.113089 env[1212]: time="2024-12-13T14:10:03.113043599Z" level=info msg="shim disconnected" id=0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a Dec 13 14:10:03.113089 env[1212]: time="2024-12-13T14:10:03.113093457Z" level=warning msg="cleaning up after shim disconnected" id=0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a namespace=k8s.io Dec 13 14:10:03.113285 env[1212]: time="2024-12-13T14:10:03.113103533Z" level=info msg="cleaning up dead shim" Dec 13 14:10:03.119724 env[1212]: time="2024-12-13T14:10:03.119686024Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1942 runtime=io.containerd.runc.v2\n" Dec 13 14:10:03.777977 kubelet[1425]: E1213 14:10:03.777940 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:04.028651 kubelet[1425]: E1213 14:10:04.028356 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:04.032373 env[1212]: time="2024-12-13T14:10:04.032328826Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 13 14:10:04.054052 env[1212]: time="2024-12-13T14:10:04.054003949Z" level=info msg="CreateContainer within sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\"" Dec 13 14:10:04.054908 env[1212]: time="2024-12-13T14:10:04.054853162Z" level=info msg="StartContainer for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\"" Dec 13 14:10:04.073527 systemd[1]: Started cri-containerd-b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445.scope. Dec 13 14:10:04.114565 env[1212]: time="2024-12-13T14:10:04.114512623Z" level=info msg="StartContainer for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" returns successfully" Dec 13 14:10:04.272830 kubelet[1425]: I1213 14:10:04.272193 1425 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Dec 13 14:10:04.356953 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Dec 13 14:10:04.589978 kernel: Initializing XFRM netlink socket Dec 13 14:10:04.591896 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Dec 13 14:10:04.779030 kubelet[1425]: E1213 14:10:04.778966 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:05.033748 kubelet[1425]: E1213 14:10:05.033086 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:05.047387 kubelet[1425]: I1213 14:10:05.047325 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-dckzx" podStartSLOduration=7.704689892 podStartE2EDuration="14.047307932s" podCreationTimestamp="2024-12-13 14:09:51 +0000 UTC" firstStartedPulling="2024-12-13 14:09:53.773582004 +0000 UTC m=+2.784026953" lastFinishedPulling="2024-12-13 14:10:00.116199963 +0000 UTC m=+9.126644993" observedRunningTime="2024-12-13 14:10:05.047148316 +0000 UTC m=+14.057593266" watchObservedRunningTime="2024-12-13 14:10:05.047307932 +0000 UTC m=+14.057752882" Dec 13 14:10:05.779653 kubelet[1425]: E1213 14:10:05.779594 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:06.034681 kubelet[1425]: E1213 14:10:06.034410 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:06.190070 systemd-networkd[1044]: cilium_host: Link UP Dec 13 14:10:06.191075 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Dec 13 14:10:06.191113 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Dec 13 14:10:06.191120 systemd-networkd[1044]: cilium_net: Link UP Dec 13 14:10:06.191325 systemd-networkd[1044]: cilium_net: Gained carrier Dec 13 14:10:06.191523 systemd-networkd[1044]: cilium_host: Gained carrier Dec 13 14:10:06.266439 systemd-networkd[1044]: cilium_vxlan: Link UP Dec 13 14:10:06.266445 systemd-networkd[1044]: cilium_vxlan: Gained carrier Dec 13 14:10:06.284984 systemd-networkd[1044]: cilium_net: Gained IPv6LL Dec 13 14:10:06.547910 kernel: NET: Registered PF_ALG protocol family Dec 13 14:10:06.780446 kubelet[1425]: E1213 14:10:06.780394 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:07.036039 kubelet[1425]: E1213 14:10:07.035958 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:07.088566 systemd-networkd[1044]: lxc_health: Link UP Dec 13 14:10:07.097414 systemd-networkd[1044]: lxc_health: Gained carrier Dec 13 14:10:07.097902 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Dec 13 14:10:07.189005 systemd-networkd[1044]: cilium_host: Gained IPv6LL Dec 13 14:10:07.445014 systemd-networkd[1044]: cilium_vxlan: Gained IPv6LL Dec 13 14:10:07.780831 kubelet[1425]: E1213 14:10:07.780723 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:07.795324 kubelet[1425]: I1213 14:10:07.795279 1425 topology_manager.go:215] "Topology Admit Handler" podUID="231c0ef0-e192-4396-a0bc-d38605ed3f25" podNamespace="default" podName="nginx-deployment-85f456d6dd-6qhn4" Dec 13 14:10:07.800032 systemd[1]: Created slice kubepods-besteffort-pod231c0ef0_e192_4396_a0bc_d38605ed3f25.slice. Dec 13 14:10:07.887035 kubelet[1425]: I1213 14:10:07.886990 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpn7w\" (UniqueName: \"kubernetes.io/projected/231c0ef0-e192-4396-a0bc-d38605ed3f25-kube-api-access-vpn7w\") pod \"nginx-deployment-85f456d6dd-6qhn4\" (UID: \"231c0ef0-e192-4396-a0bc-d38605ed3f25\") " pod="default/nginx-deployment-85f456d6dd-6qhn4" Dec 13 14:10:08.038096 kubelet[1425]: E1213 14:10:08.038013 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:08.102267 env[1212]: time="2024-12-13T14:10:08.102224431Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-6qhn4,Uid:231c0ef0-e192-4396-a0bc-d38605ed3f25,Namespace:default,Attempt:0,}" Dec 13 14:10:08.186968 systemd-networkd[1044]: lxcba7fc638d780: Link UP Dec 13 14:10:08.195935 kernel: eth0: renamed from tmpe8eaa Dec 13 14:10:08.203386 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Dec 13 14:10:08.203456 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcba7fc638d780: link becomes ready Dec 13 14:10:08.203561 systemd-networkd[1044]: lxcba7fc638d780: Gained carrier Dec 13 14:10:08.597025 systemd-networkd[1044]: lxc_health: Gained IPv6LL Dec 13 14:10:08.781417 kubelet[1425]: E1213 14:10:08.781364 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:09.039788 kubelet[1425]: E1213 14:10:09.039266 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:09.782512 kubelet[1425]: E1213 14:10:09.782484 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:09.877115 systemd-networkd[1044]: lxcba7fc638d780: Gained IPv6LL Dec 13 14:10:10.040857 kubelet[1425]: E1213 14:10:10.040547 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:10.783582 kubelet[1425]: E1213 14:10:10.783543 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:11.542469 env[1212]: time="2024-12-13T14:10:11.542396845Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:10:11.542469 env[1212]: time="2024-12-13T14:10:11.542437862Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:10:11.542839 env[1212]: time="2024-12-13T14:10:11.542448326Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:10:11.542996 env[1212]: time="2024-12-13T14:10:11.542956032Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e8eaa97cb6f4bb5ee78cfb2d0daf9b0648a82e3eb32656111d2050d4641ca0ab pid=2490 runtime=io.containerd.runc.v2 Dec 13 14:10:11.555974 systemd[1]: Started cri-containerd-e8eaa97cb6f4bb5ee78cfb2d0daf9b0648a82e3eb32656111d2050d4641ca0ab.scope. Dec 13 14:10:11.616323 systemd-resolved[1157]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 14:10:11.630396 env[1212]: time="2024-12-13T14:10:11.630353032Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-85f456d6dd-6qhn4,Uid:231c0ef0-e192-4396-a0bc-d38605ed3f25,Namespace:default,Attempt:0,} returns sandbox id \"e8eaa97cb6f4bb5ee78cfb2d0daf9b0648a82e3eb32656111d2050d4641ca0ab\"" Dec 13 14:10:11.632178 env[1212]: time="2024-12-13T14:10:11.632140665Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 14:10:11.768371 kubelet[1425]: E1213 14:10:11.768332 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:11.784643 kubelet[1425]: E1213 14:10:11.784610 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:12.785238 kubelet[1425]: E1213 14:10:12.785184 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:13.785380 kubelet[1425]: E1213 14:10:13.785332 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:13.946139 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount799625370.mount: Deactivated successfully. Dec 13 14:10:14.786145 kubelet[1425]: E1213 14:10:14.786101 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:15.098262 env[1212]: time="2024-12-13T14:10:15.098160402Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:15.099786 env[1212]: time="2024-12-13T14:10:15.099750541Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:d5cb91e7550dca840aad69277b6dbccf8dc3739757998181746daf777a8bd9de,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:15.101699 env[1212]: time="2024-12-13T14:10:15.101661114Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:15.103261 env[1212]: time="2024-12-13T14:10:15.103231430Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:15.104098 env[1212]: time="2024-12-13T14:10:15.104073278Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:d5cb91e7550dca840aad69277b6dbccf8dc3739757998181746daf777a8bd9de\"" Dec 13 14:10:15.106585 env[1212]: time="2024-12-13T14:10:15.106533759Z" level=info msg="CreateContainer within sandbox \"e8eaa97cb6f4bb5ee78cfb2d0daf9b0648a82e3eb32656111d2050d4641ca0ab\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Dec 13 14:10:15.114870 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3268829111.mount: Deactivated successfully. Dec 13 14:10:15.118379 env[1212]: time="2024-12-13T14:10:15.118345604Z" level=info msg="CreateContainer within sandbox \"e8eaa97cb6f4bb5ee78cfb2d0daf9b0648a82e3eb32656111d2050d4641ca0ab\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"59eb2b803dc6876967268bb29325b69132c0dcbf07b2d019765d0414f0c62cca\"" Dec 13 14:10:15.118843 env[1212]: time="2024-12-13T14:10:15.118819780Z" level=info msg="StartContainer for \"59eb2b803dc6876967268bb29325b69132c0dcbf07b2d019765d0414f0c62cca\"" Dec 13 14:10:15.136035 systemd[1]: Started cri-containerd-59eb2b803dc6876967268bb29325b69132c0dcbf07b2d019765d0414f0c62cca.scope. Dec 13 14:10:15.171066 env[1212]: time="2024-12-13T14:10:15.171009702Z" level=info msg="StartContainer for \"59eb2b803dc6876967268bb29325b69132c0dcbf07b2d019765d0414f0c62cca\" returns successfully" Dec 13 14:10:15.786717 kubelet[1425]: E1213 14:10:15.786672 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:16.059963 kubelet[1425]: I1213 14:10:16.059828 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nginx-deployment-85f456d6dd-6qhn4" podStartSLOduration=5.58614196 podStartE2EDuration="9.05981199s" podCreationTimestamp="2024-12-13 14:10:07 +0000 UTC" firstStartedPulling="2024-12-13 14:10:11.631584273 +0000 UTC m=+20.642029223" lastFinishedPulling="2024-12-13 14:10:15.105254303 +0000 UTC m=+24.115699253" observedRunningTime="2024-12-13 14:10:16.059649077 +0000 UTC m=+25.070093987" watchObservedRunningTime="2024-12-13 14:10:16.05981199 +0000 UTC m=+25.070256940" Dec 13 14:10:16.787174 kubelet[1425]: E1213 14:10:16.787122 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:17.788120 kubelet[1425]: E1213 14:10:17.788063 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:18.788384 kubelet[1425]: E1213 14:10:18.788329 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:19.788974 kubelet[1425]: E1213 14:10:19.788943 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:19.820512 kubelet[1425]: I1213 14:10:19.820470 1425 topology_manager.go:215] "Topology Admit Handler" podUID="8c512d1b-c729-4d64-95ec-f8cc85b43f2c" podNamespace="default" podName="nfs-server-provisioner-0" Dec 13 14:10:19.825762 systemd[1]: Created slice kubepods-besteffort-pod8c512d1b_c729_4d64_95ec_f8cc85b43f2c.slice. Dec 13 14:10:19.845453 kubelet[1425]: I1213 14:10:19.845400 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hn47\" (UniqueName: \"kubernetes.io/projected/8c512d1b-c729-4d64-95ec-f8cc85b43f2c-kube-api-access-2hn47\") pod \"nfs-server-provisioner-0\" (UID: \"8c512d1b-c729-4d64-95ec-f8cc85b43f2c\") " pod="default/nfs-server-provisioner-0" Dec 13 14:10:19.845453 kubelet[1425]: I1213 14:10:19.845450 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/8c512d1b-c729-4d64-95ec-f8cc85b43f2c-data\") pod \"nfs-server-provisioner-0\" (UID: \"8c512d1b-c729-4d64-95ec-f8cc85b43f2c\") " pod="default/nfs-server-provisioner-0" Dec 13 14:10:20.128911 env[1212]: time="2024-12-13T14:10:20.128290347Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:8c512d1b-c729-4d64-95ec-f8cc85b43f2c,Namespace:default,Attempt:0,}" Dec 13 14:10:20.157416 systemd-networkd[1044]: lxcf2dbd95cfaa2: Link UP Dec 13 14:10:20.164912 kernel: eth0: renamed from tmpda4f5 Dec 13 14:10:20.171998 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Dec 13 14:10:20.172150 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcf2dbd95cfaa2: link becomes ready Dec 13 14:10:20.172375 systemd-networkd[1044]: lxcf2dbd95cfaa2: Gained carrier Dec 13 14:10:20.343764 env[1212]: time="2024-12-13T14:10:20.343694059Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:10:20.343764 env[1212]: time="2024-12-13T14:10:20.343734473Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:10:20.343764 env[1212]: time="2024-12-13T14:10:20.343744876Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:10:20.344137 env[1212]: time="2024-12-13T14:10:20.343928977Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/da4f5f09fca53d4599ed1ee55915ff35b91a53615c7a11c1491ea4a70a498227 pid=2626 runtime=io.containerd.runc.v2 Dec 13 14:10:20.356697 systemd[1]: Started cri-containerd-da4f5f09fca53d4599ed1ee55915ff35b91a53615c7a11c1491ea4a70a498227.scope. Dec 13 14:10:20.386376 systemd-resolved[1157]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 14:10:20.402544 env[1212]: time="2024-12-13T14:10:20.402482758Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:8c512d1b-c729-4d64-95ec-f8cc85b43f2c,Namespace:default,Attempt:0,} returns sandbox id \"da4f5f09fca53d4599ed1ee55915ff35b91a53615c7a11c1491ea4a70a498227\"" Dec 13 14:10:20.404330 env[1212]: time="2024-12-13T14:10:20.404299282Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Dec 13 14:10:20.790311 kubelet[1425]: E1213 14:10:20.789969 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:20.957654 systemd[1]: run-containerd-runc-k8s.io-da4f5f09fca53d4599ed1ee55915ff35b91a53615c7a11c1491ea4a70a498227-runc.ORcGed.mount: Deactivated successfully. Dec 13 14:10:21.653069 systemd-networkd[1044]: lxcf2dbd95cfaa2: Gained IPv6LL Dec 13 14:10:21.790858 kubelet[1425]: E1213 14:10:21.790812 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:22.576106 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount568682170.mount: Deactivated successfully. Dec 13 14:10:22.791079 kubelet[1425]: E1213 14:10:22.791029 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:23.791954 kubelet[1425]: E1213 14:10:23.791908 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:24.285666 env[1212]: time="2024-12-13T14:10:24.285350836Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:24.287213 env[1212]: time="2024-12-13T14:10:24.287189445Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:5a42a519e0a8cf95c3c5f18f767c58c8c8b072aaea0a26e5e47a6f206c7df685,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:24.292767 env[1212]: time="2024-12-13T14:10:24.292741803Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:24.294934 env[1212]: time="2024-12-13T14:10:24.294905499Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:24.295598 env[1212]: time="2024-12-13T14:10:24.295561033Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:5a42a519e0a8cf95c3c5f18f767c58c8c8b072aaea0a26e5e47a6f206c7df685\"" Dec 13 14:10:24.298367 env[1212]: time="2024-12-13T14:10:24.298335252Z" level=info msg="CreateContainer within sandbox \"da4f5f09fca53d4599ed1ee55915ff35b91a53615c7a11c1491ea4a70a498227\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Dec 13 14:10:24.309011 env[1212]: time="2024-12-13T14:10:24.308978325Z" level=info msg="CreateContainer within sandbox \"da4f5f09fca53d4599ed1ee55915ff35b91a53615c7a11c1491ea4a70a498227\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"aec0770b1b8c86a4b164c6879f429f44589d90996f5cd061520f465895b780e2\"" Dec 13 14:10:24.309418 env[1212]: time="2024-12-13T14:10:24.309376551Z" level=info msg="StartContainer for \"aec0770b1b8c86a4b164c6879f429f44589d90996f5cd061520f465895b780e2\"" Dec 13 14:10:24.325745 systemd[1]: run-containerd-runc-k8s.io-aec0770b1b8c86a4b164c6879f429f44589d90996f5cd061520f465895b780e2-runc.cublLo.mount: Deactivated successfully. Dec 13 14:10:24.329147 systemd[1]: Started cri-containerd-aec0770b1b8c86a4b164c6879f429f44589d90996f5cd061520f465895b780e2.scope. Dec 13 14:10:24.375274 env[1212]: time="2024-12-13T14:10:24.375233480Z" level=info msg="StartContainer for \"aec0770b1b8c86a4b164c6879f429f44589d90996f5cd061520f465895b780e2\" returns successfully" Dec 13 14:10:24.792992 kubelet[1425]: E1213 14:10:24.792949 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:25.082032 kubelet[1425]: I1213 14:10:25.081877 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=2.188623728 podStartE2EDuration="6.081848802s" podCreationTimestamp="2024-12-13 14:10:19 +0000 UTC" firstStartedPulling="2024-12-13 14:10:20.403697802 +0000 UTC m=+29.414142752" lastFinishedPulling="2024-12-13 14:10:24.296922876 +0000 UTC m=+33.307367826" observedRunningTime="2024-12-13 14:10:25.081649232 +0000 UTC m=+34.092094182" watchObservedRunningTime="2024-12-13 14:10:25.081848802 +0000 UTC m=+34.092293752" Dec 13 14:10:25.793627 kubelet[1425]: E1213 14:10:25.793579 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:26.794687 kubelet[1425]: E1213 14:10:26.794611 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:27.795704 kubelet[1425]: E1213 14:10:27.795649 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:28.796296 kubelet[1425]: E1213 14:10:28.796247 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:29.797058 kubelet[1425]: E1213 14:10:29.797023 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:30.017905 update_engine[1204]: I1213 14:10:30.017841 1204 update_attempter.cc:509] Updating boot flags... Dec 13 14:10:30.797932 kubelet[1425]: E1213 14:10:30.797864 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:31.768879 kubelet[1425]: E1213 14:10:31.768815 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:31.798264 kubelet[1425]: E1213 14:10:31.798216 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:32.798661 kubelet[1425]: E1213 14:10:32.798614 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:33.798901 kubelet[1425]: E1213 14:10:33.798850 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:33.854794 kubelet[1425]: I1213 14:10:33.854735 1425 topology_manager.go:215] "Topology Admit Handler" podUID="f931e38b-969e-4295-bc0c-1cd9ed4413d5" podNamespace="default" podName="test-pod-1" Dec 13 14:10:33.859270 systemd[1]: Created slice kubepods-besteffort-podf931e38b_969e_4295_bc0c_1cd9ed4413d5.slice. Dec 13 14:10:33.939799 kubelet[1425]: I1213 14:10:33.939767 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chlbv\" (UniqueName: \"kubernetes.io/projected/f931e38b-969e-4295-bc0c-1cd9ed4413d5-kube-api-access-chlbv\") pod \"test-pod-1\" (UID: \"f931e38b-969e-4295-bc0c-1cd9ed4413d5\") " pod="default/test-pod-1" Dec 13 14:10:33.940018 kubelet[1425]: I1213 14:10:33.939995 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-99f999e2-46bf-4f65-9ed5-9671d33ad79a\" (UniqueName: \"kubernetes.io/nfs/f931e38b-969e-4295-bc0c-1cd9ed4413d5-pvc-99f999e2-46bf-4f65-9ed5-9671d33ad79a\") pod \"test-pod-1\" (UID: \"f931e38b-969e-4295-bc0c-1cd9ed4413d5\") " pod="default/test-pod-1" Dec 13 14:10:34.062909 kernel: FS-Cache: Loaded Dec 13 14:10:34.091264 kernel: RPC: Registered named UNIX socket transport module. Dec 13 14:10:34.091353 kernel: RPC: Registered udp transport module. Dec 13 14:10:34.091374 kernel: RPC: Registered tcp transport module. Dec 13 14:10:34.091392 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Dec 13 14:10:34.132909 kernel: FS-Cache: Netfs 'nfs' registered for caching Dec 13 14:10:34.263939 kernel: NFS: Registering the id_resolver key type Dec 13 14:10:34.264096 kernel: Key type id_resolver registered Dec 13 14:10:34.264121 kernel: Key type id_legacy registered Dec 13 14:10:34.290591 nfsidmap[2757]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Dec 13 14:10:34.296627 nfsidmap[2760]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Dec 13 14:10:34.462740 env[1212]: time="2024-12-13T14:10:34.462687360Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:f931e38b-969e-4295-bc0c-1cd9ed4413d5,Namespace:default,Attempt:0,}" Dec 13 14:10:34.485248 systemd-networkd[1044]: lxc8b103dbd747b: Link UP Dec 13 14:10:34.495935 kernel: eth0: renamed from tmp1d094 Dec 13 14:10:34.506556 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Dec 13 14:10:34.506649 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc8b103dbd747b: link becomes ready Dec 13 14:10:34.506635 systemd-networkd[1044]: lxc8b103dbd747b: Gained carrier Dec 13 14:10:34.712733 env[1212]: time="2024-12-13T14:10:34.712656419Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:10:34.712733 env[1212]: time="2024-12-13T14:10:34.712696466Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:10:34.712733 env[1212]: time="2024-12-13T14:10:34.712706707Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:10:34.714054 env[1212]: time="2024-12-13T14:10:34.713161820Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/1d0942570b24cab5c10e25dcb4e9c890ba73e7b606441176b6a315c295f3405f pid=2793 runtime=io.containerd.runc.v2 Dec 13 14:10:34.722726 systemd[1]: Started cri-containerd-1d0942570b24cab5c10e25dcb4e9c890ba73e7b606441176b6a315c295f3405f.scope. Dec 13 14:10:34.751252 systemd-resolved[1157]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 14:10:34.768771 env[1212]: time="2024-12-13T14:10:34.768726351Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:f931e38b-969e-4295-bc0c-1cd9ed4413d5,Namespace:default,Attempt:0,} returns sandbox id \"1d0942570b24cab5c10e25dcb4e9c890ba73e7b606441176b6a315c295f3405f\"" Dec 13 14:10:34.770757 env[1212]: time="2024-12-13T14:10:34.770725789Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 14:10:34.799624 kubelet[1425]: E1213 14:10:34.799558 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:34.995837 env[1212]: time="2024-12-13T14:10:34.995724831Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:34.997217 env[1212]: time="2024-12-13T14:10:34.997189864Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:d5cb91e7550dca840aad69277b6dbccf8dc3739757998181746daf777a8bd9de,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:34.999166 env[1212]: time="2024-12-13T14:10:34.999132734Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:35.001722 env[1212]: time="2024-12-13T14:10:35.001689375Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:35.002547 env[1212]: time="2024-12-13T14:10:35.002113359Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:d5cb91e7550dca840aad69277b6dbccf8dc3739757998181746daf777a8bd9de\"" Dec 13 14:10:35.004783 env[1212]: time="2024-12-13T14:10:35.004751960Z" level=info msg="CreateContainer within sandbox \"1d0942570b24cab5c10e25dcb4e9c890ba73e7b606441176b6a315c295f3405f\" for container &ContainerMetadata{Name:test,Attempt:0,}" Dec 13 14:10:35.015756 env[1212]: time="2024-12-13T14:10:35.015708825Z" level=info msg="CreateContainer within sandbox \"1d0942570b24cab5c10e25dcb4e9c890ba73e7b606441176b6a315c295f3405f\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"423867f876056399556c9edcc61e254601e1a10a803592f67a17f575658459e2\"" Dec 13 14:10:35.016215 env[1212]: time="2024-12-13T14:10:35.016159694Z" level=info msg="StartContainer for \"423867f876056399556c9edcc61e254601e1a10a803592f67a17f575658459e2\"" Dec 13 14:10:35.029420 systemd[1]: Started cri-containerd-423867f876056399556c9edcc61e254601e1a10a803592f67a17f575658459e2.scope. Dec 13 14:10:35.074634 env[1212]: time="2024-12-13T14:10:35.074583290Z" level=info msg="StartContainer for \"423867f876056399556c9edcc61e254601e1a10a803592f67a17f575658459e2\" returns successfully" Dec 13 14:10:35.122384 kubelet[1425]: I1213 14:10:35.122258 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=14.888869735 podStartE2EDuration="15.12223749s" podCreationTimestamp="2024-12-13 14:10:20 +0000 UTC" firstStartedPulling="2024-12-13 14:10:34.770179742 +0000 UTC m=+43.780624692" lastFinishedPulling="2024-12-13 14:10:35.003547497 +0000 UTC m=+44.013992447" observedRunningTime="2024-12-13 14:10:35.118187914 +0000 UTC m=+44.128632864" watchObservedRunningTime="2024-12-13 14:10:35.12223749 +0000 UTC m=+44.132682440" Dec 13 14:10:35.801099 kubelet[1425]: E1213 14:10:35.801062 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:36.373168 systemd-networkd[1044]: lxc8b103dbd747b: Gained IPv6LL Dec 13 14:10:36.802178 kubelet[1425]: E1213 14:10:36.802120 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:37.803176 kubelet[1425]: E1213 14:10:37.803118 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:38.547462 systemd[1]: run-containerd-runc-k8s.io-b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445-runc.KlfKwa.mount: Deactivated successfully. Dec 13 14:10:38.595303 env[1212]: time="2024-12-13T14:10:38.595216562Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 14:10:38.600253 env[1212]: time="2024-12-13T14:10:38.600208463Z" level=info msg="StopContainer for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" with timeout 2 (s)" Dec 13 14:10:38.600513 env[1212]: time="2024-12-13T14:10:38.600481539Z" level=info msg="Stop container \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" with signal terminated" Dec 13 14:10:38.606026 systemd-networkd[1044]: lxc_health: Link DOWN Dec 13 14:10:38.606032 systemd-networkd[1044]: lxc_health: Lost carrier Dec 13 14:10:38.637296 systemd[1]: cri-containerd-b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445.scope: Deactivated successfully. Dec 13 14:10:38.637615 systemd[1]: cri-containerd-b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445.scope: Consumed 6.298s CPU time. Dec 13 14:10:38.653506 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445-rootfs.mount: Deactivated successfully. Dec 13 14:10:38.662858 env[1212]: time="2024-12-13T14:10:38.662812797Z" level=info msg="shim disconnected" id=b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445 Dec 13 14:10:38.662858 env[1212]: time="2024-12-13T14:10:38.662858843Z" level=warning msg="cleaning up after shim disconnected" id=b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445 namespace=k8s.io Dec 13 14:10:38.663092 env[1212]: time="2024-12-13T14:10:38.662868684Z" level=info msg="cleaning up dead shim" Dec 13 14:10:38.669862 env[1212]: time="2024-12-13T14:10:38.669804883Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:38Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2925 runtime=io.containerd.runc.v2\n" Dec 13 14:10:38.671846 env[1212]: time="2024-12-13T14:10:38.671802348Z" level=info msg="StopContainer for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" returns successfully" Dec 13 14:10:38.672429 env[1212]: time="2024-12-13T14:10:38.672397266Z" level=info msg="StopPodSandbox for \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\"" Dec 13 14:10:38.672562 env[1212]: time="2024-12-13T14:10:38.672541525Z" level=info msg="Container to stop \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:10:38.672634 env[1212]: time="2024-12-13T14:10:38.672617936Z" level=info msg="Container to stop \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:10:38.672698 env[1212]: time="2024-12-13T14:10:38.672682624Z" level=info msg="Container to stop \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:10:38.672752 env[1212]: time="2024-12-13T14:10:38.672738271Z" level=info msg="Container to stop \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:10:38.672813 env[1212]: time="2024-12-13T14:10:38.672798519Z" level=info msg="Container to stop \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:10:38.674508 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909-shm.mount: Deactivated successfully. Dec 13 14:10:38.679733 systemd[1]: cri-containerd-6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909.scope: Deactivated successfully. Dec 13 14:10:38.700852 env[1212]: time="2024-12-13T14:10:38.700789948Z" level=info msg="shim disconnected" id=6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909 Dec 13 14:10:38.700852 env[1212]: time="2024-12-13T14:10:38.700842315Z" level=warning msg="cleaning up after shim disconnected" id=6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909 namespace=k8s.io Dec 13 14:10:38.700852 env[1212]: time="2024-12-13T14:10:38.700850996Z" level=info msg="cleaning up dead shim" Dec 13 14:10:38.707748 env[1212]: time="2024-12-13T14:10:38.707709744Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:38Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2956 runtime=io.containerd.runc.v2\n" Dec 13 14:10:38.708049 env[1212]: time="2024-12-13T14:10:38.708022546Z" level=info msg="TearDown network for sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" successfully" Dec 13 14:10:38.708094 env[1212]: time="2024-12-13T14:10:38.708050869Z" level=info msg="StopPodSandbox for \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" returns successfully" Dec 13 14:10:38.770509 kubelet[1425]: I1213 14:10:38.770471 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-lib-modules\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770509 kubelet[1425]: I1213 14:10:38.770509 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-kernel\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770692 kubelet[1425]: I1213 14:10:38.770529 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-run\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770692 kubelet[1425]: I1213 14:10:38.770552 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-etc-cni-netd\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770692 kubelet[1425]: I1213 14:10:38.770574 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-config-path\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770692 kubelet[1425]: I1213 14:10:38.770587 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cni-path\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770692 kubelet[1425]: I1213 14:10:38.770580 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.770692 kubelet[1425]: I1213 14:10:38.770604 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shzrw\" (UniqueName: \"kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-kube-api-access-shzrw\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770829 kubelet[1425]: I1213 14:10:38.770626 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hostproc\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770829 kubelet[1425]: I1213 14:10:38.770641 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-bpf-maps\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770829 kubelet[1425]: I1213 14:10:38.770664 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/15a016f6-3ef3-4719-9ea1-8422641ba5a5-clustermesh-secrets\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770829 kubelet[1425]: I1213 14:10:38.770678 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-net\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770829 kubelet[1425]: I1213 14:10:38.770704 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-cgroup\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.770829 kubelet[1425]: I1213 14:10:38.770721 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hubble-tls\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.771480 kubelet[1425]: I1213 14:10:38.770735 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-xtables-lock\") pod \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\" (UID: \"15a016f6-3ef3-4719-9ea1-8422641ba5a5\") " Dec 13 14:10:38.771480 kubelet[1425]: I1213 14:10:38.770759 1425 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-lib-modules\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.771480 kubelet[1425]: I1213 14:10:38.770640 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771480 kubelet[1425]: I1213 14:10:38.770655 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771480 kubelet[1425]: I1213 14:10:38.770672 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771704 kubelet[1425]: I1213 14:10:38.770686 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hostproc" (OuterVolumeSpecName: "hostproc") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771704 kubelet[1425]: I1213 14:10:38.770798 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771704 kubelet[1425]: I1213 14:10:38.770815 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771704 kubelet[1425]: I1213 14:10:38.770950 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cni-path" (OuterVolumeSpecName: "cni-path") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771704 kubelet[1425]: I1213 14:10:38.770988 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.771826 kubelet[1425]: I1213 14:10:38.771009 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:38.773485 kubelet[1425]: I1213 14:10:38.773443 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 14:10:38.774161 kubelet[1425]: I1213 14:10:38.774125 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-kube-api-access-shzrw" (OuterVolumeSpecName: "kube-api-access-shzrw") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "kube-api-access-shzrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:10:38.774274 kubelet[1425]: I1213 14:10:38.774252 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15a016f6-3ef3-4719-9ea1-8422641ba5a5-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 14:10:38.776406 kubelet[1425]: I1213 14:10:38.776360 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "15a016f6-3ef3-4719-9ea1-8422641ba5a5" (UID: "15a016f6-3ef3-4719-9ea1-8422641ba5a5"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:10:38.803417 kubelet[1425]: E1213 14:10:38.803319 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:38.871897 kubelet[1425]: I1213 14:10:38.871842 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-run\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.871897 kubelet[1425]: I1213 14:10:38.871875 1425 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-etc-cni-netd\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.871897 kubelet[1425]: I1213 14:10:38.871899 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-config-path\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871910 1425 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cni-path\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871919 1425 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-shzrw\" (UniqueName: \"kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-kube-api-access-shzrw\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871929 1425 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hostproc\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871936 1425 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-bpf-maps\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871943 1425 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/15a016f6-3ef3-4719-9ea1-8422641ba5a5-clustermesh-secrets\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871950 1425 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-net\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871957 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-cilium-cgroup\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872114 kubelet[1425]: I1213 14:10:38.871965 1425 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/15a016f6-3ef3-4719-9ea1-8422641ba5a5-hubble-tls\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872294 kubelet[1425]: I1213 14:10:38.871972 1425 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-xtables-lock\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:38.872294 kubelet[1425]: I1213 14:10:38.871980 1425 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/15a016f6-3ef3-4719-9ea1-8422641ba5a5-host-proc-sys-kernel\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:39.098014 kubelet[1425]: I1213 14:10:39.097936 1425 scope.go:117] "RemoveContainer" containerID="b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445" Dec 13 14:10:39.099373 env[1212]: time="2024-12-13T14:10:39.099330697Z" level=info msg="RemoveContainer for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\"" Dec 13 14:10:39.102063 env[1212]: time="2024-12-13T14:10:39.102028479Z" level=info msg="RemoveContainer for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" returns successfully" Dec 13 14:10:39.102301 kubelet[1425]: I1213 14:10:39.102277 1425 scope.go:117] "RemoveContainer" containerID="0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a" Dec 13 14:10:39.102849 systemd[1]: Removed slice kubepods-burstable-pod15a016f6_3ef3_4719_9ea1_8422641ba5a5.slice. Dec 13 14:10:39.102956 systemd[1]: kubepods-burstable-pod15a016f6_3ef3_4719_9ea1_8422641ba5a5.slice: Consumed 6.509s CPU time. Dec 13 14:10:39.103606 env[1212]: time="2024-12-13T14:10:39.103574555Z" level=info msg="RemoveContainer for \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\"" Dec 13 14:10:39.105707 env[1212]: time="2024-12-13T14:10:39.105666100Z" level=info msg="RemoveContainer for \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\" returns successfully" Dec 13 14:10:39.105894 kubelet[1425]: I1213 14:10:39.105855 1425 scope.go:117] "RemoveContainer" containerID="faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312" Dec 13 14:10:39.106954 env[1212]: time="2024-12-13T14:10:39.106922219Z" level=info msg="RemoveContainer for \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\"" Dec 13 14:10:39.109282 env[1212]: time="2024-12-13T14:10:39.109250835Z" level=info msg="RemoveContainer for \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\" returns successfully" Dec 13 14:10:39.109437 kubelet[1425]: I1213 14:10:39.109416 1425 scope.go:117] "RemoveContainer" containerID="ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748" Dec 13 14:10:39.110452 env[1212]: time="2024-12-13T14:10:39.110418823Z" level=info msg="RemoveContainer for \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\"" Dec 13 14:10:39.112508 env[1212]: time="2024-12-13T14:10:39.112463282Z" level=info msg="RemoveContainer for \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\" returns successfully" Dec 13 14:10:39.112690 kubelet[1425]: I1213 14:10:39.112659 1425 scope.go:117] "RemoveContainer" containerID="ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98" Dec 13 14:10:39.113639 env[1212]: time="2024-12-13T14:10:39.113600186Z" level=info msg="RemoveContainer for \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\"" Dec 13 14:10:39.115619 env[1212]: time="2024-12-13T14:10:39.115576677Z" level=info msg="RemoveContainer for \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\" returns successfully" Dec 13 14:10:39.115797 kubelet[1425]: I1213 14:10:39.115763 1425 scope.go:117] "RemoveContainer" containerID="b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445" Dec 13 14:10:39.116055 env[1212]: time="2024-12-13T14:10:39.115985848Z" level=error msg="ContainerStatus for \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\": not found" Dec 13 14:10:39.116259 kubelet[1425]: E1213 14:10:39.116151 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\": not found" containerID="b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445" Dec 13 14:10:39.116259 kubelet[1425]: I1213 14:10:39.116190 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445"} err="failed to get container status \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\": rpc error: code = NotFound desc = an error occurred when try to find container \"b5daf93e11b5a69864c49623dc0cd136f90a96975871ee467b4e4fd52d58f445\": not found" Dec 13 14:10:39.116259 kubelet[1425]: I1213 14:10:39.116260 1425 scope.go:117] "RemoveContainer" containerID="0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a" Dec 13 14:10:39.116624 kubelet[1425]: E1213 14:10:39.116579 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\": not found" containerID="0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a" Dec 13 14:10:39.116660 env[1212]: time="2024-12-13T14:10:39.116429385Z" level=error msg="ContainerStatus for \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\": not found" Dec 13 14:10:39.116689 kubelet[1425]: I1213 14:10:39.116618 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a"} err="failed to get container status \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\": rpc error: code = NotFound desc = an error occurred when try to find container \"0e8e56e04a714ed449836270c8742a7a506a7ae77070301aa13c18bc8fbf159a\": not found" Dec 13 14:10:39.116689 kubelet[1425]: I1213 14:10:39.116637 1425 scope.go:117] "RemoveContainer" containerID="faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312" Dec 13 14:10:39.116858 env[1212]: time="2024-12-13T14:10:39.116803512Z" level=error msg="ContainerStatus for \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\": not found" Dec 13 14:10:39.117002 kubelet[1425]: E1213 14:10:39.116965 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\": not found" containerID="faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312" Dec 13 14:10:39.117042 kubelet[1425]: I1213 14:10:39.117007 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312"} err="failed to get container status \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\": rpc error: code = NotFound desc = an error occurred when try to find container \"faf1a71fd38fbfc92fc7534afdffe39f9bb3f845510ae683ceb134be29949312\": not found" Dec 13 14:10:39.117042 kubelet[1425]: I1213 14:10:39.117023 1425 scope.go:117] "RemoveContainer" containerID="ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748" Dec 13 14:10:39.117271 env[1212]: time="2024-12-13T14:10:39.117213484Z" level=error msg="ContainerStatus for \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\": not found" Dec 13 14:10:39.117382 kubelet[1425]: E1213 14:10:39.117364 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\": not found" containerID="ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748" Dec 13 14:10:39.117436 kubelet[1425]: I1213 14:10:39.117397 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748"} err="failed to get container status \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\": rpc error: code = NotFound desc = an error occurred when try to find container \"ecf5a55c868a4d52be6f7d5926f65da1bab23266b82a84be3329a88c819e2748\": not found" Dec 13 14:10:39.117436 kubelet[1425]: I1213 14:10:39.117421 1425 scope.go:117] "RemoveContainer" containerID="ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98" Dec 13 14:10:39.117610 env[1212]: time="2024-12-13T14:10:39.117567889Z" level=error msg="ContainerStatus for \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\": not found" Dec 13 14:10:39.117696 kubelet[1425]: E1213 14:10:39.117680 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\": not found" containerID="ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98" Dec 13 14:10:39.117733 kubelet[1425]: I1213 14:10:39.117700 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98"} err="failed to get container status \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\": rpc error: code = NotFound desc = an error occurred when try to find container \"ac6f4c484143f5552814d2a0df6e89743fec2ed1601b8706aecdf93794f2ee98\": not found" Dec 13 14:10:39.539231 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909-rootfs.mount: Deactivated successfully. Dec 13 14:10:39.539325 systemd[1]: var-lib-kubelet-pods-15a016f6\x2d3ef3\x2d4719\x2d9ea1\x2d8422641ba5a5-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dshzrw.mount: Deactivated successfully. Dec 13 14:10:39.539381 systemd[1]: var-lib-kubelet-pods-15a016f6\x2d3ef3\x2d4719\x2d9ea1\x2d8422641ba5a5-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 13 14:10:39.539439 systemd[1]: var-lib-kubelet-pods-15a016f6\x2d3ef3\x2d4719\x2d9ea1\x2d8422641ba5a5-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 13 14:10:39.803800 kubelet[1425]: E1213 14:10:39.803705 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:40.000231 kubelet[1425]: I1213 14:10:40.000192 1425 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" path="/var/lib/kubelet/pods/15a016f6-3ef3-4719-9ea1-8422641ba5a5/volumes" Dec 13 14:10:40.804222 kubelet[1425]: E1213 14:10:40.804180 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:41.628881 kubelet[1425]: I1213 14:10:41.628830 1425 topology_manager.go:215] "Topology Admit Handler" podUID="06922b4e-69be-4972-b8a3-c316b30c5a13" podNamespace="kube-system" podName="cilium-jbr9z" Dec 13 14:10:41.629021 kubelet[1425]: E1213 14:10:41.628900 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" containerName="mount-bpf-fs" Dec 13 14:10:41.629021 kubelet[1425]: E1213 14:10:41.628911 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" containerName="clean-cilium-state" Dec 13 14:10:41.629021 kubelet[1425]: E1213 14:10:41.628918 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" containerName="cilium-agent" Dec 13 14:10:41.629021 kubelet[1425]: E1213 14:10:41.628924 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" containerName="mount-cgroup" Dec 13 14:10:41.629021 kubelet[1425]: E1213 14:10:41.628930 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" containerName="apply-sysctl-overwrites" Dec 13 14:10:41.629021 kubelet[1425]: I1213 14:10:41.628952 1425 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a016f6-3ef3-4719-9ea1-8422641ba5a5" containerName="cilium-agent" Dec 13 14:10:41.634558 systemd[1]: Created slice kubepods-burstable-pod06922b4e_69be_4972_b8a3_c316b30c5a13.slice. Dec 13 14:10:41.640315 kubelet[1425]: I1213 14:10:41.640272 1425 topology_manager.go:215] "Topology Admit Handler" podUID="15590612-cba4-4d63-be34-40272ba49a51" podNamespace="kube-system" podName="cilium-operator-599987898-xrkwc" Dec 13 14:10:41.645312 systemd[1]: Created slice kubepods-besteffort-pod15590612_cba4_4d63_be34_40272ba49a51.slice. Dec 13 14:10:41.686321 kubelet[1425]: I1213 14:10:41.686277 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-config-path\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686321 kubelet[1425]: I1213 14:10:41.686321 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rxb\" (UniqueName: \"kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-kube-api-access-j8rxb\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686488 kubelet[1425]: I1213 14:10:41.686344 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/15590612-cba4-4d63-be34-40272ba49a51-cilium-config-path\") pod \"cilium-operator-599987898-xrkwc\" (UID: \"15590612-cba4-4d63-be34-40272ba49a51\") " pod="kube-system/cilium-operator-599987898-xrkwc" Dec 13 14:10:41.686488 kubelet[1425]: I1213 14:10:41.686363 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cni-path\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686488 kubelet[1425]: I1213 14:10:41.686381 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-xtables-lock\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686488 kubelet[1425]: I1213 14:10:41.686397 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-etc-cni-netd\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686488 kubelet[1425]: I1213 14:10:41.686412 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-clustermesh-secrets\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686600 kubelet[1425]: I1213 14:10:41.686428 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-kernel\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686600 kubelet[1425]: I1213 14:10:41.686442 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-hubble-tls\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686600 kubelet[1425]: I1213 14:10:41.686456 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-hostproc\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686600 kubelet[1425]: I1213 14:10:41.686472 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-cgroup\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686600 kubelet[1425]: I1213 14:10:41.686486 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqgwf\" (UniqueName: \"kubernetes.io/projected/15590612-cba4-4d63-be34-40272ba49a51-kube-api-access-sqgwf\") pod \"cilium-operator-599987898-xrkwc\" (UID: \"15590612-cba4-4d63-be34-40272ba49a51\") " pod="kube-system/cilium-operator-599987898-xrkwc" Dec 13 14:10:41.686701 kubelet[1425]: I1213 14:10:41.686500 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-lib-modules\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686701 kubelet[1425]: I1213 14:10:41.686515 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-ipsec-secrets\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686701 kubelet[1425]: I1213 14:10:41.686531 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-net\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686701 kubelet[1425]: I1213 14:10:41.686549 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-run\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.686701 kubelet[1425]: I1213 14:10:41.686564 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-bpf-maps\") pod \"cilium-jbr9z\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " pod="kube-system/cilium-jbr9z" Dec 13 14:10:41.778566 kubelet[1425]: E1213 14:10:41.778521 1425 pod_workers.go:1298] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-ipsec-secrets cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-j8rxb lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-jbr9z" podUID="06922b4e-69be-4972-b8a3-c316b30c5a13" Dec 13 14:10:41.804565 kubelet[1425]: E1213 14:10:41.804535 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:41.947273 kubelet[1425]: E1213 14:10:41.947239 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:41.947780 env[1212]: time="2024-12-13T14:10:41.947719284Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-xrkwc,Uid:15590612-cba4-4d63-be34-40272ba49a51,Namespace:kube-system,Attempt:0,}" Dec 13 14:10:41.959915 env[1212]: time="2024-12-13T14:10:41.959701479Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:10:41.959915 env[1212]: time="2024-12-13T14:10:41.959790009Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:10:41.959915 env[1212]: time="2024-12-13T14:10:41.959816492Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:10:41.960053 env[1212]: time="2024-12-13T14:10:41.959986992Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0f9a4af00b1c1510db7a8650f94ac70d3da6a0b65d6f5a51831187398677c174 pid=2987 runtime=io.containerd.runc.v2 Dec 13 14:10:41.969594 systemd[1]: Started cri-containerd-0f9a4af00b1c1510db7a8650f94ac70d3da6a0b65d6f5a51831187398677c174.scope. Dec 13 14:10:42.008671 kubelet[1425]: E1213 14:10:42.008623 1425 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 13 14:10:42.011153 env[1212]: time="2024-12-13T14:10:42.010693376Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-xrkwc,Uid:15590612-cba4-4d63-be34-40272ba49a51,Namespace:kube-system,Attempt:0,} returns sandbox id \"0f9a4af00b1c1510db7a8650f94ac70d3da6a0b65d6f5a51831187398677c174\"" Dec 13 14:10:42.011224 kubelet[1425]: E1213 14:10:42.011193 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:42.012113 env[1212]: time="2024-12-13T14:10:42.012077971Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Dec 13 14:10:42.189636 kubelet[1425]: I1213 14:10:42.189604 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cni-path\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.189917 kubelet[1425]: I1213 14:10:42.189701 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cni-path" (OuterVolumeSpecName: "cni-path") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190184 kubelet[1425]: I1213 14:10:42.189822 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-xtables-lock\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190237 kubelet[1425]: I1213 14:10:42.190198 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-hostproc\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190274 kubelet[1425]: I1213 14:10:42.190236 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-config-path\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190274 kubelet[1425]: I1213 14:10:42.190262 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8rxb\" (UniqueName: \"kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-kube-api-access-j8rxb\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190325 kubelet[1425]: I1213 14:10:42.190282 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-hubble-tls\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190325 kubelet[1425]: I1213 14:10:42.190301 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-lib-modules\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190325 kubelet[1425]: I1213 14:10:42.190322 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-net\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190393 kubelet[1425]: I1213 14:10:42.190338 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-bpf-maps\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190393 kubelet[1425]: I1213 14:10:42.190350 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190492 kubelet[1425]: I1213 14:10:42.190358 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-etc-cni-netd\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190525 kubelet[1425]: I1213 14:10:42.190501 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-cgroup\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190554 kubelet[1425]: I1213 14:10:42.190523 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-ipsec-secrets\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190554 kubelet[1425]: I1213 14:10:42.190541 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-run\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190600 kubelet[1425]: I1213 14:10:42.190562 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-clustermesh-secrets\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190600 kubelet[1425]: I1213 14:10:42.190578 1425 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-kernel\") pod \"06922b4e-69be-4972-b8a3-c316b30c5a13\" (UID: \"06922b4e-69be-4972-b8a3-c316b30c5a13\") " Dec 13 14:10:42.190645 kubelet[1425]: I1213 14:10:42.190629 1425 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-xtables-lock\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.190645 kubelet[1425]: I1213 14:10:42.190641 1425 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cni-path\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.190687 kubelet[1425]: I1213 14:10:42.190671 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190725 kubelet[1425]: I1213 14:10:42.190708 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190776 kubelet[1425]: I1213 14:10:42.190750 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190806 kubelet[1425]: I1213 14:10:42.190788 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190832 kubelet[1425]: I1213 14:10:42.190819 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190856 kubelet[1425]: I1213 14:10:42.190847 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.190951 kubelet[1425]: I1213 14:10:42.190925 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.191085 kubelet[1425]: I1213 14:10:42.191059 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-hostproc" (OuterVolumeSpecName: "hostproc") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:10:42.192750 kubelet[1425]: I1213 14:10:42.192700 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 14:10:42.196220 kubelet[1425]: I1213 14:10:42.196165 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-kube-api-access-j8rxb" (OuterVolumeSpecName: "kube-api-access-j8rxb") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "kube-api-access-j8rxb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:10:42.196220 kubelet[1425]: I1213 14:10:42.196193 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:10:42.196331 kubelet[1425]: I1213 14:10:42.196273 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 14:10:42.198403 kubelet[1425]: I1213 14:10:42.198322 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "06922b4e-69be-4972-b8a3-c316b30c5a13" (UID: "06922b4e-69be-4972-b8a3-c316b30c5a13"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 14:10:42.291138 kubelet[1425]: I1213 14:10:42.291101 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-ipsec-secrets\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291138 kubelet[1425]: I1213 14:10:42.291130 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-run\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291138 kubelet[1425]: I1213 14:10:42.291139 1425 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/06922b4e-69be-4972-b8a3-c316b30c5a13-clustermesh-secrets\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291168 1425 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-kernel\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291178 1425 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-hostproc\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291186 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-config-path\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291194 1425 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-j8rxb\" (UniqueName: \"kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-kube-api-access-j8rxb\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291201 1425 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/06922b4e-69be-4972-b8a3-c316b30c5a13-hubble-tls\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291209 1425 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-lib-modules\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291216 1425 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-host-proc-sys-net\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291311 kubelet[1425]: I1213 14:10:42.291223 1425 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-bpf-maps\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291483 kubelet[1425]: I1213 14:10:42.291230 1425 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-etc-cni-netd\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.291483 kubelet[1425]: I1213 14:10:42.291238 1425 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/06922b4e-69be-4972-b8a3-c316b30c5a13-cilium-cgroup\") on node \"10.0.0.79\" DevicePath \"\"" Dec 13 14:10:42.792686 systemd[1]: var-lib-kubelet-pods-06922b4e\x2d69be\x2d4972\x2db8a3\x2dc316b30c5a13-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dj8rxb.mount: Deactivated successfully. Dec 13 14:10:42.792778 systemd[1]: var-lib-kubelet-pods-06922b4e\x2d69be\x2d4972\x2db8a3\x2dc316b30c5a13-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 13 14:10:42.792828 systemd[1]: var-lib-kubelet-pods-06922b4e\x2d69be\x2d4972\x2db8a3\x2dc316b30c5a13-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 13 14:10:42.792897 systemd[1]: var-lib-kubelet-pods-06922b4e\x2d69be\x2d4972\x2db8a3\x2dc316b30c5a13-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Dec 13 14:10:42.805254 kubelet[1425]: E1213 14:10:42.805211 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:43.011515 kubelet[1425]: I1213 14:10:43.011450 1425 setters.go:580] "Node became not ready" node="10.0.0.79" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-12-13T14:10:43Z","lastTransitionTime":"2024-12-13T14:10:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Dec 13 14:10:43.089934 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3654909561.mount: Deactivated successfully. Dec 13 14:10:43.109493 systemd[1]: Removed slice kubepods-burstable-pod06922b4e_69be_4972_b8a3_c316b30c5a13.slice. Dec 13 14:10:43.135561 kubelet[1425]: I1213 14:10:43.135520 1425 topology_manager.go:215] "Topology Admit Handler" podUID="d267b82b-0cb5-4a10-875d-991aed40368f" podNamespace="kube-system" podName="cilium-jmg88" Dec 13 14:10:43.141066 systemd[1]: Created slice kubepods-burstable-podd267b82b_0cb5_4a10_875d_991aed40368f.slice. Dec 13 14:10:43.197293 kubelet[1425]: I1213 14:10:43.197237 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d267b82b-0cb5-4a10-875d-991aed40368f-hubble-tls\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197293 kubelet[1425]: I1213 14:10:43.197286 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-cilium-run\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197453 kubelet[1425]: I1213 14:10:43.197304 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-xtables-lock\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197453 kubelet[1425]: I1213 14:10:43.197321 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d267b82b-0cb5-4a10-875d-991aed40368f-clustermesh-secrets\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197453 kubelet[1425]: I1213 14:10:43.197339 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/d267b82b-0cb5-4a10-875d-991aed40368f-cilium-ipsec-secrets\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197453 kubelet[1425]: I1213 14:10:43.197353 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-bpf-maps\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197453 kubelet[1425]: I1213 14:10:43.197367 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-hostproc\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197453 kubelet[1425]: I1213 14:10:43.197382 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d267b82b-0cb5-4a10-875d-991aed40368f-cilium-config-path\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197591 kubelet[1425]: I1213 14:10:43.197422 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-etc-cni-netd\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197591 kubelet[1425]: I1213 14:10:43.197451 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-host-proc-sys-net\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197591 kubelet[1425]: I1213 14:10:43.197468 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-host-proc-sys-kernel\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197591 kubelet[1425]: I1213 14:10:43.197482 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phm4k\" (UniqueName: \"kubernetes.io/projected/d267b82b-0cb5-4a10-875d-991aed40368f-kube-api-access-phm4k\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197591 kubelet[1425]: I1213 14:10:43.197498 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-cilium-cgroup\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197591 kubelet[1425]: I1213 14:10:43.197513 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-cni-path\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.197735 kubelet[1425]: I1213 14:10:43.197528 1425 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d267b82b-0cb5-4a10-875d-991aed40368f-lib-modules\") pod \"cilium-jmg88\" (UID: \"d267b82b-0cb5-4a10-875d-991aed40368f\") " pod="kube-system/cilium-jmg88" Dec 13 14:10:43.452611 kubelet[1425]: E1213 14:10:43.452569 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:43.453169 env[1212]: time="2024-12-13T14:10:43.453121622Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jmg88,Uid:d267b82b-0cb5-4a10-875d-991aed40368f,Namespace:kube-system,Attempt:0,}" Dec 13 14:10:43.464166 env[1212]: time="2024-12-13T14:10:43.464085839Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:10:43.464166 env[1212]: time="2024-12-13T14:10:43.464133724Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:10:43.464166 env[1212]: time="2024-12-13T14:10:43.464143965Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:10:43.464949 env[1212]: time="2024-12-13T14:10:43.464910248Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5 pid=3038 runtime=io.containerd.runc.v2 Dec 13 14:10:43.474272 systemd[1]: Started cri-containerd-b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5.scope. Dec 13 14:10:43.516052 env[1212]: time="2024-12-13T14:10:43.516003093Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jmg88,Uid:d267b82b-0cb5-4a10-875d-991aed40368f,Namespace:kube-system,Attempt:0,} returns sandbox id \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\"" Dec 13 14:10:43.516590 kubelet[1425]: E1213 14:10:43.516558 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:43.518808 env[1212]: time="2024-12-13T14:10:43.518759469Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 13 14:10:43.571305 env[1212]: time="2024-12-13T14:10:43.571247504Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61\"" Dec 13 14:10:43.571923 env[1212]: time="2024-12-13T14:10:43.571877531Z" level=info msg="StartContainer for \"b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61\"" Dec 13 14:10:43.577149 env[1212]: time="2024-12-13T14:10:43.577112813Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:43.578619 env[1212]: time="2024-12-13T14:10:43.578575690Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:43.580471 env[1212]: time="2024-12-13T14:10:43.580437730Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:10:43.580679 env[1212]: time="2024-12-13T14:10:43.580655594Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Dec 13 14:10:43.584077 env[1212]: time="2024-12-13T14:10:43.584035636Z" level=info msg="CreateContainer within sandbox \"0f9a4af00b1c1510db7a8650f94ac70d3da6a0b65d6f5a51831187398677c174\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Dec 13 14:10:43.586559 systemd[1]: Started cri-containerd-b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61.scope. Dec 13 14:10:43.611287 env[1212]: time="2024-12-13T14:10:43.611224915Z" level=info msg="CreateContainer within sandbox \"0f9a4af00b1c1510db7a8650f94ac70d3da6a0b65d6f5a51831187398677c174\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"faea5a6fd495d512fbd974746b552431b6474df781db4fe1d798492a43893521\"" Dec 13 14:10:43.611982 env[1212]: time="2024-12-13T14:10:43.611943152Z" level=info msg="StartContainer for \"faea5a6fd495d512fbd974746b552431b6474df781db4fe1d798492a43893521\"" Dec 13 14:10:43.627508 systemd[1]: Started cri-containerd-faea5a6fd495d512fbd974746b552431b6474df781db4fe1d798492a43893521.scope. Dec 13 14:10:43.632342 env[1212]: time="2024-12-13T14:10:43.632298858Z" level=info msg="StartContainer for \"b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61\" returns successfully" Dec 13 14:10:43.649558 systemd[1]: cri-containerd-b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61.scope: Deactivated successfully. Dec 13 14:10:43.675461 env[1212]: time="2024-12-13T14:10:43.675394364Z" level=info msg="shim disconnected" id=b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61 Dec 13 14:10:43.675461 env[1212]: time="2024-12-13T14:10:43.675444290Z" level=warning msg="cleaning up after shim disconnected" id=b04c21052316a51f49c90a7dfc1f04d76960d946075eebe77fedd8e57bff9b61 namespace=k8s.io Dec 13 14:10:43.675461 env[1212]: time="2024-12-13T14:10:43.675453931Z" level=info msg="cleaning up dead shim" Dec 13 14:10:43.676402 env[1212]: time="2024-12-13T14:10:43.676359588Z" level=info msg="StartContainer for \"faea5a6fd495d512fbd974746b552431b6474df781db4fe1d798492a43893521\" returns successfully" Dec 13 14:10:43.682562 env[1212]: time="2024-12-13T14:10:43.682523890Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:43Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3155 runtime=io.containerd.runc.v2\n" Dec 13 14:10:43.806626 kubelet[1425]: E1213 14:10:43.806169 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:43.999616 kubelet[1425]: I1213 14:10:43.999566 1425 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06922b4e-69be-4972-b8a3-c316b30c5a13" path="/var/lib/kubelet/pods/06922b4e-69be-4972-b8a3-c316b30c5a13/volumes" Dec 13 14:10:44.109797 kubelet[1425]: E1213 14:10:44.109578 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:44.111535 kubelet[1425]: E1213 14:10:44.111490 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:44.111632 env[1212]: time="2024-12-13T14:10:44.111538411Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 13 14:10:44.125399 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3592758639.mount: Deactivated successfully. Dec 13 14:10:44.131999 env[1212]: time="2024-12-13T14:10:44.131946277Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea\"" Dec 13 14:10:44.132646 env[1212]: time="2024-12-13T14:10:44.132596985Z" level=info msg="StartContainer for \"856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea\"" Dec 13 14:10:44.133297 kubelet[1425]: I1213 14:10:44.133163 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-xrkwc" podStartSLOduration=1.5631924289999999 podStartE2EDuration="3.133147401s" podCreationTimestamp="2024-12-13 14:10:41 +0000 UTC" firstStartedPulling="2024-12-13 14:10:42.011769896 +0000 UTC m=+51.022214806" lastFinishedPulling="2024-12-13 14:10:43.581724868 +0000 UTC m=+52.592169778" observedRunningTime="2024-12-13 14:10:44.132313915 +0000 UTC m=+53.142758825" watchObservedRunningTime="2024-12-13 14:10:44.133147401 +0000 UTC m=+53.143592311" Dec 13 14:10:44.146121 systemd[1]: Started cri-containerd-856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea.scope. Dec 13 14:10:44.178987 env[1212]: time="2024-12-13T14:10:44.178932688Z" level=info msg="StartContainer for \"856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea\" returns successfully" Dec 13 14:10:44.188440 systemd[1]: cri-containerd-856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea.scope: Deactivated successfully. Dec 13 14:10:44.205866 env[1212]: time="2024-12-13T14:10:44.205793221Z" level=info msg="shim disconnected" id=856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea Dec 13 14:10:44.205866 env[1212]: time="2024-12-13T14:10:44.205842146Z" level=warning msg="cleaning up after shim disconnected" id=856137b46b628e71a233a4653a3d8eecc78914334574f9fef398a1d364edabea namespace=k8s.io Dec 13 14:10:44.205866 env[1212]: time="2024-12-13T14:10:44.205851067Z" level=info msg="cleaning up dead shim" Dec 13 14:10:44.212059 env[1212]: time="2024-12-13T14:10:44.212024864Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:44Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3226 runtime=io.containerd.runc.v2\n" Dec 13 14:10:44.806738 kubelet[1425]: E1213 14:10:44.806694 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:45.115717 kubelet[1425]: E1213 14:10:45.115012 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:45.115717 kubelet[1425]: E1213 14:10:45.115263 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:45.117094 env[1212]: time="2024-12-13T14:10:45.117043642Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 13 14:10:45.135087 env[1212]: time="2024-12-13T14:10:45.135026669Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165\"" Dec 13 14:10:45.136067 env[1212]: time="2024-12-13T14:10:45.136036409Z" level=info msg="StartContainer for \"1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165\"" Dec 13 14:10:45.153082 systemd[1]: Started cri-containerd-1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165.scope. Dec 13 14:10:45.185207 env[1212]: time="2024-12-13T14:10:45.185154130Z" level=info msg="StartContainer for \"1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165\" returns successfully" Dec 13 14:10:45.186814 systemd[1]: cri-containerd-1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165.scope: Deactivated successfully. Dec 13 14:10:45.207997 env[1212]: time="2024-12-13T14:10:45.207949035Z" level=info msg="shim disconnected" id=1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165 Dec 13 14:10:45.208173 env[1212]: time="2024-12-13T14:10:45.208002160Z" level=warning msg="cleaning up after shim disconnected" id=1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165 namespace=k8s.io Dec 13 14:10:45.208173 env[1212]: time="2024-12-13T14:10:45.208013201Z" level=info msg="cleaning up dead shim" Dec 13 14:10:45.216029 env[1212]: time="2024-12-13T14:10:45.215989274Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:45Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3284 runtime=io.containerd.runc.v2\n" Dec 13 14:10:45.795697 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1241465fd90c70f3e8fe001ebc269ee7fac6d1a43223567f068f7096dbe83165-rootfs.mount: Deactivated successfully. Dec 13 14:10:45.807863 kubelet[1425]: E1213 14:10:45.807814 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:46.119633 kubelet[1425]: E1213 14:10:46.118997 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:46.121001 env[1212]: time="2024-12-13T14:10:46.120964365Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 13 14:10:46.135179 env[1212]: time="2024-12-13T14:10:46.135128761Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b\"" Dec 13 14:10:46.135599 env[1212]: time="2024-12-13T14:10:46.135572003Z" level=info msg="StartContainer for \"f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b\"" Dec 13 14:10:46.151183 systemd[1]: Started cri-containerd-f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b.scope. Dec 13 14:10:46.179738 systemd[1]: cri-containerd-f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b.scope: Deactivated successfully. Dec 13 14:10:46.180809 env[1212]: time="2024-12-13T14:10:46.180772931Z" level=info msg="StartContainer for \"f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b\" returns successfully" Dec 13 14:10:46.197987 env[1212]: time="2024-12-13T14:10:46.197946135Z" level=info msg="shim disconnected" id=f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b Dec 13 14:10:46.198187 env[1212]: time="2024-12-13T14:10:46.198155436Z" level=warning msg="cleaning up after shim disconnected" id=f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b namespace=k8s.io Dec 13 14:10:46.198252 env[1212]: time="2024-12-13T14:10:46.198238844Z" level=info msg="cleaning up dead shim" Dec 13 14:10:46.204375 env[1212]: time="2024-12-13T14:10:46.204343188Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:10:46Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3337 runtime=io.containerd.runc.v2\n" Dec 13 14:10:46.795783 systemd[1]: run-containerd-runc-k8s.io-f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b-runc.3TJ3gk.mount: Deactivated successfully. Dec 13 14:10:46.795875 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f8565920f06de9169e964ff4c532b79d26b76f772e88c761996799497dffb49b-rootfs.mount: Deactivated successfully. Dec 13 14:10:46.808592 kubelet[1425]: E1213 14:10:46.808554 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:47.009927 kubelet[1425]: E1213 14:10:47.009879 1425 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 13 14:10:47.122725 kubelet[1425]: E1213 14:10:47.122623 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:47.125141 env[1212]: time="2024-12-13T14:10:47.125098772Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 13 14:10:47.142473 env[1212]: time="2024-12-13T14:10:47.142421052Z" level=info msg="CreateContainer within sandbox \"b8d4344ca4bfd3c0d21108a5278b2b3f8c4f923d1a259dc6fb497b45f532ead5\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f\"" Dec 13 14:10:47.143028 env[1212]: time="2024-12-13T14:10:47.142991105Z" level=info msg="StartContainer for \"6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f\"" Dec 13 14:10:47.158045 systemd[1]: Started cri-containerd-6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f.scope. Dec 13 14:10:47.190653 env[1212]: time="2024-12-13T14:10:47.190602022Z" level=info msg="StartContainer for \"6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f\" returns successfully" Dec 13 14:10:47.431943 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106(gcm-aes-ce))) Dec 13 14:10:47.809343 kubelet[1425]: E1213 14:10:47.809136 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:48.127720 kubelet[1425]: E1213 14:10:48.127618 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:48.140584 kubelet[1425]: I1213 14:10:48.140509 1425 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-jmg88" podStartSLOduration=5.140491778 podStartE2EDuration="5.140491778s" podCreationTimestamp="2024-12-13 14:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 14:10:48.140427812 +0000 UTC m=+57.150872762" watchObservedRunningTime="2024-12-13 14:10:48.140491778 +0000 UTC m=+57.150936728" Dec 13 14:10:48.810304 kubelet[1425]: E1213 14:10:48.810256 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:49.453842 kubelet[1425]: E1213 14:10:49.453800 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:49.810894 kubelet[1425]: E1213 14:10:49.810780 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:50.168484 systemd[1]: run-containerd-runc-k8s.io-6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f-runc.tyfB7f.mount: Deactivated successfully. Dec 13 14:10:50.275590 systemd-networkd[1044]: lxc_health: Link UP Dec 13 14:10:50.276935 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Dec 13 14:10:50.276970 systemd-networkd[1044]: lxc_health: Gained carrier Dec 13 14:10:50.811702 kubelet[1425]: E1213 14:10:50.811648 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:51.454440 kubelet[1425]: E1213 14:10:51.454386 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:51.477043 systemd-networkd[1044]: lxc_health: Gained IPv6LL Dec 13 14:10:51.768997 kubelet[1425]: E1213 14:10:51.768856 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:51.806418 env[1212]: time="2024-12-13T14:10:51.806376702Z" level=info msg="StopPodSandbox for \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\"" Dec 13 14:10:51.806748 env[1212]: time="2024-12-13T14:10:51.806459908Z" level=info msg="TearDown network for sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" successfully" Dec 13 14:10:51.806748 env[1212]: time="2024-12-13T14:10:51.806493151Z" level=info msg="StopPodSandbox for \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" returns successfully" Dec 13 14:10:51.806920 env[1212]: time="2024-12-13T14:10:51.806869822Z" level=info msg="RemovePodSandbox for \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\"" Dec 13 14:10:51.806960 env[1212]: time="2024-12-13T14:10:51.806927666Z" level=info msg="Forcibly stopping sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\"" Dec 13 14:10:51.807014 env[1212]: time="2024-12-13T14:10:51.806997032Z" level=info msg="TearDown network for sandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" successfully" Dec 13 14:10:51.810985 env[1212]: time="2024-12-13T14:10:51.810947071Z" level=info msg="RemovePodSandbox \"6a1e77c31ffe28b556a6560d9375ec667bf2136787dc79a9d0a24573f04fc909\" returns successfully" Dec 13 14:10:51.812217 kubelet[1425]: E1213 14:10:51.812198 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:52.137514 kubelet[1425]: E1213 14:10:52.137486 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:10:52.320036 systemd[1]: run-containerd-runc-k8s.io-6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f-runc.bUvd44.mount: Deactivated successfully. Dec 13 14:10:52.812755 kubelet[1425]: E1213 14:10:52.812718 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:53.813501 kubelet[1425]: E1213 14:10:53.813470 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:54.450539 systemd[1]: run-containerd-runc-k8s.io-6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f-runc.b9Uxkf.mount: Deactivated successfully. Dec 13 14:10:54.815316 kubelet[1425]: E1213 14:10:54.815024 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:55.816071 kubelet[1425]: E1213 14:10:55.816030 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:56.570596 systemd[1]: run-containerd-runc-k8s.io-6b17bf0d20f6a7388166cb475734e8909a671e2de679444ceaa8c8620fea577f-runc.kQGCOa.mount: Deactivated successfully. Dec 13 14:10:56.817202 kubelet[1425]: E1213 14:10:56.817137 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:10:57.817799 kubelet[1425]: E1213 14:10:57.817755 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"