Sep 6 00:33:37.125370 kernel: Linux version 5.15.190-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri Sep 5 22:53:38 -00 2025 Sep 6 00:33:37.125420 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=a807e3b6c1f608bcead7858f1ad5b6908e6d312e2d99c0ec0e5454f978e611a7 Sep 6 00:33:37.125438 kernel: BIOS-provided physical RAM map: Sep 6 00:33:37.125451 kernel: BIOS-e820: [mem 0x0000000000000000-0x0000000000000fff] reserved Sep 6 00:33:37.125463 kernel: BIOS-e820: [mem 0x0000000000001000-0x0000000000054fff] usable Sep 6 00:33:37.125476 kernel: BIOS-e820: [mem 0x0000000000055000-0x000000000005ffff] reserved Sep 6 00:33:37.125496 kernel: BIOS-e820: [mem 0x0000000000060000-0x0000000000097fff] usable Sep 6 00:33:37.125510 kernel: BIOS-e820: [mem 0x0000000000098000-0x000000000009ffff] reserved Sep 6 00:33:37.125523 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000bd27afff] usable Sep 6 00:33:37.125537 kernel: BIOS-e820: [mem 0x00000000bd27b000-0x00000000bd284fff] ACPI data Sep 6 00:33:37.125550 kernel: BIOS-e820: [mem 0x00000000bd285000-0x00000000bf8ecfff] usable Sep 6 00:33:37.125563 kernel: BIOS-e820: [mem 0x00000000bf8ed000-0x00000000bfb6cfff] reserved Sep 6 00:33:37.125577 kernel: BIOS-e820: [mem 0x00000000bfb6d000-0x00000000bfb7efff] ACPI data Sep 6 00:33:37.125590 kernel: BIOS-e820: [mem 0x00000000bfb7f000-0x00000000bfbfefff] ACPI NVS Sep 6 00:33:37.125610 kernel: BIOS-e820: [mem 0x00000000bfbff000-0x00000000bffdffff] usable Sep 6 00:33:37.125625 kernel: BIOS-e820: [mem 0x00000000bffe0000-0x00000000bfffffff] reserved Sep 6 00:33:37.125639 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000021fffffff] usable Sep 6 00:33:37.125654 kernel: NX (Execute Disable) protection: active Sep 6 00:33:37.125668 kernel: efi: EFI v2.70 by EDK II Sep 6 00:33:37.125683 kernel: efi: TPMFinalLog=0xbfbf7000 ACPI=0xbfb7e000 ACPI 2.0=0xbfb7e014 SMBIOS=0xbf9e8000 RNG=0xbfb73018 TPMEventLog=0xbd27b018 Sep 6 00:33:37.125698 kernel: random: crng init done Sep 6 00:33:37.125712 kernel: SMBIOS 2.4 present. Sep 6 00:33:37.125730 kernel: DMI: Google Google Compute Engine/Google Compute Engine, BIOS Google 08/14/2025 Sep 6 00:33:37.125744 kernel: Hypervisor detected: KVM Sep 6 00:33:37.125758 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Sep 6 00:33:37.125773 kernel: kvm-clock: cpu 0, msr 1e719f001, primary cpu clock Sep 6 00:33:37.125787 kernel: kvm-clock: using sched offset of 12918082040 cycles Sep 6 00:33:37.125802 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Sep 6 00:33:37.125817 kernel: tsc: Detected 2299.998 MHz processor Sep 6 00:33:37.125832 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Sep 6 00:33:37.125847 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Sep 6 00:33:37.125861 kernel: last_pfn = 0x220000 max_arch_pfn = 0x400000000 Sep 6 00:33:37.125879 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Sep 6 00:33:37.125893 kernel: last_pfn = 0xbffe0 max_arch_pfn = 0x400000000 Sep 6 00:33:37.125908 kernel: Using GB pages for direct mapping Sep 6 00:33:37.125922 kernel: Secure boot disabled Sep 6 00:33:37.125937 kernel: ACPI: Early table checksum verification disabled Sep 6 00:33:37.125951 kernel: ACPI: RSDP 0x00000000BFB7E014 000024 (v02 Google) Sep 6 00:33:37.125965 kernel: ACPI: XSDT 0x00000000BFB7D0E8 00005C (v01 Google GOOGFACP 00000001 01000013) Sep 6 00:33:37.125981 kernel: ACPI: FACP 0x00000000BFB78000 0000F4 (v02 Google GOOGFACP 00000001 GOOG 00000001) Sep 6 00:33:37.126005 kernel: ACPI: DSDT 0x00000000BFB79000 001A64 (v01 Google GOOGDSDT 00000001 GOOG 00000001) Sep 6 00:33:37.126020 kernel: ACPI: FACS 0x00000000BFBF2000 000040 Sep 6 00:33:37.126036 kernel: ACPI: SSDT 0x00000000BFB7C000 000316 (v02 GOOGLE Tpm2Tabl 00001000 INTL 20250404) Sep 6 00:33:37.126052 kernel: ACPI: TPM2 0x00000000BFB7B000 000034 (v04 GOOGLE 00000001 GOOG 00000001) Sep 6 00:33:37.126068 kernel: ACPI: SRAT 0x00000000BFB77000 0000C8 (v03 Google GOOGSRAT 00000001 GOOG 00000001) Sep 6 00:33:37.126084 kernel: ACPI: APIC 0x00000000BFB76000 000076 (v05 Google GOOGAPIC 00000001 GOOG 00000001) Sep 6 00:33:37.126102 kernel: ACPI: SSDT 0x00000000BFB75000 000980 (v01 Google GOOGSSDT 00000001 GOOG 00000001) Sep 6 00:33:37.126118 kernel: ACPI: WAET 0x00000000BFB74000 000028 (v01 Google GOOGWAET 00000001 GOOG 00000001) Sep 6 00:33:37.126134 kernel: ACPI: Reserving FACP table memory at [mem 0xbfb78000-0xbfb780f3] Sep 6 00:33:37.126149 kernel: ACPI: Reserving DSDT table memory at [mem 0xbfb79000-0xbfb7aa63] Sep 6 00:33:37.126165 kernel: ACPI: Reserving FACS table memory at [mem 0xbfbf2000-0xbfbf203f] Sep 6 00:33:37.126180 kernel: ACPI: Reserving SSDT table memory at [mem 0xbfb7c000-0xbfb7c315] Sep 6 00:33:37.126196 kernel: ACPI: Reserving TPM2 table memory at [mem 0xbfb7b000-0xbfb7b033] Sep 6 00:33:37.126219 kernel: ACPI: Reserving SRAT table memory at [mem 0xbfb77000-0xbfb770c7] Sep 6 00:33:37.126235 kernel: ACPI: Reserving APIC table memory at [mem 0xbfb76000-0xbfb76075] Sep 6 00:33:37.126254 kernel: ACPI: Reserving SSDT table memory at [mem 0xbfb75000-0xbfb7597f] Sep 6 00:33:37.126270 kernel: ACPI: Reserving WAET table memory at [mem 0xbfb74000-0xbfb74027] Sep 6 00:33:37.126285 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Sep 6 00:33:37.126300 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Sep 6 00:33:37.126316 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Sep 6 00:33:37.126332 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0xbfffffff] Sep 6 00:33:37.126348 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x21fffffff] Sep 6 00:33:37.126364 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0xbfffffff] -> [mem 0x00000000-0xbfffffff] Sep 6 00:33:37.126380 kernel: NUMA: Node 0 [mem 0x00000000-0xbfffffff] + [mem 0x100000000-0x21fffffff] -> [mem 0x00000000-0x21fffffff] Sep 6 00:33:37.126409 kernel: NODE_DATA(0) allocated [mem 0x21fffa000-0x21fffffff] Sep 6 00:33:37.126426 kernel: Zone ranges: Sep 6 00:33:37.126442 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Sep 6 00:33:37.126458 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Sep 6 00:33:37.126473 kernel: Normal [mem 0x0000000100000000-0x000000021fffffff] Sep 6 00:33:37.126489 kernel: Movable zone start for each node Sep 6 00:33:37.126505 kernel: Early memory node ranges Sep 6 00:33:37.126520 kernel: node 0: [mem 0x0000000000001000-0x0000000000054fff] Sep 6 00:33:37.126536 kernel: node 0: [mem 0x0000000000060000-0x0000000000097fff] Sep 6 00:33:37.126554 kernel: node 0: [mem 0x0000000000100000-0x00000000bd27afff] Sep 6 00:33:37.126570 kernel: node 0: [mem 0x00000000bd285000-0x00000000bf8ecfff] Sep 6 00:33:37.126586 kernel: node 0: [mem 0x00000000bfbff000-0x00000000bffdffff] Sep 6 00:33:37.126601 kernel: node 0: [mem 0x0000000100000000-0x000000021fffffff] Sep 6 00:33:37.126617 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000021fffffff] Sep 6 00:33:37.126632 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Sep 6 00:33:37.126648 kernel: On node 0, zone DMA: 11 pages in unavailable ranges Sep 6 00:33:37.126663 kernel: On node 0, zone DMA: 104 pages in unavailable ranges Sep 6 00:33:37.126679 kernel: On node 0, zone DMA32: 10 pages in unavailable ranges Sep 6 00:33:37.126698 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Sep 6 00:33:37.126714 kernel: On node 0, zone Normal: 32 pages in unavailable ranges Sep 6 00:33:37.126729 kernel: ACPI: PM-Timer IO Port: 0xb008 Sep 6 00:33:37.126745 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Sep 6 00:33:37.126761 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Sep 6 00:33:37.126776 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Sep 6 00:33:37.126792 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Sep 6 00:33:37.126808 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Sep 6 00:33:37.126823 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Sep 6 00:33:37.126842 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Sep 6 00:33:37.126857 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Sep 6 00:33:37.126873 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Sep 6 00:33:37.126889 kernel: Booting paravirtualized kernel on KVM Sep 6 00:33:37.126905 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Sep 6 00:33:37.126921 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Sep 6 00:33:37.126937 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 Sep 6 00:33:37.126953 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 Sep 6 00:33:37.126968 kernel: pcpu-alloc: [0] 0 1 Sep 6 00:33:37.126988 kernel: kvm-guest: PV spinlocks enabled Sep 6 00:33:37.127003 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Sep 6 00:33:37.127017 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1932270 Sep 6 00:33:37.127030 kernel: Policy zone: Normal Sep 6 00:33:37.127045 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=a807e3b6c1f608bcead7858f1ad5b6908e6d312e2d99c0ec0e5454f978e611a7 Sep 6 00:33:37.127060 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Sep 6 00:33:37.127074 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Sep 6 00:33:37.127089 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Sep 6 00:33:37.127102 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Sep 6 00:33:37.127122 kernel: Memory: 7515424K/7860544K available (12295K kernel code, 2276K rwdata, 13732K rodata, 47492K init, 4088K bss, 344860K reserved, 0K cma-reserved) Sep 6 00:33:37.127138 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Sep 6 00:33:37.127153 kernel: Kernel/User page tables isolation: enabled Sep 6 00:33:37.127168 kernel: ftrace: allocating 34612 entries in 136 pages Sep 6 00:33:37.127183 kernel: ftrace: allocated 136 pages with 2 groups Sep 6 00:33:37.127199 kernel: rcu: Hierarchical RCU implementation. Sep 6 00:33:37.127222 kernel: rcu: RCU event tracing is enabled. Sep 6 00:33:37.127237 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Sep 6 00:33:37.127258 kernel: Rude variant of Tasks RCU enabled. Sep 6 00:33:37.127288 kernel: Tracing variant of Tasks RCU enabled. Sep 6 00:33:37.127305 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Sep 6 00:33:37.127326 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Sep 6 00:33:37.127344 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Sep 6 00:33:37.127360 kernel: Console: colour dummy device 80x25 Sep 6 00:33:37.127378 kernel: printk: console [ttyS0] enabled Sep 6 00:33:37.127393 kernel: ACPI: Core revision 20210730 Sep 6 00:33:37.127423 kernel: APIC: Switch to symmetric I/O mode setup Sep 6 00:33:37.127439 kernel: x2apic enabled Sep 6 00:33:37.127458 kernel: Switched APIC routing to physical x2apic. Sep 6 00:33:37.127474 kernel: ..TIMER: vector=0x30 apic1=0 pin1=0 apic2=-1 pin2=-1 Sep 6 00:33:37.127491 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x212733415c7, max_idle_ns: 440795236380 ns Sep 6 00:33:37.127509 kernel: Calibrating delay loop (skipped) preset value.. 4599.99 BogoMIPS (lpj=2299998) Sep 6 00:33:37.127525 kernel: Last level iTLB entries: 4KB 1024, 2MB 1024, 4MB 1024 Sep 6 00:33:37.127542 kernel: Last level dTLB entries: 4KB 1024, 2MB 1024, 4MB 1024, 1GB 4 Sep 6 00:33:37.127560 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Sep 6 00:33:37.127581 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Sep 6 00:33:37.127599 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Sep 6 00:33:37.127616 kernel: Spectre V2 : Mitigation: IBRS Sep 6 00:33:37.127634 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT Sep 6 00:33:37.127651 kernel: RETBleed: Mitigation: IBRS Sep 6 00:33:37.127669 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Sep 6 00:33:37.127687 kernel: Spectre V2 : User space: Mitigation: STIBP via seccomp and prctl Sep 6 00:33:37.127705 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Sep 6 00:33:37.127722 kernel: MDS: Mitigation: Clear CPU buffers Sep 6 00:33:37.127745 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Sep 6 00:33:37.127763 kernel: active return thunk: its_return_thunk Sep 6 00:33:37.127781 kernel: ITS: Mitigation: Aligned branch/return thunks Sep 6 00:33:37.127797 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Sep 6 00:33:37.127812 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Sep 6 00:33:37.127828 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Sep 6 00:33:37.127846 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Sep 6 00:33:37.127864 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Sep 6 00:33:37.127880 kernel: Freeing SMP alternatives memory: 32K Sep 6 00:33:37.127899 kernel: pid_max: default: 32768 minimum: 301 Sep 6 00:33:37.127915 kernel: LSM: Security Framework initializing Sep 6 00:33:37.127931 kernel: SELinux: Initializing. Sep 6 00:33:37.127948 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 6 00:33:37.127963 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Sep 6 00:33:37.127980 kernel: smpboot: CPU0: Intel(R) Xeon(R) CPU @ 2.30GHz (family: 0x6, model: 0x3f, stepping: 0x0) Sep 6 00:33:37.135242 kernel: Performance Events: unsupported p6 CPU model 63 no PMU driver, software events only. Sep 6 00:33:37.135269 kernel: signal: max sigframe size: 1776 Sep 6 00:33:37.135287 kernel: rcu: Hierarchical SRCU implementation. Sep 6 00:33:37.135311 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Sep 6 00:33:37.135327 kernel: smp: Bringing up secondary CPUs ... Sep 6 00:33:37.135344 kernel: x86: Booting SMP configuration: Sep 6 00:33:37.135360 kernel: .... node #0, CPUs: #1 Sep 6 00:33:37.135376 kernel: kvm-clock: cpu 1, msr 1e719f041, secondary cpu clock Sep 6 00:33:37.135394 kernel: Transient Scheduler Attacks: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Sep 6 00:33:37.135430 kernel: Transient Scheduler Attacks: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Sep 6 00:33:37.135447 kernel: smp: Brought up 1 node, 2 CPUs Sep 6 00:33:37.135467 kernel: smpboot: Max logical packages: 1 Sep 6 00:33:37.135484 kernel: smpboot: Total of 2 processors activated (9199.99 BogoMIPS) Sep 6 00:33:37.135501 kernel: devtmpfs: initialized Sep 6 00:33:37.135517 kernel: x86/mm: Memory block size: 128MB Sep 6 00:33:37.135534 kernel: ACPI: PM: Registering ACPI NVS region [mem 0xbfb7f000-0xbfbfefff] (524288 bytes) Sep 6 00:33:37.135550 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Sep 6 00:33:37.135567 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Sep 6 00:33:37.135584 kernel: pinctrl core: initialized pinctrl subsystem Sep 6 00:33:37.135601 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Sep 6 00:33:37.135621 kernel: audit: initializing netlink subsys (disabled) Sep 6 00:33:37.135638 kernel: audit: type=2000 audit(1757118815.779:1): state=initialized audit_enabled=0 res=1 Sep 6 00:33:37.135655 kernel: thermal_sys: Registered thermal governor 'step_wise' Sep 6 00:33:37.135672 kernel: thermal_sys: Registered thermal governor 'user_space' Sep 6 00:33:37.135689 kernel: cpuidle: using governor menu Sep 6 00:33:37.135706 kernel: ACPI: bus type PCI registered Sep 6 00:33:37.135723 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Sep 6 00:33:37.135740 kernel: dca service started, version 1.12.1 Sep 6 00:33:37.135756 kernel: PCI: Using configuration type 1 for base access Sep 6 00:33:37.135777 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Sep 6 00:33:37.135794 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Sep 6 00:33:37.135810 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Sep 6 00:33:37.135826 kernel: ACPI: Added _OSI(Module Device) Sep 6 00:33:37.135844 kernel: ACPI: Added _OSI(Processor Device) Sep 6 00:33:37.135860 kernel: ACPI: Added _OSI(Processor Aggregator Device) Sep 6 00:33:37.135877 kernel: ACPI: Added _OSI(Linux-Dell-Video) Sep 6 00:33:37.135894 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Sep 6 00:33:37.135910 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Sep 6 00:33:37.135929 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Sep 6 00:33:37.135946 kernel: ACPI: Interpreter enabled Sep 6 00:33:37.135962 kernel: ACPI: PM: (supports S0 S3 S5) Sep 6 00:33:37.135979 kernel: ACPI: Using IOAPIC for interrupt routing Sep 6 00:33:37.135995 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Sep 6 00:33:37.136011 kernel: ACPI: Enabled 16 GPEs in block 00 to 0F Sep 6 00:33:37.136027 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Sep 6 00:33:37.136289 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Sep 6 00:33:37.136808 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. Sep 6 00:33:37.136834 kernel: PCI host bridge to bus 0000:00 Sep 6 00:33:37.137254 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Sep 6 00:33:37.137512 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Sep 6 00:33:37.137696 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Sep 6 00:33:37.138113 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfefff window] Sep 6 00:33:37.138310 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Sep 6 00:33:37.138545 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Sep 6 00:33:37.138731 kernel: pci 0000:00:01.0: [8086:7110] type 00 class 0x060100 Sep 6 00:33:37.138914 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Sep 6 00:33:37.139098 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Sep 6 00:33:37.139283 kernel: pci 0000:00:03.0: [1af4:1004] type 00 class 0x000000 Sep 6 00:33:37.139477 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc040-0xc07f] Sep 6 00:33:37.139666 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc0001000-0xc000107f] Sep 6 00:33:37.139847 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Sep 6 00:33:37.140030 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc03f] Sep 6 00:33:37.140202 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc0000000-0xc000007f] Sep 6 00:33:37.140391 kernel: pci 0000:00:05.0: [1af4:1005] type 00 class 0x00ff00 Sep 6 00:33:37.158025 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc080-0xc09f] Sep 6 00:33:37.158237 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xc0002000-0xc000203f] Sep 6 00:33:37.158282 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Sep 6 00:33:37.158303 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Sep 6 00:33:37.158322 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Sep 6 00:33:37.158339 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Sep 6 00:33:37.158357 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Sep 6 00:33:37.158374 kernel: iommu: Default domain type: Translated Sep 6 00:33:37.158392 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Sep 6 00:33:37.159547 kernel: vgaarb: loaded Sep 6 00:33:37.159566 kernel: pps_core: LinuxPPS API ver. 1 registered Sep 6 00:33:37.159591 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Sep 6 00:33:37.159608 kernel: PTP clock support registered Sep 6 00:33:37.159624 kernel: Registered efivars operations Sep 6 00:33:37.159641 kernel: PCI: Using ACPI for IRQ routing Sep 6 00:33:37.159964 kernel: PCI: pci_cache_line_size set to 64 bytes Sep 6 00:33:37.159983 kernel: e820: reserve RAM buffer [mem 0x00055000-0x0005ffff] Sep 6 00:33:37.160001 kernel: e820: reserve RAM buffer [mem 0x00098000-0x0009ffff] Sep 6 00:33:37.160017 kernel: e820: reserve RAM buffer [mem 0xbd27b000-0xbfffffff] Sep 6 00:33:37.160035 kernel: e820: reserve RAM buffer [mem 0xbf8ed000-0xbfffffff] Sep 6 00:33:37.160057 kernel: e820: reserve RAM buffer [mem 0xbffe0000-0xbfffffff] Sep 6 00:33:37.160074 kernel: clocksource: Switched to clocksource kvm-clock Sep 6 00:33:37.160092 kernel: VFS: Disk quotas dquot_6.6.0 Sep 6 00:33:37.160108 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Sep 6 00:33:37.160124 kernel: pnp: PnP ACPI init Sep 6 00:33:37.160140 kernel: pnp: PnP ACPI: found 7 devices Sep 6 00:33:37.160157 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Sep 6 00:33:37.160175 kernel: NET: Registered PF_INET protocol family Sep 6 00:33:37.160196 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Sep 6 00:33:37.160212 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Sep 6 00:33:37.160228 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Sep 6 00:33:37.160244 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Sep 6 00:33:37.160260 kernel: TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear) Sep 6 00:33:37.160277 kernel: TCP: Hash tables configured (established 65536 bind 65536) Sep 6 00:33:37.160294 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Sep 6 00:33:37.160312 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Sep 6 00:33:37.160330 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Sep 6 00:33:37.160352 kernel: NET: Registered PF_XDP protocol family Sep 6 00:33:37.166250 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Sep 6 00:33:37.166462 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Sep 6 00:33:37.166655 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Sep 6 00:33:37.166839 kernel: pci_bus 0000:00: resource 7 [mem 0xc0000000-0xfebfefff window] Sep 6 00:33:37.167025 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Sep 6 00:33:37.167049 kernel: PCI: CLS 0 bytes, default 64 Sep 6 00:33:37.167073 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Sep 6 00:33:37.167092 kernel: software IO TLB: mapped [mem 0x00000000b7f7f000-0x00000000bbf7f000] (64MB) Sep 6 00:33:37.167110 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Sep 6 00:33:37.167128 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x212733415c7, max_idle_ns: 440795236380 ns Sep 6 00:33:37.167146 kernel: clocksource: Switched to clocksource tsc Sep 6 00:33:37.167163 kernel: Initialise system trusted keyrings Sep 6 00:33:37.167181 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Sep 6 00:33:37.167198 kernel: Key type asymmetric registered Sep 6 00:33:37.167215 kernel: Asymmetric key parser 'x509' registered Sep 6 00:33:37.167236 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Sep 6 00:33:37.167254 kernel: io scheduler mq-deadline registered Sep 6 00:33:37.167272 kernel: io scheduler kyber registered Sep 6 00:33:37.167289 kernel: io scheduler bfq registered Sep 6 00:33:37.167306 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Sep 6 00:33:37.167325 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Sep 6 00:33:37.175161 kernel: virtio-pci 0000:00:03.0: virtio_pci: leaving for legacy driver Sep 6 00:33:37.175197 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 10 Sep 6 00:33:37.175378 kernel: virtio-pci 0000:00:04.0: virtio_pci: leaving for legacy driver Sep 6 00:33:37.175429 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Sep 6 00:33:37.175615 kernel: virtio-pci 0000:00:05.0: virtio_pci: leaving for legacy driver Sep 6 00:33:37.175639 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Sep 6 00:33:37.175658 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Sep 6 00:33:37.175674 kernel: 00:04: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Sep 6 00:33:37.175690 kernel: 00:05: ttyS2 at I/O 0x3e8 (irq = 6, base_baud = 115200) is a 16550A Sep 6 00:33:37.175707 kernel: 00:06: ttyS3 at I/O 0x2e8 (irq = 7, base_baud = 115200) is a 16550A Sep 6 00:33:37.175888 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x9009, rev-id 0) Sep 6 00:33:37.175920 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Sep 6 00:33:37.175936 kernel: i8042: Warning: Keylock active Sep 6 00:33:37.175952 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Sep 6 00:33:37.175969 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Sep 6 00:33:37.176247 kernel: rtc_cmos 00:00: RTC can wake from S4 Sep 6 00:33:37.184505 kernel: rtc_cmos 00:00: registered as rtc0 Sep 6 00:33:37.184687 kernel: rtc_cmos 00:00: setting system clock to 2025-09-06T00:33:36 UTC (1757118816) Sep 6 00:33:37.184846 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Sep 6 00:33:37.184875 kernel: intel_pstate: CPU model not supported Sep 6 00:33:37.184893 kernel: pstore: Registered efi as persistent store backend Sep 6 00:33:37.184911 kernel: NET: Registered PF_INET6 protocol family Sep 6 00:33:37.184930 kernel: Segment Routing with IPv6 Sep 6 00:33:37.184946 kernel: In-situ OAM (IOAM) with IPv6 Sep 6 00:33:37.184976 kernel: NET: Registered PF_PACKET protocol family Sep 6 00:33:37.184994 kernel: Key type dns_resolver registered Sep 6 00:33:37.185011 kernel: IPI shorthand broadcast: enabled Sep 6 00:33:37.185029 kernel: sched_clock: Marking stable (754778937, 161226154)->(949467212, -33462121) Sep 6 00:33:37.185050 kernel: registered taskstats version 1 Sep 6 00:33:37.185068 kernel: Loading compiled-in X.509 certificates Sep 6 00:33:37.185086 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Sep 6 00:33:37.185104 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.190-flatcar: 59a3efd48c75422889eb056cb9758fbe471623cb' Sep 6 00:33:37.185122 kernel: Key type .fscrypt registered Sep 6 00:33:37.185140 kernel: Key type fscrypt-provisioning registered Sep 6 00:33:37.185158 kernel: pstore: Using crash dump compression: deflate Sep 6 00:33:37.185176 kernel: ima: Allocated hash algorithm: sha1 Sep 6 00:33:37.185193 kernel: ima: No architecture policies found Sep 6 00:33:37.185212 kernel: clk: Disabling unused clocks Sep 6 00:33:37.185229 kernel: Freeing unused kernel image (initmem) memory: 47492K Sep 6 00:33:37.185246 kernel: Write protecting the kernel read-only data: 28672k Sep 6 00:33:37.185264 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Sep 6 00:33:37.185282 kernel: Freeing unused kernel image (rodata/data gap) memory: 604K Sep 6 00:33:37.185298 kernel: Run /init as init process Sep 6 00:33:37.185316 kernel: with arguments: Sep 6 00:33:37.185333 kernel: /init Sep 6 00:33:37.185349 kernel: with environment: Sep 6 00:33:37.185370 kernel: HOME=/ Sep 6 00:33:37.185388 kernel: TERM=linux Sep 6 00:33:37.185428 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Sep 6 00:33:37.185449 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 6 00:33:37.185470 systemd[1]: Detected virtualization kvm. Sep 6 00:33:37.185489 systemd[1]: Detected architecture x86-64. Sep 6 00:33:37.185505 systemd[1]: Running in initrd. Sep 6 00:33:37.185526 systemd[1]: No hostname configured, using default hostname. Sep 6 00:33:37.185542 systemd[1]: Hostname set to . Sep 6 00:33:37.185560 systemd[1]: Initializing machine ID from VM UUID. Sep 6 00:33:37.185577 systemd[1]: Queued start job for default target initrd.target. Sep 6 00:33:37.185594 systemd[1]: Started systemd-ask-password-console.path. Sep 6 00:33:37.185612 systemd[1]: Reached target cryptsetup.target. Sep 6 00:33:37.185627 systemd[1]: Reached target paths.target. Sep 6 00:33:37.185644 systemd[1]: Reached target slices.target. Sep 6 00:33:37.185665 systemd[1]: Reached target swap.target. Sep 6 00:33:37.185684 systemd[1]: Reached target timers.target. Sep 6 00:33:37.185703 systemd[1]: Listening on iscsid.socket. Sep 6 00:33:37.185720 systemd[1]: Listening on iscsiuio.socket. Sep 6 00:33:37.185737 systemd[1]: Listening on systemd-journald-audit.socket. Sep 6 00:33:37.185755 systemd[1]: Listening on systemd-journald-dev-log.socket. Sep 6 00:33:37.185772 systemd[1]: Listening on systemd-journald.socket. Sep 6 00:33:37.185790 systemd[1]: Listening on systemd-networkd.socket. Sep 6 00:33:37.185812 systemd[1]: Listening on systemd-udevd-control.socket. Sep 6 00:33:37.185832 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 6 00:33:37.185869 systemd[1]: Reached target sockets.target. Sep 6 00:33:37.185891 systemd[1]: Starting kmod-static-nodes.service... Sep 6 00:33:37.185910 systemd[1]: Finished network-cleanup.service. Sep 6 00:33:37.185928 systemd[1]: Starting systemd-fsck-usr.service... Sep 6 00:33:37.185968 systemd[1]: Starting systemd-journald.service... Sep 6 00:33:37.185987 systemd[1]: Starting systemd-modules-load.service... Sep 6 00:33:37.186005 systemd[1]: Starting systemd-resolved.service... Sep 6 00:33:37.186024 systemd[1]: Starting systemd-vconsole-setup.service... Sep 6 00:33:37.186042 systemd[1]: Finished kmod-static-nodes.service. Sep 6 00:33:37.186061 kernel: audit: type=1130 audit(1757118817.129:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.186080 systemd[1]: Finished systemd-fsck-usr.service. Sep 6 00:33:37.186099 systemd[1]: Finished systemd-vconsole-setup.service. Sep 6 00:33:37.186118 kernel: audit: type=1130 audit(1757118817.139:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.186140 kernel: audit: type=1130 audit(1757118817.149:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.186159 systemd[1]: Starting dracut-cmdline-ask.service... Sep 6 00:33:37.186178 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 6 00:33:37.186202 systemd-journald[190]: Journal started Sep 6 00:33:37.186298 systemd-journald[190]: Runtime Journal (/run/log/journal/54d118c5b4df6aa1e26d8995714142bc) is 8.0M, max 148.8M, 140.8M free. Sep 6 00:33:37.197258 systemd[1]: Started systemd-journald.service. Sep 6 00:33:37.197307 kernel: audit: type=1130 audit(1757118817.186:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.197333 kernel: audit: type=1130 audit(1757118817.195:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.139000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.149000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.186000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.195000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.146657 systemd-modules-load[191]: Inserted module 'overlay' Sep 6 00:33:37.228551 kernel: audit: type=1130 audit(1757118817.220:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.220000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.188605 systemd-resolved[192]: Positive Trust Anchors: Sep 6 00:33:37.188618 systemd-resolved[192]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 6 00:33:37.188682 systemd-resolved[192]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 6 00:33:37.249565 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Sep 6 00:33:37.194337 systemd-resolved[192]: Defaulting to hostname 'linux'. Sep 6 00:33:37.254512 kernel: Bridge firewalling registered Sep 6 00:33:37.196867 systemd[1]: Started systemd-resolved.service. Sep 6 00:33:37.197166 systemd[1]: Reached target nss-lookup.target. Sep 6 00:33:37.218069 systemd[1]: Finished dracut-cmdline-ask.service. Sep 6 00:33:37.225053 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 6 00:33:37.251738 systemd-modules-load[191]: Inserted module 'br_netfilter' Sep 6 00:33:37.283680 kernel: audit: type=1130 audit(1757118817.275:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.283718 kernel: SCSI subsystem initialized Sep 6 00:33:37.275000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.284015 systemd[1]: Starting dracut-cmdline.service... Sep 6 00:33:37.301979 dracut-cmdline[206]: dracut-dracut-053 Sep 6 00:33:37.307202 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Sep 6 00:33:37.307253 kernel: device-mapper: uevent: version 1.0.3 Sep 6 00:33:37.307876 dracut-cmdline[206]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=a807e3b6c1f608bcead7858f1ad5b6908e6d312e2d99c0ec0e5454f978e611a7 Sep 6 00:33:37.315203 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Sep 6 00:33:37.315350 systemd-modules-load[191]: Inserted module 'dm_multipath' Sep 6 00:33:37.317200 systemd[1]: Finished systemd-modules-load.service. Sep 6 00:33:37.325000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.327752 systemd[1]: Starting systemd-sysctl.service... Sep 6 00:33:37.336523 kernel: audit: type=1130 audit(1757118817.325:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.342921 systemd[1]: Finished systemd-sysctl.service. Sep 6 00:33:37.353552 kernel: audit: type=1130 audit(1757118817.345:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.345000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.405435 kernel: Loading iSCSI transport class v2.0-870. Sep 6 00:33:37.427431 kernel: iscsi: registered transport (tcp) Sep 6 00:33:37.455757 kernel: iscsi: registered transport (qla4xxx) Sep 6 00:33:37.455834 kernel: QLogic iSCSI HBA Driver Sep 6 00:33:37.505206 systemd[1]: Finished dracut-cmdline.service. Sep 6 00:33:37.507000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.510237 systemd[1]: Starting dracut-pre-udev.service... Sep 6 00:33:37.569473 kernel: raid6: avx2x4 gen() 18162 MB/s Sep 6 00:33:37.586441 kernel: raid6: avx2x4 xor() 8288 MB/s Sep 6 00:33:37.604450 kernel: raid6: avx2x2 gen() 18180 MB/s Sep 6 00:33:37.622440 kernel: raid6: avx2x2 xor() 18563 MB/s Sep 6 00:33:37.639468 kernel: raid6: avx2x1 gen() 14039 MB/s Sep 6 00:33:37.656445 kernel: raid6: avx2x1 xor() 16114 MB/s Sep 6 00:33:37.673496 kernel: raid6: sse2x4 gen() 11080 MB/s Sep 6 00:33:37.690452 kernel: raid6: sse2x4 xor() 6500 MB/s Sep 6 00:33:37.707467 kernel: raid6: sse2x2 gen() 11819 MB/s Sep 6 00:33:37.724500 kernel: raid6: sse2x2 xor() 7390 MB/s Sep 6 00:33:37.741474 kernel: raid6: sse2x1 gen() 9622 MB/s Sep 6 00:33:37.759785 kernel: raid6: sse2x1 xor() 5156 MB/s Sep 6 00:33:37.759855 kernel: raid6: using algorithm avx2x2 gen() 18180 MB/s Sep 6 00:33:37.759880 kernel: raid6: .... xor() 18563 MB/s, rmw enabled Sep 6 00:33:37.760698 kernel: raid6: using avx2x2 recovery algorithm Sep 6 00:33:37.776441 kernel: xor: automatically using best checksumming function avx Sep 6 00:33:37.883453 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Sep 6 00:33:37.895988 systemd[1]: Finished dracut-pre-udev.service. Sep 6 00:33:37.899000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.899000 audit: BPF prog-id=7 op=LOAD Sep 6 00:33:37.899000 audit: BPF prog-id=8 op=LOAD Sep 6 00:33:37.901939 systemd[1]: Starting systemd-udevd.service... Sep 6 00:33:37.919030 systemd-udevd[388]: Using default interface naming scheme 'v252'. Sep 6 00:33:37.926147 systemd[1]: Started systemd-udevd.service. Sep 6 00:33:37.928000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.930792 systemd[1]: Starting dracut-pre-trigger.service... Sep 6 00:33:37.948890 dracut-pre-trigger[393]: rd.md=0: removing MD RAID activation Sep 6 00:33:37.983854 systemd[1]: Finished dracut-pre-trigger.service. Sep 6 00:33:37.982000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:37.985209 systemd[1]: Starting systemd-udev-trigger.service... Sep 6 00:33:38.051384 systemd[1]: Finished systemd-udev-trigger.service. Sep 6 00:33:38.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:38.129424 kernel: scsi host0: Virtio SCSI HBA Sep 6 00:33:38.136422 kernel: scsi 0:0:1:0: Direct-Access Google PersistentDisk 1 PQ: 0 ANSI: 6 Sep 6 00:33:38.143448 kernel: cryptd: max_cpu_qlen set to 1000 Sep 6 00:33:38.206444 kernel: AVX2 version of gcm_enc/dec engaged. Sep 6 00:33:38.212439 kernel: AES CTR mode by8 optimization enabled Sep 6 00:33:38.270215 kernel: sd 0:0:1:0: [sda] 25165824 512-byte logical blocks: (12.9 GB/12.0 GiB) Sep 6 00:33:38.329211 kernel: sd 0:0:1:0: [sda] 4096-byte physical blocks Sep 6 00:33:38.329483 kernel: sd 0:0:1:0: [sda] Write Protect is off Sep 6 00:33:38.329697 kernel: sd 0:0:1:0: [sda] Mode Sense: 1f 00 00 08 Sep 6 00:33:38.329926 kernel: sd 0:0:1:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Sep 6 00:33:38.330162 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Sep 6 00:33:38.330205 kernel: GPT:17805311 != 25165823 Sep 6 00:33:38.330228 kernel: GPT:Alternate GPT header not at the end of the disk. Sep 6 00:33:38.330250 kernel: GPT:17805311 != 25165823 Sep 6 00:33:38.330273 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 6 00:33:38.330294 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 6 00:33:38.330318 kernel: sd 0:0:1:0: [sda] Attached SCSI disk Sep 6 00:33:38.394424 kernel: BTRFS: device label OEM devid 1 transid 10 /dev/sda6 scanned by (udev-worker) (429) Sep 6 00:33:38.396834 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Sep 6 00:33:38.419798 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Sep 6 00:33:38.429561 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Sep 6 00:33:38.452826 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Sep 6 00:33:38.479883 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 6 00:33:38.490612 systemd[1]: Starting disk-uuid.service... Sep 6 00:33:38.516537 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 6 00:33:38.516741 disk-uuid[506]: Primary Header is updated. Sep 6 00:33:38.516741 disk-uuid[506]: Secondary Entries is updated. Sep 6 00:33:38.516741 disk-uuid[506]: Secondary Header is updated. Sep 6 00:33:38.550522 kernel: GPT:disk_guids don't match. Sep 6 00:33:38.550549 kernel: GPT: Use GNU Parted to correct GPT errors. Sep 6 00:33:38.550564 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 6 00:33:38.563421 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 6 00:33:39.554244 disk-uuid[507]: The operation has completed successfully. Sep 6 00:33:39.562551 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Sep 6 00:33:39.619770 systemd[1]: disk-uuid.service: Deactivated successfully. Sep 6 00:33:39.626000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:39.626000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:39.619928 systemd[1]: Finished disk-uuid.service. Sep 6 00:33:39.643119 systemd[1]: Starting verity-setup.service... Sep 6 00:33:39.671787 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Sep 6 00:33:39.752884 systemd[1]: Found device dev-mapper-usr.device. Sep 6 00:33:39.755565 systemd[1]: Mounting sysusr-usr.mount... Sep 6 00:33:39.767143 systemd[1]: Finished verity-setup.service. Sep 6 00:33:39.793000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:39.861463 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Sep 6 00:33:39.862507 systemd[1]: Mounted sysusr-usr.mount. Sep 6 00:33:39.862871 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Sep 6 00:33:39.864076 systemd[1]: Starting ignition-setup.service... Sep 6 00:33:39.924582 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 6 00:33:39.924625 kernel: BTRFS info (device sda6): using free space tree Sep 6 00:33:39.924657 kernel: BTRFS info (device sda6): has skinny extents Sep 6 00:33:39.924687 kernel: BTRFS info (device sda6): enabling ssd optimizations Sep 6 00:33:39.917990 systemd[1]: Starting parse-ip-for-networkd.service... Sep 6 00:33:39.961103 systemd[1]: Finished ignition-setup.service. Sep 6 00:33:39.959000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:39.962892 systemd[1]: Starting ignition-fetch-offline.service... Sep 6 00:33:40.039765 systemd[1]: Finished parse-ip-for-networkd.service. Sep 6 00:33:40.038000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.039000 audit: BPF prog-id=9 op=LOAD Sep 6 00:33:40.042293 systemd[1]: Starting systemd-networkd.service... Sep 6 00:33:40.079829 systemd-networkd[681]: lo: Link UP Sep 6 00:33:40.079843 systemd-networkd[681]: lo: Gained carrier Sep 6 00:33:40.086000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.080938 systemd-networkd[681]: Enumeration completed Sep 6 00:33:40.081057 systemd[1]: Started systemd-networkd.service. Sep 6 00:33:40.081467 systemd-networkd[681]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 6 00:33:40.084247 systemd-networkd[681]: eth0: Link UP Sep 6 00:33:40.084255 systemd-networkd[681]: eth0: Gained carrier Sep 6 00:33:40.087883 systemd[1]: Reached target network.target. Sep 6 00:33:40.094528 systemd-networkd[681]: eth0: Overlong DHCP hostname received, shortened from 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6.c.flatcar-212911.internal' to 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' Sep 6 00:33:40.094543 systemd-networkd[681]: eth0: DHCPv4 address 10.128.0.94/32, gateway 10.128.0.1 acquired from 169.254.169.254 Sep 6 00:33:40.109713 systemd[1]: Starting iscsiuio.service... Sep 6 00:33:40.199924 systemd[1]: Started iscsiuio.service. Sep 6 00:33:40.219000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.222120 systemd[1]: Starting iscsid.service... Sep 6 00:33:40.224141 ignition[609]: Ignition 2.14.0 Sep 6 00:33:40.240000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.232090 systemd[1]: Started iscsid.service. Sep 6 00:33:40.255550 iscsid[691]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Sep 6 00:33:40.255550 iscsid[691]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log Sep 6 00:33:40.255550 iscsid[691]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Sep 6 00:33:40.255550 iscsid[691]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Sep 6 00:33:40.255550 iscsid[691]: If using hardware iscsi like qla4xxx this message can be ignored. Sep 6 00:33:40.255550 iscsid[691]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Sep 6 00:33:40.255550 iscsid[691]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Sep 6 00:33:40.308000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.337000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.224162 ignition[609]: Stage: fetch-offline Sep 6 00:33:40.243197 systemd[1]: Starting dracut-initqueue.service... Sep 6 00:33:40.224286 ignition[609]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:40.274914 systemd[1]: Finished ignition-fetch-offline.service. Sep 6 00:33:40.224353 ignition[609]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:40.309954 systemd[1]: Finished dracut-initqueue.service. Sep 6 00:33:40.244138 ignition[609]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:40.338793 systemd[1]: Reached target remote-fs-pre.target. Sep 6 00:33:40.479000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.244646 ignition[609]: parsed url from cmdline: "" Sep 6 00:33:40.355583 systemd[1]: Reached target remote-cryptsetup.target. Sep 6 00:33:40.495000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.244655 ignition[609]: no config URL provided Sep 6 00:33:40.355723 systemd[1]: Reached target remote-fs.target. Sep 6 00:33:40.244668 ignition[609]: reading system config file "/usr/lib/ignition/user.ign" Sep 6 00:33:40.382178 systemd[1]: Starting dracut-pre-mount.service... Sep 6 00:33:40.244767 ignition[609]: no config at "/usr/lib/ignition/user.ign" Sep 6 00:33:40.548000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.414472 systemd[1]: Starting ignition-fetch.service... Sep 6 00:33:40.244785 ignition[609]: failed to fetch config: resource requires networking Sep 6 00:33:40.454535 unknown[706]: fetched base config from "system" Sep 6 00:33:40.590000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.245185 ignition[609]: Ignition finished successfully Sep 6 00:33:40.454545 unknown[706]: fetched base config from "system" Sep 6 00:33:40.427431 ignition[706]: Ignition 2.14.0 Sep 6 00:33:40.454552 unknown[706]: fetched user config from "gcp" Sep 6 00:33:40.427445 ignition[706]: Stage: fetch Sep 6 00:33:40.463165 systemd[1]: Finished dracut-pre-mount.service. Sep 6 00:33:40.427603 ignition[706]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:40.480887 systemd[1]: Finished ignition-fetch.service. Sep 6 00:33:40.427639 ignition[706]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:40.497948 systemd[1]: Starting ignition-kargs.service... Sep 6 00:33:40.437063 ignition[706]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:40.528096 systemd[1]: Finished ignition-kargs.service. Sep 6 00:33:40.437272 ignition[706]: parsed url from cmdline: "" Sep 6 00:33:40.550979 systemd[1]: Starting ignition-disks.service... Sep 6 00:33:40.437278 ignition[706]: no config URL provided Sep 6 00:33:40.571388 systemd[1]: Finished ignition-disks.service. Sep 6 00:33:40.437285 ignition[706]: reading system config file "/usr/lib/ignition/user.ign" Sep 6 00:33:40.591794 systemd[1]: Reached target initrd-root-device.target. Sep 6 00:33:40.437297 ignition[706]: no config at "/usr/lib/ignition/user.ign" Sep 6 00:33:40.607576 systemd[1]: Reached target local-fs-pre.target. Sep 6 00:33:40.437336 ignition[706]: GET http://169.254.169.254/computeMetadata/v1/instance/attributes/user-data: attempt #1 Sep 6 00:33:40.621565 systemd[1]: Reached target local-fs.target. Sep 6 00:33:40.442476 ignition[706]: GET result: OK Sep 6 00:33:40.634572 systemd[1]: Reached target sysinit.target. Sep 6 00:33:40.442604 ignition[706]: parsing config with SHA512: e6628eafd5958a3aeca54ac9471dc15e0bdfc0845461d76a98f745c34affb02134925dd0ef57eba43136f7d4c51dfa2478798b767bd554c40f90c33526caa0a6 Sep 6 00:33:40.645556 systemd[1]: Reached target basic.target. Sep 6 00:33:40.455230 ignition[706]: fetch: fetch complete Sep 6 00:33:40.646915 systemd[1]: Starting systemd-fsck-root.service... Sep 6 00:33:40.455237 ignition[706]: fetch: fetch passed Sep 6 00:33:40.455284 ignition[706]: Ignition finished successfully Sep 6 00:33:40.512889 ignition[712]: Ignition 2.14.0 Sep 6 00:33:40.512898 ignition[712]: Stage: kargs Sep 6 00:33:40.513037 ignition[712]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:40.513178 ignition[712]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:40.521111 ignition[712]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:40.522513 ignition[712]: kargs: kargs passed Sep 6 00:33:40.522562 ignition[712]: Ignition finished successfully Sep 6 00:33:40.562647 ignition[718]: Ignition 2.14.0 Sep 6 00:33:40.562658 ignition[718]: Stage: disks Sep 6 00:33:40.562780 ignition[718]: reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:40.562805 ignition[718]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:40.569173 ignition[718]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:40.570478 ignition[718]: disks: disks passed Sep 6 00:33:40.570530 ignition[718]: Ignition finished successfully Sep 6 00:33:40.689798 systemd-fsck[726]: ROOT: clean, 629/1628000 files, 124065/1617920 blocks Sep 6 00:33:40.876546 systemd[1]: Finished systemd-fsck-root.service. Sep 6 00:33:40.875000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:40.877816 systemd[1]: Mounting sysroot.mount... Sep 6 00:33:40.912557 kernel: EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Sep 6 00:33:40.905440 systemd[1]: Mounted sysroot.mount. Sep 6 00:33:40.919838 systemd[1]: Reached target initrd-root-fs.target. Sep 6 00:33:40.936702 systemd[1]: Mounting sysroot-usr.mount... Sep 6 00:33:40.950116 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Sep 6 00:33:40.950165 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Sep 6 00:33:40.950196 systemd[1]: Reached target ignition-diskful.target. Sep 6 00:33:41.032946 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by mount (732) Sep 6 00:33:41.032977 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 6 00:33:41.032993 kernel: BTRFS info (device sda6): using free space tree Sep 6 00:33:41.033007 kernel: BTRFS info (device sda6): has skinny extents Sep 6 00:33:40.965818 systemd[1]: Mounted sysroot-usr.mount. Sep 6 00:33:40.989644 systemd[1]: Mounting sysroot-usr-share-oem.mount... Sep 6 00:33:41.054437 kernel: BTRFS info (device sda6): enabling ssd optimizations Sep 6 00:33:41.061832 systemd[1]: Starting initrd-setup-root.service... Sep 6 00:33:41.074262 systemd[1]: Mounted sysroot-usr-share-oem.mount. Sep 6 00:33:41.087896 initrd-setup-root[755]: cut: /sysroot/etc/passwd: No such file or directory Sep 6 00:33:41.097524 initrd-setup-root[763]: cut: /sysroot/etc/group: No such file or directory Sep 6 00:33:41.107539 initrd-setup-root[771]: cut: /sysroot/etc/shadow: No such file or directory Sep 6 00:33:41.118538 initrd-setup-root[779]: cut: /sysroot/etc/gshadow: No such file or directory Sep 6 00:33:41.172591 systemd[1]: Finished initrd-setup-root.service. Sep 6 00:33:41.212602 kernel: kauditd_printk_skb: 23 callbacks suppressed Sep 6 00:33:41.212667 kernel: audit: type=1130 audit(1757118821.171:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:41.171000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:41.174223 systemd[1]: Starting ignition-mount.service... Sep 6 00:33:41.220588 systemd[1]: Starting sysroot-boot.service... Sep 6 00:33:41.235376 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. Sep 6 00:33:41.235704 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. Sep 6 00:33:41.263190 ignition[798]: INFO : Ignition 2.14.0 Sep 6 00:33:41.263190 ignition[798]: INFO : Stage: mount Sep 6 00:33:41.263190 ignition[798]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:41.263190 ignition[798]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:41.395553 kernel: audit: type=1130 audit(1757118821.269:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:41.395600 kernel: audit: type=1130 audit(1757118821.302:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:41.395626 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (807) Sep 6 00:33:41.395648 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Sep 6 00:33:41.395671 kernel: BTRFS info (device sda6): using free space tree Sep 6 00:33:41.395692 kernel: BTRFS info (device sda6): has skinny extents Sep 6 00:33:41.269000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:41.302000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:41.263768 systemd[1]: Finished sysroot-boot.service. Sep 6 00:33:41.415631 kernel: BTRFS info (device sda6): enabling ssd optimizations Sep 6 00:33:41.415740 ignition[798]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:41.415740 ignition[798]: INFO : mount: mount passed Sep 6 00:33:41.415740 ignition[798]: INFO : Ignition finished successfully Sep 6 00:33:41.271060 systemd[1]: Finished ignition-mount.service. Sep 6 00:33:41.305448 systemd[1]: Starting ignition-files.service... Sep 6 00:33:41.344683 systemd[1]: Mounting sysroot-usr-share-oem.mount... Sep 6 00:33:41.475519 ignition[826]: INFO : Ignition 2.14.0 Sep 6 00:33:41.475519 ignition[826]: INFO : Stage: files Sep 6 00:33:41.475519 ignition[826]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:41.475519 ignition[826]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:41.475519 ignition[826]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:41.475519 ignition[826]: DEBUG : files: compiled without relabeling support, skipping Sep 6 00:33:41.475519 ignition[826]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Sep 6 00:33:41.475519 ignition[826]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Sep 6 00:33:41.475519 ignition[826]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Sep 6 00:33:41.475519 ignition[826]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Sep 6 00:33:41.475519 ignition[826]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Sep 6 00:33:41.475519 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/hosts" Sep 6 00:33:41.475519 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): oem config not found in "/usr/share/oem", looking on oem partition Sep 6 00:33:41.475519 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(4): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4083573330" Sep 6 00:33:41.475519 ignition[826]: CRITICAL : files: createFilesystemsFiles: createFiles: op(3): op(4): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4083573330": device or resource busy Sep 6 00:33:41.475519 ignition[826]: ERROR : files: createFilesystemsFiles: createFiles: op(3): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem4083573330", trying btrfs: device or resource busy Sep 6 00:33:41.475519 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(5): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4083573330" Sep 6 00:33:41.475519 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(5): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem4083573330" Sep 6 00:33:41.475519 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(6): [started] unmounting "/mnt/oem4083573330" Sep 6 00:33:41.412190 systemd[1]: Mounted sysroot-usr-share-oem.mount. Sep 6 00:33:41.748566 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(6): [finished] unmounting "/mnt/oem4083573330" Sep 6 00:33:41.748566 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/hosts" Sep 6 00:33:41.748566 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 6 00:33:41.748566 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Sep 6 00:33:41.748566 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Sep 6 00:33:41.463296 unknown[826]: wrote ssh authorized keys file for user: core Sep 6 00:33:41.533607 systemd-networkd[681]: eth0: Gained IPv6LL Sep 6 00:33:42.064041 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Sep 6 00:33:42.080556 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 6 00:33:42.080556 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Sep 6 00:33:42.283504 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET result: OK Sep 6 00:33:42.474793 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Sep 6 00:33:42.474793 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/profile.d/google-cloud-sdk.sh" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): oem config not found in "/usr/share/oem", looking on oem partition Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2424899731" Sep 6 00:33:42.507553 ignition[826]: CRITICAL : files: createFilesystemsFiles: createFiles: op(9): op(a): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2424899731": device or resource busy Sep 6 00:33:42.507553 ignition[826]: ERROR : files: createFilesystemsFiles: createFiles: op(9): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem2424899731", trying btrfs: device or resource busy Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2424899731" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(b): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2424899731" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(c): [started] unmounting "/mnt/oem2424899731" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(c): [finished] unmounting "/mnt/oem2424899731" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/profile.d/google-cloud-sdk.sh" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/home/core/install.sh" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/home/core/install.sh" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/home/core/nginx.yaml" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/home/core/nginx.yaml" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 6 00:33:42.507553 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(10): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(10): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(11): [started] writing file "/sysroot/etc/flatcar/update.conf" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(11): [finished] writing file "/sysroot/etc/flatcar/update.conf" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(12): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(12): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): [started] writing file "/sysroot/etc/systemd/system/oem-gce-enable-oslogin.service" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): oem config not found in "/usr/share/oem", looking on oem partition Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(14): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3727504367" Sep 6 00:33:42.752773 ignition[826]: CRITICAL : files: createFilesystemsFiles: createFiles: op(13): op(14): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3727504367": device or resource busy Sep 6 00:33:42.752773 ignition[826]: ERROR : files: createFilesystemsFiles: createFiles: op(13): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem3727504367", trying btrfs: device or resource busy Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(15): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3727504367" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(15): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3727504367" Sep 6 00:33:42.752773 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(16): [started] unmounting "/mnt/oem3727504367" Sep 6 00:33:42.508195 systemd[1]: mnt-oem3727504367.mount: Deactivated successfully. Sep 6 00:33:43.010660 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(16): [finished] unmounting "/mnt/oem3727504367" Sep 6 00:33:43.010660 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(13): [finished] writing file "/sysroot/etc/systemd/system/oem-gce-enable-oslogin.service" Sep 6 00:33:43.010660 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(17): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 6 00:33:43.010660 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(17): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 Sep 6 00:33:43.010660 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(17): GET result: OK Sep 6 00:33:43.335516 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(17): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" Sep 6 00:33:43.335516 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): [started] writing file "/sysroot/etc/systemd/system/oem-gce.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): oem config not found in "/usr/share/oem", looking on oem partition Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(19): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3304790790" Sep 6 00:33:43.370592 ignition[826]: CRITICAL : files: createFilesystemsFiles: createFiles: op(18): op(19): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3304790790": device or resource busy Sep 6 00:33:43.370592 ignition[826]: ERROR : files: createFilesystemsFiles: createFiles: op(18): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem3304790790", trying btrfs: device or resource busy Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3304790790" Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1a): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3304790790" Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1b): [started] unmounting "/mnt/oem3304790790" Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1b): [finished] unmounting "/mnt/oem3304790790" Sep 6 00:33:43.370592 ignition[826]: INFO : files: createFilesystemsFiles: createFiles: op(18): [finished] writing file "/sysroot/etc/systemd/system/oem-gce.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1c): [started] processing unit "coreos-metadata-sshkeys@.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1c): [finished] processing unit "coreos-metadata-sshkeys@.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1d): [started] processing unit "oem-gce.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1d): [finished] processing unit "oem-gce.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1e): [started] processing unit "oem-gce-enable-oslogin.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1e): [finished] processing unit "oem-gce-enable-oslogin.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1f): [started] processing unit "prepare-helm.service" Sep 6 00:33:43.370592 ignition[826]: INFO : files: op(1f): op(20): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 6 00:33:43.862728 kernel: audit: type=1130 audit(1757118823.369:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.862772 kernel: audit: type=1130 audit(1757118823.472:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.862800 kernel: audit: type=1130 audit(1757118823.514:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.862817 kernel: audit: type=1131 audit(1757118823.514:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.862831 kernel: audit: type=1130 audit(1757118823.623:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.862846 kernel: audit: type=1131 audit(1757118823.623:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.862860 kernel: audit: type=1130 audit(1757118823.754:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.369000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.472000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.514000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.514000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.623000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.623000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.754000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.354461 systemd[1]: Finished ignition-files.service. Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(1f): op(20): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(1f): [finished] processing unit "prepare-helm.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(21): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(21): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(22): [started] setting preset to enabled for "oem-gce.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(22): [finished] setting preset to enabled for "oem-gce.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(23): [started] setting preset to enabled for "oem-gce-enable-oslogin.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(23): [finished] setting preset to enabled for "oem-gce-enable-oslogin.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(24): [started] setting preset to enabled for "prepare-helm.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: op(24): [finished] setting preset to enabled for "prepare-helm.service" Sep 6 00:33:43.876691 ignition[826]: INFO : files: createResultFile: createFiles: op(25): [started] writing file "/sysroot/etc/.ignition-result.json" Sep 6 00:33:43.876691 ignition[826]: INFO : files: createResultFile: createFiles: op(25): [finished] writing file "/sysroot/etc/.ignition-result.json" Sep 6 00:33:43.876691 ignition[826]: INFO : files: files passed Sep 6 00:33:43.876691 ignition[826]: INFO : Ignition finished successfully Sep 6 00:33:43.883000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.120000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.381358 systemd[1]: Starting initrd-setup-root-after-ignition.service... Sep 6 00:33:43.412770 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Sep 6 00:33:44.166000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.184677 initrd-setup-root-after-ignition[849]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Sep 6 00:33:44.193000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.413843 systemd[1]: Starting ignition-quench.service... Sep 6 00:33:44.215000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.443947 systemd[1]: Finished initrd-setup-root-after-ignition.service. Sep 6 00:33:43.473999 systemd[1]: ignition-quench.service: Deactivated successfully. Sep 6 00:33:44.256555 iscsid[691]: iscsid shutting down. Sep 6 00:33:43.474145 systemd[1]: Finished ignition-quench.service. Sep 6 00:33:44.278000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.516047 systemd[1]: Reached target ignition-complete.target. Sep 6 00:33:44.294000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.579892 systemd[1]: Starting initrd-parse-etc.service... Sep 6 00:33:44.310000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.623425 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Sep 6 00:33:44.327000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.336662 ignition[864]: INFO : Ignition 2.14.0 Sep 6 00:33:44.336662 ignition[864]: INFO : Stage: umount Sep 6 00:33:44.336662 ignition[864]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Sep 6 00:33:44.336662 ignition[864]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Sep 6 00:33:44.336662 ignition[864]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" Sep 6 00:33:44.336662 ignition[864]: INFO : umount: umount passed Sep 6 00:33:44.336662 ignition[864]: INFO : Ignition finished successfully Sep 6 00:33:44.342000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.358000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.365000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.407000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.418000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.623549 systemd[1]: Finished initrd-parse-etc.service. Sep 6 00:33:43.625143 systemd[1]: Reached target initrd-fs.target. Sep 6 00:33:43.707763 systemd[1]: Reached target initrd.target. Sep 6 00:33:43.720852 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Sep 6 00:33:44.504000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.722154 systemd[1]: Starting dracut-pre-pivot.service... Sep 6 00:33:44.523000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.739819 systemd[1]: Finished dracut-pre-pivot.service. Sep 6 00:33:43.757632 systemd[1]: Starting initrd-cleanup.service... Sep 6 00:33:44.545000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.802896 systemd[1]: Stopped target nss-lookup.target. Sep 6 00:33:44.566000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.566000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.810083 systemd[1]: Stopped target remote-cryptsetup.target. Sep 6 00:33:43.845912 systemd[1]: Stopped target timers.target. Sep 6 00:33:43.869746 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Sep 6 00:33:43.869940 systemd[1]: Stopped dracut-pre-pivot.service. Sep 6 00:33:43.884919 systemd[1]: Stopped target initrd.target. Sep 6 00:33:44.633000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.908759 systemd[1]: Stopped target basic.target. Sep 6 00:33:44.654000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.654000 audit: BPF prog-id=6 op=UNLOAD Sep 6 00:33:43.926752 systemd[1]: Stopped target ignition-complete.target. Sep 6 00:33:43.948751 systemd[1]: Stopped target ignition-diskful.target. Sep 6 00:33:43.970734 systemd[1]: Stopped target initrd-root-device.target. Sep 6 00:33:44.700000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:43.991866 systemd[1]: Stopped target remote-fs.target. Sep 6 00:33:44.717000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.004932 systemd[1]: Stopped target remote-fs-pre.target. Sep 6 00:33:44.732000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.023910 systemd[1]: Stopped target sysinit.target. Sep 6 00:33:44.043891 systemd[1]: Stopped target local-fs.target. Sep 6 00:33:44.761000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.061906 systemd[1]: Stopped target local-fs-pre.target. Sep 6 00:33:44.080053 systemd[1]: Stopped target swap.target. Sep 6 00:33:44.100900 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Sep 6 00:33:44.807000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.101093 systemd[1]: Stopped dracut-pre-mount.service. Sep 6 00:33:44.825000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.122048 systemd[1]: Stopped target cryptsetup.target. Sep 6 00:33:44.840000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.149798 systemd[1]: dracut-initqueue.service: Deactivated successfully. Sep 6 00:33:44.149998 systemd[1]: Stopped dracut-initqueue.service. Sep 6 00:33:44.866000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.167944 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Sep 6 00:33:44.889000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.168136 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Sep 6 00:33:44.907000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.194914 systemd[1]: ignition-files.service: Deactivated successfully. Sep 6 00:33:44.924000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.195085 systemd[1]: Stopped ignition-files.service. Sep 6 00:33:44.940000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.940000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:44.218171 systemd[1]: Stopping ignition-mount.service... Sep 6 00:33:44.230904 systemd[1]: Stopping iscsid.service... Sep 6 00:33:44.239112 systemd[1]: Stopping sysroot-boot.service... Sep 6 00:33:44.264571 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Sep 6 00:33:44.264841 systemd[1]: Stopped systemd-udev-trigger.service. Sep 6 00:33:44.279803 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Sep 6 00:33:45.018538 systemd-journald[190]: Received SIGTERM from PID 1 (systemd). Sep 6 00:33:44.279972 systemd[1]: Stopped dracut-pre-trigger.service. Sep 6 00:33:44.299791 systemd[1]: sysroot-boot.mount: Deactivated successfully. Sep 6 00:33:44.300891 systemd[1]: iscsid.service: Deactivated successfully. Sep 6 00:33:44.301004 systemd[1]: Stopped iscsid.service. Sep 6 00:33:44.312354 systemd[1]: ignition-mount.service: Deactivated successfully. Sep 6 00:33:44.312490 systemd[1]: Stopped ignition-mount.service. Sep 6 00:33:44.329238 systemd[1]: sysroot-boot.service: Deactivated successfully. Sep 6 00:33:44.329342 systemd[1]: Stopped sysroot-boot.service. Sep 6 00:33:44.344323 systemd[1]: ignition-disks.service: Deactivated successfully. Sep 6 00:33:44.344477 systemd[1]: Stopped ignition-disks.service. Sep 6 00:33:44.359648 systemd[1]: ignition-kargs.service: Deactivated successfully. Sep 6 00:33:44.359731 systemd[1]: Stopped ignition-kargs.service. Sep 6 00:33:44.366774 systemd[1]: ignition-fetch.service: Deactivated successfully. Sep 6 00:33:44.366823 systemd[1]: Stopped ignition-fetch.service. Sep 6 00:33:44.408703 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Sep 6 00:33:44.408785 systemd[1]: Stopped ignition-fetch-offline.service. Sep 6 00:33:44.419930 systemd[1]: Stopped target paths.target. Sep 6 00:33:44.440558 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Sep 6 00:33:44.444547 systemd[1]: Stopped systemd-ask-password-console.path. Sep 6 00:33:44.456553 systemd[1]: Stopped target slices.target. Sep 6 00:33:44.456737 systemd[1]: Stopped target sockets.target. Sep 6 00:33:44.477591 systemd[1]: iscsid.socket: Deactivated successfully. Sep 6 00:33:44.477664 systemd[1]: Closed iscsid.socket. Sep 6 00:33:44.490588 systemd[1]: ignition-setup.service: Deactivated successfully. Sep 6 00:33:44.490689 systemd[1]: Stopped ignition-setup.service. Sep 6 00:33:44.505746 systemd[1]: initrd-setup-root.service: Deactivated successfully. Sep 6 00:33:44.505846 systemd[1]: Stopped initrd-setup-root.service. Sep 6 00:33:44.524874 systemd[1]: Stopping iscsiuio.service... Sep 6 00:33:44.540045 systemd[1]: iscsiuio.service: Deactivated successfully. Sep 6 00:33:44.540166 systemd[1]: Stopped iscsiuio.service. Sep 6 00:33:44.547305 systemd[1]: initrd-cleanup.service: Deactivated successfully. Sep 6 00:33:44.547466 systemd[1]: Finished initrd-cleanup.service. Sep 6 00:33:44.569035 systemd[1]: Stopped target network.target. Sep 6 00:33:44.583678 systemd[1]: iscsiuio.socket: Deactivated successfully. Sep 6 00:33:44.583777 systemd[1]: Closed iscsiuio.socket. Sep 6 00:33:44.600002 systemd[1]: Stopping systemd-networkd.service... Sep 6 00:33:44.603544 systemd-networkd[681]: eth0: DHCPv6 lease lost Sep 6 00:33:44.608002 systemd[1]: Stopping systemd-resolved.service... Sep 6 00:33:44.622077 systemd[1]: systemd-resolved.service: Deactivated successfully. Sep 6 00:33:44.622201 systemd[1]: Stopped systemd-resolved.service. Sep 6 00:33:44.635465 systemd[1]: systemd-networkd.service: Deactivated successfully. Sep 6 00:33:44.635663 systemd[1]: Stopped systemd-networkd.service. Sep 6 00:33:44.656260 systemd[1]: systemd-networkd.socket: Deactivated successfully. Sep 6 00:33:44.656302 systemd[1]: Closed systemd-networkd.socket. Sep 6 00:33:44.671626 systemd[1]: Stopping network-cleanup.service... Sep 6 00:33:44.685623 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Sep 6 00:33:44.685774 systemd[1]: Stopped parse-ip-for-networkd.service. Sep 6 00:33:44.701709 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 6 00:33:44.701798 systemd[1]: Stopped systemd-sysctl.service. Sep 6 00:33:44.718825 systemd[1]: systemd-modules-load.service: Deactivated successfully. Sep 6 00:33:44.718894 systemd[1]: Stopped systemd-modules-load.service. Sep 6 00:33:44.733804 systemd[1]: Stopping systemd-udevd.service... Sep 6 00:33:44.749594 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Sep 6 00:33:44.750250 systemd[1]: systemd-udevd.service: Deactivated successfully. Sep 6 00:33:44.750428 systemd[1]: Stopped systemd-udevd.service. Sep 6 00:33:44.764390 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Sep 6 00:33:44.764668 systemd[1]: Closed systemd-udevd-control.socket. Sep 6 00:33:44.777738 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Sep 6 00:33:44.777789 systemd[1]: Closed systemd-udevd-kernel.socket. Sep 6 00:33:44.793685 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Sep 6 00:33:44.793768 systemd[1]: Stopped dracut-pre-udev.service. Sep 6 00:33:44.808750 systemd[1]: dracut-cmdline.service: Deactivated successfully. Sep 6 00:33:44.808810 systemd[1]: Stopped dracut-cmdline.service. Sep 6 00:33:44.826776 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Sep 6 00:33:44.826847 systemd[1]: Stopped dracut-cmdline-ask.service. Sep 6 00:33:44.843565 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Sep 6 00:33:44.867544 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Sep 6 00:33:44.867654 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Sep 6 00:33:44.867893 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Sep 6 00:33:44.867940 systemd[1]: Stopped kmod-static-nodes.service. Sep 6 00:33:44.890738 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Sep 6 00:33:44.890803 systemd[1]: Stopped systemd-vconsole-setup.service. Sep 6 00:33:44.910179 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Sep 6 00:33:44.910864 systemd[1]: network-cleanup.service: Deactivated successfully. Sep 6 00:33:44.910978 systemd[1]: Stopped network-cleanup.service. Sep 6 00:33:44.925963 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Sep 6 00:33:44.926187 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Sep 6 00:33:44.941813 systemd[1]: Reached target initrd-switch-root.target. Sep 6 00:33:44.958702 systemd[1]: Starting initrd-switch-root.service... Sep 6 00:33:44.980856 systemd[1]: Switching root. Sep 6 00:33:45.029095 systemd-journald[190]: Journal stopped Sep 6 00:33:49.720640 kernel: SELinux: Class mctp_socket not defined in policy. Sep 6 00:33:49.722177 kernel: SELinux: Class anon_inode not defined in policy. Sep 6 00:33:49.722220 kernel: SELinux: the above unknown classes and permissions will be allowed Sep 6 00:33:49.722264 kernel: SELinux: policy capability network_peer_controls=1 Sep 6 00:33:49.722300 kernel: SELinux: policy capability open_perms=1 Sep 6 00:33:49.722325 kernel: SELinux: policy capability extended_socket_class=1 Sep 6 00:33:49.722351 kernel: SELinux: policy capability always_check_network=0 Sep 6 00:33:49.722374 kernel: SELinux: policy capability cgroup_seclabel=1 Sep 6 00:33:49.722416 kernel: SELinux: policy capability nnp_nosuid_transition=1 Sep 6 00:33:49.722443 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Sep 6 00:33:49.722467 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Sep 6 00:33:49.722493 systemd[1]: Successfully loaded SELinux policy in 110.975ms. Sep 6 00:33:49.722537 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.392ms. Sep 6 00:33:49.722563 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Sep 6 00:33:49.722588 systemd[1]: Detected virtualization kvm. Sep 6 00:33:49.722619 systemd[1]: Detected architecture x86-64. Sep 6 00:33:49.722643 systemd[1]: Detected first boot. Sep 6 00:33:49.722667 systemd[1]: Initializing machine ID from VM UUID. Sep 6 00:33:49.722693 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Sep 6 00:33:49.722717 systemd[1]: Populated /etc with preset unit settings. Sep 6 00:33:49.722754 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 6 00:33:49.722795 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 6 00:33:49.722821 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 6 00:33:49.722857 kernel: kauditd_printk_skb: 52 callbacks suppressed Sep 6 00:33:49.722879 kernel: audit: type=1334 audit(1757118828.832:89): prog-id=12 op=LOAD Sep 6 00:33:49.724508 kernel: audit: type=1334 audit(1757118828.832:90): prog-id=3 op=UNLOAD Sep 6 00:33:49.724554 kernel: audit: type=1334 audit(1757118828.844:91): prog-id=13 op=LOAD Sep 6 00:33:49.724580 kernel: audit: type=1334 audit(1757118828.851:92): prog-id=14 op=LOAD Sep 6 00:33:49.724602 kernel: audit: type=1334 audit(1757118828.851:93): prog-id=4 op=UNLOAD Sep 6 00:33:49.724626 kernel: audit: type=1334 audit(1757118828.851:94): prog-id=5 op=UNLOAD Sep 6 00:33:49.724648 kernel: audit: type=1334 audit(1757118828.858:95): prog-id=15 op=LOAD Sep 6 00:33:49.724670 kernel: audit: type=1334 audit(1757118828.858:96): prog-id=12 op=UNLOAD Sep 6 00:33:49.724706 kernel: audit: type=1334 audit(1757118828.886:97): prog-id=16 op=LOAD Sep 6 00:33:49.724729 kernel: audit: type=1334 audit(1757118828.900:98): prog-id=17 op=LOAD Sep 6 00:33:49.724754 systemd[1]: initrd-switch-root.service: Deactivated successfully. Sep 6 00:33:49.724779 systemd[1]: Stopped initrd-switch-root.service. Sep 6 00:33:49.724805 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Sep 6 00:33:49.724830 systemd[1]: Created slice system-addon\x2dconfig.slice. Sep 6 00:33:49.724853 systemd[1]: Created slice system-addon\x2drun.slice. Sep 6 00:33:49.724879 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Sep 6 00:33:49.730444 systemd[1]: Created slice system-getty.slice. Sep 6 00:33:49.730485 systemd[1]: Created slice system-modprobe.slice. Sep 6 00:33:49.730512 systemd[1]: Created slice system-serial\x2dgetty.slice. Sep 6 00:33:49.730537 systemd[1]: Created slice system-system\x2dcloudinit.slice. Sep 6 00:33:49.730561 systemd[1]: Created slice system-systemd\x2dfsck.slice. Sep 6 00:33:49.730586 systemd[1]: Created slice user.slice. Sep 6 00:33:49.730611 systemd[1]: Started systemd-ask-password-console.path. Sep 6 00:33:49.730635 systemd[1]: Started systemd-ask-password-wall.path. Sep 6 00:33:49.730666 systemd[1]: Set up automount boot.automount. Sep 6 00:33:49.730691 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Sep 6 00:33:49.730718 systemd[1]: Stopped target initrd-switch-root.target. Sep 6 00:33:49.730743 systemd[1]: Stopped target initrd-fs.target. Sep 6 00:33:49.730767 systemd[1]: Stopped target initrd-root-fs.target. Sep 6 00:33:49.730791 systemd[1]: Reached target integritysetup.target. Sep 6 00:33:49.730817 systemd[1]: Reached target remote-cryptsetup.target. Sep 6 00:33:49.730842 systemd[1]: Reached target remote-fs.target. Sep 6 00:33:49.730867 systemd[1]: Reached target slices.target. Sep 6 00:33:49.730895 systemd[1]: Reached target swap.target. Sep 6 00:33:49.730929 systemd[1]: Reached target torcx.target. Sep 6 00:33:49.730959 systemd[1]: Reached target veritysetup.target. Sep 6 00:33:49.730981 systemd[1]: Listening on systemd-coredump.socket. Sep 6 00:33:49.731004 systemd[1]: Listening on systemd-initctl.socket. Sep 6 00:33:49.731030 systemd[1]: Listening on systemd-networkd.socket. Sep 6 00:33:49.731055 systemd[1]: Listening on systemd-udevd-control.socket. Sep 6 00:33:49.731078 systemd[1]: Listening on systemd-udevd-kernel.socket. Sep 6 00:33:49.731103 systemd[1]: Listening on systemd-userdbd.socket. Sep 6 00:33:49.731130 systemd[1]: Mounting dev-hugepages.mount... Sep 6 00:33:49.731159 systemd[1]: Mounting dev-mqueue.mount... Sep 6 00:33:49.731184 systemd[1]: Mounting media.mount... Sep 6 00:33:49.731209 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:49.731233 systemd[1]: Mounting sys-kernel-debug.mount... Sep 6 00:33:49.731259 systemd[1]: Mounting sys-kernel-tracing.mount... Sep 6 00:33:49.731283 systemd[1]: Mounting tmp.mount... Sep 6 00:33:49.731307 systemd[1]: Starting flatcar-tmpfiles.service... Sep 6 00:33:49.731331 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 6 00:33:49.731354 systemd[1]: Starting kmod-static-nodes.service... Sep 6 00:33:49.731385 systemd[1]: Starting modprobe@configfs.service... Sep 6 00:33:49.731448 systemd[1]: Starting modprobe@dm_mod.service... Sep 6 00:33:49.731477 systemd[1]: Starting modprobe@drm.service... Sep 6 00:33:49.731502 systemd[1]: Starting modprobe@efi_pstore.service... Sep 6 00:33:49.731527 systemd[1]: Starting modprobe@fuse.service... Sep 6 00:33:49.731567 systemd[1]: Starting modprobe@loop.service... Sep 6 00:33:49.731592 kernel: fuse: init (API version 7.34) Sep 6 00:33:49.731617 kernel: loop: module loaded Sep 6 00:33:49.731642 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Sep 6 00:33:49.731673 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Sep 6 00:33:49.731696 systemd[1]: Stopped systemd-fsck-root.service. Sep 6 00:33:49.731721 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Sep 6 00:33:49.731745 systemd[1]: Stopped systemd-fsck-usr.service. Sep 6 00:33:49.731768 systemd[1]: Stopped systemd-journald.service. Sep 6 00:33:49.731792 systemd[1]: Starting systemd-journald.service... Sep 6 00:33:49.731816 systemd[1]: Starting systemd-modules-load.service... Sep 6 00:33:49.731841 systemd[1]: Starting systemd-network-generator.service... Sep 6 00:33:49.731865 systemd[1]: Starting systemd-remount-fs.service... Sep 6 00:33:49.731901 systemd-journald[987]: Journal started Sep 6 00:33:49.732031 systemd-journald[987]: Runtime Journal (/run/log/journal/54d118c5b4df6aa1e26d8995714142bc) is 8.0M, max 148.8M, 140.8M free. Sep 6 00:33:45.028000 audit: BPF prog-id=9 op=UNLOAD Sep 6 00:33:45.312000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Sep 6 00:33:45.456000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 6 00:33:45.456000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Sep 6 00:33:45.456000 audit: BPF prog-id=10 op=LOAD Sep 6 00:33:45.456000 audit: BPF prog-id=10 op=UNLOAD Sep 6 00:33:45.456000 audit: BPF prog-id=11 op=LOAD Sep 6 00:33:45.456000 audit: BPF prog-id=11 op=UNLOAD Sep 6 00:33:45.623000 audit[897]: AVC avc: denied { associate } for pid=897 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Sep 6 00:33:45.623000 audit[897]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c0001078cc a1=c00002ae40 a2=c000029100 a3=32 items=0 ppid=880 pid=897 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 6 00:33:45.623000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Sep 6 00:33:45.633000 audit[897]: AVC avc: denied { associate } for pid=897 comm="torcx-generator" name="lib" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Sep 6 00:33:45.633000 audit[897]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c0001079a5 a2=1ed a3=0 items=2 ppid=880 pid=897 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 6 00:33:45.633000 audit: CWD cwd="/" Sep 6 00:33:45.633000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:45.633000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:45.633000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Sep 6 00:33:48.832000 audit: BPF prog-id=12 op=LOAD Sep 6 00:33:48.832000 audit: BPF prog-id=3 op=UNLOAD Sep 6 00:33:48.844000 audit: BPF prog-id=13 op=LOAD Sep 6 00:33:48.851000 audit: BPF prog-id=14 op=LOAD Sep 6 00:33:48.851000 audit: BPF prog-id=4 op=UNLOAD Sep 6 00:33:48.851000 audit: BPF prog-id=5 op=UNLOAD Sep 6 00:33:48.858000 audit: BPF prog-id=15 op=LOAD Sep 6 00:33:48.858000 audit: BPF prog-id=12 op=UNLOAD Sep 6 00:33:48.886000 audit: BPF prog-id=16 op=LOAD Sep 6 00:33:48.900000 audit: BPF prog-id=17 op=LOAD Sep 6 00:33:48.900000 audit: BPF prog-id=13 op=UNLOAD Sep 6 00:33:48.900000 audit: BPF prog-id=14 op=UNLOAD Sep 6 00:33:48.902000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:48.922000 audit: BPF prog-id=15 op=UNLOAD Sep 6 00:33:48.931000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:48.931000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.635000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.657000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.671000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.671000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.672000 audit: BPF prog-id=18 op=LOAD Sep 6 00:33:49.672000 audit: BPF prog-id=19 op=LOAD Sep 6 00:33:49.672000 audit: BPF prog-id=20 op=LOAD Sep 6 00:33:49.672000 audit: BPF prog-id=16 op=UNLOAD Sep 6 00:33:49.672000 audit: BPF prog-id=17 op=UNLOAD Sep 6 00:33:49.715000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Sep 6 00:33:49.715000 audit[987]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=5 a1=7ffcb53aeee0 a2=4000 a3=7ffcb53aef7c items=0 ppid=1 pid=987 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 6 00:33:49.715000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Sep 6 00:33:48.831254 systemd[1]: Queued start job for default target multi-user.target. Sep 6 00:33:45.619123 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 6 00:33:48.831275 systemd[1]: Unnecessary job was removed for dev-sda6.device. Sep 6 00:33:45.620213 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Sep 6 00:33:48.903869 systemd[1]: systemd-journald.service: Deactivated successfully. Sep 6 00:33:45.620254 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Sep 6 00:33:45.620320 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Sep 6 00:33:45.620341 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="skipped missing lower profile" missing profile=oem Sep 6 00:33:45.620421 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Sep 6 00:33:45.620447 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Sep 6 00:33:45.620778 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Sep 6 00:33:45.620847 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Sep 6 00:33:45.620872 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Sep 6 00:33:45.623215 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Sep 6 00:33:45.623284 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Sep 6 00:33:45.623331 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.8: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.8 Sep 6 00:33:45.623361 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Sep 6 00:33:45.623407 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.8: no such file or directory" path=/var/lib/torcx/store/3510.3.8 Sep 6 00:33:45.623447 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:45Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Sep 6 00:33:48.183223 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:48Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 6 00:33:48.183558 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:48Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 6 00:33:48.183973 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:48Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 6 00:33:48.184513 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:48Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Sep 6 00:33:48.184604 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:48Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Sep 6 00:33:48.184685 /usr/lib/systemd/system-generators/torcx-generator[897]: time="2025-09-06T00:33:48Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Sep 6 00:33:49.745444 systemd[1]: Starting systemd-udev-trigger.service... Sep 6 00:33:49.763699 systemd[1]: verity-setup.service: Deactivated successfully. Sep 6 00:33:49.763779 systemd[1]: Stopped verity-setup.service. Sep 6 00:33:49.772000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.786420 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:49.796849 systemd[1]: Started systemd-journald.service. Sep 6 00:33:49.804000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.806820 systemd[1]: Mounted dev-hugepages.mount. Sep 6 00:33:49.813732 systemd[1]: Mounted dev-mqueue.mount. Sep 6 00:33:49.821686 systemd[1]: Mounted media.mount. Sep 6 00:33:49.828673 systemd[1]: Mounted sys-kernel-debug.mount. Sep 6 00:33:49.837668 systemd[1]: Mounted sys-kernel-tracing.mount. Sep 6 00:33:49.846682 systemd[1]: Mounted tmp.mount. Sep 6 00:33:49.854113 systemd[1]: Finished flatcar-tmpfiles.service. Sep 6 00:33:49.861000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.862966 systemd[1]: Finished kmod-static-nodes.service. Sep 6 00:33:49.870000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.872155 systemd[1]: modprobe@configfs.service: Deactivated successfully. Sep 6 00:33:49.872462 systemd[1]: Finished modprobe@configfs.service. Sep 6 00:33:49.880000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.880000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.881946 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 6 00:33:49.882173 systemd[1]: Finished modprobe@dm_mod.service. Sep 6 00:33:49.889000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.889000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.891220 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 6 00:33:49.891489 systemd[1]: Finished modprobe@drm.service. Sep 6 00:33:49.899000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.899000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.900973 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 6 00:33:49.901195 systemd[1]: Finished modprobe@efi_pstore.service. Sep 6 00:33:49.908000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.908000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.909911 systemd[1]: modprobe@fuse.service: Deactivated successfully. Sep 6 00:33:49.910133 systemd[1]: Finished modprobe@fuse.service. Sep 6 00:33:49.917000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.917000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.918883 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 6 00:33:49.919101 systemd[1]: Finished modprobe@loop.service. Sep 6 00:33:49.926000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.926000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.927942 systemd[1]: Finished systemd-modules-load.service. Sep 6 00:33:49.935000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.936979 systemd[1]: Finished systemd-network-generator.service. Sep 6 00:33:49.944000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.945882 systemd[1]: Finished systemd-remount-fs.service. Sep 6 00:33:49.953000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.954915 systemd[1]: Finished systemd-udev-trigger.service. Sep 6 00:33:49.962000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:49.964631 systemd[1]: Reached target network-pre.target. Sep 6 00:33:49.975069 systemd[1]: Mounting sys-fs-fuse-connections.mount... Sep 6 00:33:49.985000 systemd[1]: Mounting sys-kernel-config.mount... Sep 6 00:33:49.992562 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Sep 6 00:33:49.995138 systemd[1]: Starting systemd-hwdb-update.service... Sep 6 00:33:50.004616 systemd[1]: Starting systemd-journal-flush.service... Sep 6 00:33:50.012651 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 6 00:33:50.014580 systemd[1]: Starting systemd-random-seed.service... Sep 6 00:33:50.021583 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 6 00:33:50.023726 systemd[1]: Starting systemd-sysctl.service... Sep 6 00:33:50.029617 systemd-journald[987]: Time spent on flushing to /var/log/journal/54d118c5b4df6aa1e26d8995714142bc is 53.279ms for 1167 entries. Sep 6 00:33:50.029617 systemd-journald[987]: System Journal (/var/log/journal/54d118c5b4df6aa1e26d8995714142bc) is 8.0M, max 584.8M, 576.8M free. Sep 6 00:33:50.124767 systemd-journald[987]: Received client request to flush runtime journal. Sep 6 00:33:50.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.093000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.117000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.039583 systemd[1]: Starting systemd-sysusers.service... Sep 6 00:33:50.048044 systemd[1]: Starting systemd-udev-settle.service... Sep 6 00:33:50.058862 systemd[1]: Mounted sys-fs-fuse-connections.mount. Sep 6 00:33:50.132134 udevadm[1002]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Sep 6 00:33:50.067715 systemd[1]: Mounted sys-kernel-config.mount. Sep 6 00:33:50.076840 systemd[1]: Finished systemd-random-seed.service. Sep 6 00:33:50.085902 systemd[1]: Finished systemd-sysctl.service. Sep 6 00:33:50.098073 systemd[1]: Reached target first-boot-complete.target. Sep 6 00:33:50.109827 systemd[1]: Finished systemd-sysusers.service. Sep 6 00:33:50.120419 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Sep 6 00:33:50.131058 systemd[1]: Finished systemd-journal-flush.service. Sep 6 00:33:50.138000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.176149 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Sep 6 00:33:50.183000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.702682 systemd[1]: Finished systemd-hwdb-update.service. Sep 6 00:33:50.710000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.710000 audit: BPF prog-id=21 op=LOAD Sep 6 00:33:50.711000 audit: BPF prog-id=22 op=LOAD Sep 6 00:33:50.711000 audit: BPF prog-id=7 op=UNLOAD Sep 6 00:33:50.711000 audit: BPF prog-id=8 op=UNLOAD Sep 6 00:33:50.713447 systemd[1]: Starting systemd-udevd.service... Sep 6 00:33:50.737141 systemd-udevd[1006]: Using default interface naming scheme 'v252'. Sep 6 00:33:50.788201 systemd[1]: Started systemd-udevd.service. Sep 6 00:33:50.795000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.797000 audit: BPF prog-id=23 op=LOAD Sep 6 00:33:50.800055 systemd[1]: Starting systemd-networkd.service... Sep 6 00:33:50.810000 audit: BPF prog-id=24 op=LOAD Sep 6 00:33:50.810000 audit: BPF prog-id=25 op=LOAD Sep 6 00:33:50.810000 audit: BPF prog-id=26 op=LOAD Sep 6 00:33:50.813214 systemd[1]: Starting systemd-userdbd.service... Sep 6 00:33:50.864028 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Sep 6 00:33:50.895283 systemd[1]: Started systemd-userdbd.service. Sep 6 00:33:50.902000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:50.980426 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Sep 6 00:33:51.016830 kernel: ACPI: button: Power Button [PWRF] Sep 6 00:33:51.016949 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input3 Sep 6 00:33:51.021857 kernel: ACPI: button: Sleep Button [SLPF] Sep 6 00:33:51.031000 audit[1018]: AVC avc: denied { confidentiality } for pid=1018 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Sep 6 00:33:51.067558 systemd-networkd[1017]: lo: Link UP Sep 6 00:33:51.067959 systemd-networkd[1017]: lo: Gained carrier Sep 6 00:33:51.068975 systemd-networkd[1017]: Enumeration completed Sep 6 00:33:51.069247 systemd[1]: Started systemd-networkd.service. Sep 6 00:33:51.069670 systemd-networkd[1017]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Sep 6 00:33:51.071913 systemd-networkd[1017]: eth0: Link UP Sep 6 00:33:51.072037 systemd-networkd[1017]: eth0: Gained carrier Sep 6 00:33:51.031000 audit[1018]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=564c86aab340 a1=338ec a2=7f55e7222bc5 a3=5 items=110 ppid=1006 pid=1018 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 6 00:33:51.031000 audit: CWD cwd="/" Sep 6 00:33:51.031000 audit: PATH item=0 name=(null) inode=40 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=1 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=2 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=3 name=(null) inode=13714 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=4 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=5 name=(null) inode=13715 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=6 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=7 name=(null) inode=13716 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=8 name=(null) inode=13716 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=9 name=(null) inode=13717 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=10 name=(null) inode=13716 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=11 name=(null) inode=13718 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=12 name=(null) inode=13716 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=13 name=(null) inode=13719 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=14 name=(null) inode=13716 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=15 name=(null) inode=13720 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=16 name=(null) inode=13716 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=17 name=(null) inode=13721 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=18 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=19 name=(null) inode=13722 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=20 name=(null) inode=13722 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=21 name=(null) inode=13723 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=22 name=(null) inode=13722 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=23 name=(null) inode=13724 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=24 name=(null) inode=13722 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=25 name=(null) inode=13725 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.076000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.031000 audit: PATH item=26 name=(null) inode=13722 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=27 name=(null) inode=13726 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=28 name=(null) inode=13722 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=29 name=(null) inode=13727 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=30 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=31 name=(null) inode=13728 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=32 name=(null) inode=13728 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=33 name=(null) inode=13729 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=34 name=(null) inode=13728 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=35 name=(null) inode=13730 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=36 name=(null) inode=13728 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=37 name=(null) inode=13731 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=38 name=(null) inode=13728 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=39 name=(null) inode=13732 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=40 name=(null) inode=13728 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=41 name=(null) inode=13733 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=42 name=(null) inode=13713 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=43 name=(null) inode=13734 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=44 name=(null) inode=13734 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=45 name=(null) inode=13735 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.083576 systemd-networkd[1017]: eth0: Overlong DHCP hostname received, shortened from 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6.c.flatcar-212911.internal' to 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' Sep 6 00:33:51.083596 systemd-networkd[1017]: eth0: DHCPv4 address 10.128.0.94/32, gateway 10.128.0.1 acquired from 169.254.169.254 Sep 6 00:33:51.031000 audit: PATH item=46 name=(null) inode=13734 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=47 name=(null) inode=13736 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=48 name=(null) inode=13734 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=49 name=(null) inode=13737 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=50 name=(null) inode=13734 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=51 name=(null) inode=13738 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=52 name=(null) inode=13734 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=53 name=(null) inode=13739 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=54 name=(null) inode=40 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=55 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=56 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=57 name=(null) inode=13741 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=58 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=59 name=(null) inode=13742 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=60 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=61 name=(null) inode=13743 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=62 name=(null) inode=13743 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=63 name=(null) inode=13744 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=64 name=(null) inode=13743 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=65 name=(null) inode=13745 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=66 name=(null) inode=13743 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=67 name=(null) inode=13746 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=68 name=(null) inode=13743 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=69 name=(null) inode=13747 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=70 name=(null) inode=13743 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=71 name=(null) inode=13748 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=72 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=73 name=(null) inode=13749 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=74 name=(null) inode=13749 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=75 name=(null) inode=13750 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=76 name=(null) inode=13749 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=77 name=(null) inode=13751 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=78 name=(null) inode=13749 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=79 name=(null) inode=13752 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=80 name=(null) inode=13749 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=81 name=(null) inode=13753 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=82 name=(null) inode=13749 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=83 name=(null) inode=13754 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=84 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=85 name=(null) inode=13755 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=86 name=(null) inode=13755 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=87 name=(null) inode=13756 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=88 name=(null) inode=13755 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=89 name=(null) inode=13757 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=90 name=(null) inode=13755 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=91 name=(null) inode=13758 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=92 name=(null) inode=13755 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=93 name=(null) inode=13759 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=94 name=(null) inode=13755 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=95 name=(null) inode=13760 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=96 name=(null) inode=13740 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=97 name=(null) inode=13761 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=98 name=(null) inode=13761 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=99 name=(null) inode=13762 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=100 name=(null) inode=13761 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=101 name=(null) inode=13763 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=102 name=(null) inode=13761 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=103 name=(null) inode=13764 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=104 name=(null) inode=13761 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=105 name=(null) inode=13765 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=106 name=(null) inode=13761 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=107 name=(null) inode=13766 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PATH item=109 name=(null) inode=13767 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Sep 6 00:33:51.031000 audit: PROCTITLE proctitle="(udev-worker)" Sep 6 00:33:51.115431 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Sep 6 00:33:51.130427 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Sep 6 00:33:51.148478 kernel: EDAC MC: Ver: 3.0.0 Sep 6 00:33:51.184430 kernel: mousedev: PS/2 mouse device common for all mice Sep 6 00:33:51.212168 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Sep 6 00:33:51.221017 systemd[1]: Finished systemd-udev-settle.service. Sep 6 00:33:51.228000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.231176 systemd[1]: Starting lvm2-activation-early.service... Sep 6 00:33:51.258954 lvm[1043]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 6 00:33:51.294598 systemd[1]: Finished lvm2-activation-early.service. Sep 6 00:33:51.302000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.303974 systemd[1]: Reached target cryptsetup.target. Sep 6 00:33:51.314626 systemd[1]: Starting lvm2-activation.service... Sep 6 00:33:51.320779 lvm[1044]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Sep 6 00:33:51.350818 systemd[1]: Finished lvm2-activation.service. Sep 6 00:33:51.358000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.359773 systemd[1]: Reached target local-fs-pre.target. Sep 6 00:33:51.368679 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Sep 6 00:33:51.368827 systemd[1]: Reached target local-fs.target. Sep 6 00:33:51.377635 systemd[1]: Reached target machines.target. Sep 6 00:33:51.388427 systemd[1]: Starting ldconfig.service... Sep 6 00:33:51.396869 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 6 00:33:51.397107 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:51.398753 systemd[1]: Starting systemd-boot-update.service... Sep 6 00:33:51.407208 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Sep 6 00:33:51.418430 systemd[1]: Starting systemd-machine-id-commit.service... Sep 6 00:33:51.428412 systemd[1]: Starting systemd-sysext.service... Sep 6 00:33:51.436194 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1046 (bootctl) Sep 6 00:33:51.438168 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Sep 6 00:33:51.455438 systemd[1]: Unmounting usr-share-oem.mount... Sep 6 00:33:51.465697 systemd[1]: usr-share-oem.mount: Deactivated successfully. Sep 6 00:33:51.465960 systemd[1]: Unmounted usr-share-oem.mount. Sep 6 00:33:51.490506 kernel: loop0: detected capacity change from 0 to 221472 Sep 6 00:33:51.492620 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Sep 6 00:33:51.491000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.599664 systemd-fsck[1059]: fsck.fat 4.2 (2021-01-31) Sep 6 00:33:51.599664 systemd-fsck[1059]: /dev/sda1: 790 files, 120761/258078 clusters Sep 6 00:33:51.603496 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Sep 6 00:33:51.612000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.615689 systemd[1]: Mounting boot.mount... Sep 6 00:33:51.696474 systemd[1]: Mounted boot.mount. Sep 6 00:33:51.723301 systemd[1]: Finished systemd-boot-update.service. Sep 6 00:33:51.730000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.821304 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Sep 6 00:33:51.822210 systemd[1]: Finished systemd-machine-id-commit.service. Sep 6 00:33:51.829000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:51.858686 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Sep 6 00:33:51.884464 kernel: loop1: detected capacity change from 0 to 221472 Sep 6 00:33:51.909807 (sd-sysext)[1064]: Using extensions 'kubernetes'. Sep 6 00:33:51.911927 (sd-sysext)[1064]: Merged extensions into '/usr'. Sep 6 00:33:51.938073 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:51.941610 systemd[1]: Mounting usr-share-oem.mount... Sep 6 00:33:51.947638 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 6 00:33:51.951594 systemd[1]: Starting modprobe@dm_mod.service... Sep 6 00:33:51.962175 systemd[1]: Starting modprobe@efi_pstore.service... Sep 6 00:33:51.972180 systemd[1]: Starting modprobe@loop.service... Sep 6 00:33:51.979609 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 6 00:33:51.979857 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:51.980063 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:51.985329 systemd[1]: Mounted usr-share-oem.mount. Sep 6 00:33:51.993122 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 6 00:33:51.993352 systemd[1]: Finished modprobe@dm_mod.service. Sep 6 00:33:52.000000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.000000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.002354 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 6 00:33:52.002614 systemd[1]: Finished modprobe@efi_pstore.service. Sep 6 00:33:52.006357 ldconfig[1045]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Sep 6 00:33:52.010000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.010000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.012373 systemd[1]: Finished ldconfig.service. Sep 6 00:33:52.018000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.020190 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 6 00:33:52.020430 systemd[1]: Finished modprobe@loop.service. Sep 6 00:33:52.027000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.027000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.029776 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 6 00:33:52.029962 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.031693 systemd[1]: Finished systemd-sysext.service. Sep 6 00:33:52.039000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.042613 systemd[1]: Starting ensure-sysext.service... Sep 6 00:33:52.051101 systemd[1]: Starting systemd-tmpfiles-setup.service... Sep 6 00:33:52.063002 systemd[1]: Reloading. Sep 6 00:33:52.069513 systemd-tmpfiles[1071]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Sep 6 00:33:52.073338 systemd-tmpfiles[1071]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Sep 6 00:33:52.077295 systemd-tmpfiles[1071]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Sep 6 00:33:52.202936 /usr/lib/systemd/system-generators/torcx-generator[1091]: time="2025-09-06T00:33:52Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 6 00:33:52.203007 /usr/lib/systemd/system-generators/torcx-generator[1091]: time="2025-09-06T00:33:52Z" level=info msg="torcx already run" Sep 6 00:33:52.364231 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 6 00:33:52.364257 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 6 00:33:52.388871 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 6 00:33:52.466000 audit: BPF prog-id=27 op=LOAD Sep 6 00:33:52.466000 audit: BPF prog-id=23 op=UNLOAD Sep 6 00:33:52.467000 audit: BPF prog-id=28 op=LOAD Sep 6 00:33:52.467000 audit: BPF prog-id=18 op=UNLOAD Sep 6 00:33:52.467000 audit: BPF prog-id=29 op=LOAD Sep 6 00:33:52.467000 audit: BPF prog-id=30 op=LOAD Sep 6 00:33:52.467000 audit: BPF prog-id=19 op=UNLOAD Sep 6 00:33:52.467000 audit: BPF prog-id=20 op=UNLOAD Sep 6 00:33:52.468000 audit: BPF prog-id=31 op=LOAD Sep 6 00:33:52.469000 audit: BPF prog-id=24 op=UNLOAD Sep 6 00:33:52.469000 audit: BPF prog-id=32 op=LOAD Sep 6 00:33:52.469000 audit: BPF prog-id=33 op=LOAD Sep 6 00:33:52.469000 audit: BPF prog-id=25 op=UNLOAD Sep 6 00:33:52.469000 audit: BPF prog-id=26 op=UNLOAD Sep 6 00:33:52.471000 audit: BPF prog-id=34 op=LOAD Sep 6 00:33:52.471000 audit: BPF prog-id=35 op=LOAD Sep 6 00:33:52.471000 audit: BPF prog-id=21 op=UNLOAD Sep 6 00:33:52.471000 audit: BPF prog-id=22 op=UNLOAD Sep 6 00:33:52.482891 systemd[1]: Finished systemd-tmpfiles-setup.service. Sep 6 00:33:52.491000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.500050 systemd[1]: Starting audit-rules.service... Sep 6 00:33:52.510441 systemd[1]: Starting clean-ca-certificates.service... Sep 6 00:33:52.520846 systemd[1]: Starting oem-gce-enable-oslogin.service... Sep 6 00:33:52.531893 systemd[1]: Starting systemd-journal-catalog-update.service... Sep 6 00:33:52.540000 audit: BPF prog-id=36 op=LOAD Sep 6 00:33:52.543770 systemd[1]: Starting systemd-resolved.service... Sep 6 00:33:52.550000 audit: BPF prog-id=37 op=LOAD Sep 6 00:33:52.553249 systemd[1]: Starting systemd-timesyncd.service... Sep 6 00:33:52.562681 systemd[1]: Starting systemd-update-utmp.service... Sep 6 00:33:52.571386 systemd[1]: Finished clean-ca-certificates.service. Sep 6 00:33:52.573000 audit[1161]: SYSTEM_BOOT pid=1161 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.578000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Sep 6 00:33:52.580314 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Sep 6 00:33:52.580609 systemd[1]: Finished oem-gce-enable-oslogin.service. Sep 6 00:33:52.589089 augenrules[1165]: No rules Sep 6 00:33:52.587000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Sep 6 00:33:52.587000 audit[1165]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffcd8687c00 a2=420 a3=0 items=0 ppid=1135 pid=1165 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Sep 6 00:33:52.587000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Sep 6 00:33:52.590344 systemd[1]: Finished audit-rules.service. Sep 6 00:33:52.599267 systemd[1]: Finished systemd-journal-catalog-update.service. Sep 6 00:33:52.605585 systemd-networkd[1017]: eth0: Gained IPv6LL Sep 6 00:33:52.615347 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:52.615878 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.618351 systemd[1]: Starting modprobe@dm_mod.service... Sep 6 00:33:52.628866 systemd[1]: Starting modprobe@efi_pstore.service... Sep 6 00:33:52.637855 systemd[1]: Starting modprobe@loop.service... Sep 6 00:33:52.648292 systemd[1]: Starting oem-gce-enable-oslogin.service... Sep 6 00:33:52.654756 enable-oslogin[1173]: /etc/pam.d/sshd already exists. Not enabling OS Login Sep 6 00:33:52.657654 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.658046 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:52.661018 systemd[1]: Starting systemd-update-done.service... Sep 6 00:33:52.668519 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 6 00:33:52.668831 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:52.672622 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 6 00:33:52.672877 systemd[1]: Finished modprobe@dm_mod.service. Sep 6 00:33:52.682588 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 6 00:33:52.682825 systemd[1]: Finished modprobe@efi_pstore.service. Sep 6 00:33:52.692702 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 6 00:33:52.692965 systemd[1]: Finished modprobe@loop.service. Sep 6 00:33:52.702559 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Sep 6 00:33:52.702831 systemd[1]: Finished oem-gce-enable-oslogin.service. Sep 6 00:33:52.712833 systemd[1]: Finished systemd-update-done.service. Sep 6 00:33:52.724316 systemd[1]: Finished systemd-update-utmp.service. Sep 6 00:33:52.735538 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:52.735983 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.738369 systemd[1]: Starting modprobe@dm_mod.service... Sep 6 00:33:52.747457 systemd[1]: Starting modprobe@efi_pstore.service... Sep 6 00:33:52.756650 systemd[1]: Starting modprobe@loop.service... Sep 6 00:33:52.765652 systemd[1]: Starting oem-gce-enable-oslogin.service... Sep 6 00:33:52.769571 systemd-resolved[1154]: Positive Trust Anchors: Sep 6 00:33:52.769995 systemd-resolved[1154]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Sep 6 00:33:52.770170 systemd-resolved[1154]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Sep 6 00:33:52.771526 enable-oslogin[1178]: /etc/pam.d/sshd already exists. Not enabling OS Login Sep 6 00:33:52.774611 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.774884 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:52.775078 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 6 00:33:52.775226 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:52.777472 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 6 00:33:52.777723 systemd[1]: Finished modprobe@dm_mod.service. Sep 6 00:33:52.784988 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 6 00:33:52.785206 systemd[1]: Finished modprobe@efi_pstore.service. Sep 6 00:33:52.787606 systemd-resolved[1154]: Defaulting to hostname 'linux'. Sep 6 00:33:52.794119 systemd[1]: Started systemd-resolved.service. Sep 6 00:33:52.803222 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 6 00:33:52.803455 systemd[1]: Finished modprobe@loop.service. Sep 6 00:33:52.812308 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Sep 6 00:33:52.812598 systemd[1]: Finished oem-gce-enable-oslogin.service. Sep 6 00:33:52.817752 systemd-timesyncd[1158]: Contacted time server 169.254.169.254:123 (169.254.169.254). Sep 6 00:33:52.817852 systemd-timesyncd[1158]: Initial clock synchronization to Sat 2025-09-06 00:33:52.725642 UTC. Sep 6 00:33:52.821975 systemd[1]: Started systemd-timesyncd.service. Sep 6 00:33:52.836236 systemd[1]: Reached target network.target. Sep 6 00:33:52.844773 systemd[1]: Reached target nss-lookup.target. Sep 6 00:33:52.853838 systemd[1]: Reached target time-set.target. Sep 6 00:33:52.862783 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:52.863279 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.865371 systemd[1]: Starting modprobe@dm_mod.service... Sep 6 00:33:52.874650 systemd[1]: Starting modprobe@drm.service... Sep 6 00:33:52.883709 systemd[1]: Starting modprobe@efi_pstore.service... Sep 6 00:33:52.895459 systemd[1]: Starting modprobe@loop.service... Sep 6 00:33:52.904645 systemd[1]: Starting oem-gce-enable-oslogin.service... Sep 6 00:33:52.909299 enable-oslogin[1184]: /etc/pam.d/sshd already exists. Not enabling OS Login Sep 6 00:33:52.913706 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Sep 6 00:33:52.913990 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:52.916062 systemd[1]: Starting systemd-networkd-wait-online.service... Sep 6 00:33:52.924777 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Sep 6 00:33:52.925043 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Sep 6 00:33:52.927914 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Sep 6 00:33:52.928154 systemd[1]: Finished modprobe@dm_mod.service. Sep 6 00:33:52.937284 systemd[1]: modprobe@drm.service: Deactivated successfully. Sep 6 00:33:52.937527 systemd[1]: Finished modprobe@drm.service. Sep 6 00:33:52.947255 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Sep 6 00:33:52.947500 systemd[1]: Finished modprobe@efi_pstore.service. Sep 6 00:33:52.956203 systemd[1]: modprobe@loop.service: Deactivated successfully. Sep 6 00:33:52.956464 systemd[1]: Finished modprobe@loop.service. Sep 6 00:33:52.966241 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Sep 6 00:33:52.966528 systemd[1]: Finished oem-gce-enable-oslogin.service. Sep 6 00:33:52.975228 systemd[1]: Finished systemd-networkd-wait-online.service. Sep 6 00:33:52.987664 systemd[1]: Reached target network-online.target. Sep 6 00:33:52.996726 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Sep 6 00:33:52.996796 systemd[1]: Reached target sysinit.target. Sep 6 00:33:53.005731 systemd[1]: Started motdgen.path. Sep 6 00:33:53.012660 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Sep 6 00:33:53.022826 systemd[1]: Started logrotate.timer. Sep 6 00:33:53.029700 systemd[1]: Started mdadm.timer. Sep 6 00:33:53.036595 systemd[1]: Started systemd-tmpfiles-clean.timer. Sep 6 00:33:53.045595 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Sep 6 00:33:53.045664 systemd[1]: Reached target paths.target. Sep 6 00:33:53.052705 systemd[1]: Reached target timers.target. Sep 6 00:33:53.060356 systemd[1]: Listening on dbus.socket. Sep 6 00:33:53.069027 systemd[1]: Starting docker.socket... Sep 6 00:33:53.080587 systemd[1]: Listening on sshd.socket. Sep 6 00:33:53.087736 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:53.087839 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Sep 6 00:33:53.089106 systemd[1]: Finished ensure-sysext.service. Sep 6 00:33:53.097888 systemd[1]: Listening on docker.socket. Sep 6 00:33:53.106129 systemd[1]: Reached target sockets.target. Sep 6 00:33:53.114605 systemd[1]: Reached target basic.target. Sep 6 00:33:53.121703 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 6 00:33:53.121757 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Sep 6 00:33:53.123979 systemd[1]: Starting containerd.service... Sep 6 00:33:53.134215 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Sep 6 00:33:53.145892 systemd[1]: Starting dbus.service... Sep 6 00:33:53.153377 systemd[1]: Starting enable-oem-cloudinit.service... Sep 6 00:33:53.162543 systemd[1]: Starting extend-filesystems.service... Sep 6 00:33:53.171336 jq[1191]: false Sep 6 00:33:53.169611 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Sep 6 00:33:53.172154 systemd[1]: Starting kubelet.service... Sep 6 00:33:53.182495 systemd[1]: Starting motdgen.service... Sep 6 00:33:53.192439 systemd[1]: Starting oem-gce.service... Sep 6 00:33:53.201456 systemd[1]: Starting prepare-helm.service... Sep 6 00:33:53.210541 systemd[1]: Starting ssh-key-proc-cmdline.service... Sep 6 00:33:53.217025 extend-filesystems[1192]: Found loop1 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda1 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda2 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda3 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found usr Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda4 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda6 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda7 Sep 6 00:33:53.217025 extend-filesystems[1192]: Found sda9 Sep 6 00:33:53.217025 extend-filesystems[1192]: Checking size of /dev/sda9 Sep 6 00:33:53.220816 systemd[1]: Starting sshd-keygen.service... Sep 6 00:33:53.290351 dbus-daemon[1190]: [system] SELinux support is enabled Sep 6 00:33:53.232105 systemd[1]: Starting systemd-logind.service... Sep 6 00:33:53.296841 dbus-daemon[1190]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.1' (uid=244 pid=1017 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Sep 6 00:33:53.239062 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Sep 6 00:33:53.301700 dbus-daemon[1190]: [system] Successfully activated service 'org.freedesktop.systemd1' Sep 6 00:33:53.239181 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionSecurity=!tpm2). Sep 6 00:33:53.240128 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Sep 6 00:33:53.347157 jq[1216]: true Sep 6 00:33:53.241717 systemd[1]: Starting update-engine.service... Sep 6 00:33:53.255520 systemd[1]: Starting update-ssh-keys-after-ignition.service... Sep 6 00:33:53.268780 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Sep 6 00:33:53.269195 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Sep 6 00:33:53.275804 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Sep 6 00:33:53.349732 jq[1221]: true Sep 6 00:33:53.276340 systemd[1]: Finished ssh-key-proc-cmdline.service. Sep 6 00:33:53.350346 mkfs.ext4[1223]: mke2fs 1.46.5 (30-Dec-2021) Sep 6 00:33:53.350346 mkfs.ext4[1223]: Discarding device blocks: 0/262144\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008 \u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008done Sep 6 00:33:53.350346 mkfs.ext4[1223]: Creating filesystem with 262144 4k blocks and 65536 inodes Sep 6 00:33:53.350346 mkfs.ext4[1223]: Filesystem UUID: 353b71f4-971a-45b4-9fb2-8261bd87ecfd Sep 6 00:33:53.350346 mkfs.ext4[1223]: Superblock backups stored on blocks: Sep 6 00:33:53.350346 mkfs.ext4[1223]: 32768, 98304, 163840, 229376 Sep 6 00:33:53.350346 mkfs.ext4[1223]: Allocating group tables: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done Sep 6 00:33:53.350346 mkfs.ext4[1223]: Writing inode tables: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done Sep 6 00:33:53.350346 mkfs.ext4[1223]: Creating journal (8192 blocks): done Sep 6 00:33:53.350346 mkfs.ext4[1223]: Writing superblocks and filesystem accounting information: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done Sep 6 00:33:53.290684 systemd[1]: Started dbus.service. Sep 6 00:33:53.300216 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Sep 6 00:33:53.300260 systemd[1]: Reached target system-config.target. Sep 6 00:33:53.309139 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Sep 6 00:33:53.309180 systemd[1]: Reached target user-config.target. Sep 6 00:33:53.332691 systemd[1]: Starting systemd-hostnamed.service... Sep 6 00:33:53.369639 extend-filesystems[1192]: Resized partition /dev/sda9 Sep 6 00:33:53.385071 extend-filesystems[1232]: resize2fs 1.46.5 (30-Dec-2021) Sep 6 00:33:53.408643 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 2538491 blocks Sep 6 00:33:53.370725 systemd[1]: motdgen.service: Deactivated successfully. Sep 6 00:33:53.408866 umount[1238]: umount: /var/lib/flatcar-oem-gce.img: not mounted. Sep 6 00:33:53.371014 systemd[1]: Finished motdgen.service. Sep 6 00:33:53.414473 tar[1220]: linux-amd64/helm Sep 6 00:33:53.432450 kernel: loop2: detected capacity change from 0 to 2097152 Sep 6 00:33:53.483704 update_engine[1212]: I0906 00:33:53.483649 1212 main.cc:92] Flatcar Update Engine starting Sep 6 00:33:53.490857 systemd[1]: Started update-engine.service. Sep 6 00:33:53.491600 update_engine[1212]: I0906 00:33:53.491568 1212 update_check_scheduler.cc:74] Next update check in 11m31s Sep 6 00:33:53.502334 systemd[1]: Started locksmithd.service. Sep 6 00:33:53.519978 kernel: EXT4-fs (sda9): resized filesystem to 2538491 Sep 6 00:33:53.552500 extend-filesystems[1232]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Sep 6 00:33:53.552500 extend-filesystems[1232]: old_desc_blocks = 1, new_desc_blocks = 2 Sep 6 00:33:53.552500 extend-filesystems[1232]: The filesystem on /dev/sda9 is now 2538491 (4k) blocks long. Sep 6 00:33:53.630732 kernel: EXT4-fs (loop2): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Sep 6 00:33:53.631141 bash[1255]: Updated "/home/core/.ssh/authorized_keys" Sep 6 00:33:53.554549 systemd[1]: extend-filesystems.service: Deactivated successfully. Sep 6 00:33:53.631463 extend-filesystems[1192]: Resized filesystem in /dev/sda9 Sep 6 00:33:53.554862 systemd[1]: Finished extend-filesystems.service. Sep 6 00:33:53.596788 systemd[1]: Finished update-ssh-keys-after-ignition.service. Sep 6 00:33:53.649812 coreos-metadata[1189]: Sep 06 00:33:53.649 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/sshKeys: Attempt #1 Sep 6 00:33:53.652981 coreos-metadata[1189]: Sep 06 00:33:53.652 INFO Fetch failed with 404: resource not found Sep 6 00:33:53.652981 coreos-metadata[1189]: Sep 06 00:33:53.652 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/ssh-keys: Attempt #1 Sep 6 00:33:53.652981 coreos-metadata[1189]: Sep 06 00:33:53.652 INFO Fetch successful Sep 6 00:33:53.652981 coreos-metadata[1189]: Sep 06 00:33:53.652 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/block-project-ssh-keys: Attempt #1 Sep 6 00:33:53.654087 coreos-metadata[1189]: Sep 06 00:33:53.653 INFO Fetch failed with 404: resource not found Sep 6 00:33:53.654087 coreos-metadata[1189]: Sep 06 00:33:53.653 INFO Fetching http://169.254.169.254/computeMetadata/v1/project/attributes/sshKeys: Attempt #1 Sep 6 00:33:53.654087 coreos-metadata[1189]: Sep 06 00:33:53.653 INFO Fetch failed with 404: resource not found Sep 6 00:33:53.654087 coreos-metadata[1189]: Sep 06 00:33:53.653 INFO Fetching http://169.254.169.254/computeMetadata/v1/project/attributes/ssh-keys: Attempt #1 Sep 6 00:33:53.654467 coreos-metadata[1189]: Sep 06 00:33:53.654 INFO Fetch successful Sep 6 00:33:53.656275 unknown[1189]: wrote ssh authorized keys file for user: core Sep 6 00:33:53.676103 update-ssh-keys[1263]: Updated "/home/core/.ssh/authorized_keys" Sep 6 00:33:53.676734 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Sep 6 00:33:53.695948 systemd-logind[1211]: Watching system buttons on /dev/input/event1 (Power Button) Sep 6 00:33:53.696306 systemd-logind[1211]: Watching system buttons on /dev/input/event2 (Sleep Button) Sep 6 00:33:53.696458 systemd-logind[1211]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Sep 6 00:33:53.700583 systemd-logind[1211]: New seat seat0. Sep 6 00:33:53.723652 systemd[1]: Started systemd-logind.service. Sep 6 00:33:53.736302 dbus-daemon[1190]: [system] Successfully activated service 'org.freedesktop.hostname1' Sep 6 00:33:53.736510 systemd[1]: Started systemd-hostnamed.service. Sep 6 00:33:53.737590 dbus-daemon[1190]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=1227 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Sep 6 00:33:53.748704 env[1222]: time="2025-09-06T00:33:53.748641982Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Sep 6 00:33:53.750914 systemd[1]: Starting polkit.service... Sep 6 00:33:53.872995 env[1222]: time="2025-09-06T00:33:53.872934369Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Sep 6 00:33:53.873157 env[1222]: time="2025-09-06T00:33:53.873121242Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Sep 6 00:33:53.875275 env[1222]: time="2025-09-06T00:33:53.875223971Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.190-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Sep 6 00:33:53.875275 env[1222]: time="2025-09-06T00:33:53.875271892Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Sep 6 00:33:53.875714 env[1222]: time="2025-09-06T00:33:53.875634149Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 6 00:33:53.875714 env[1222]: time="2025-09-06T00:33:53.875675358Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Sep 6 00:33:53.875714 env[1222]: time="2025-09-06T00:33:53.875697391Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Sep 6 00:33:53.875714 env[1222]: time="2025-09-06T00:33:53.875714434Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Sep 6 00:33:53.875954 env[1222]: time="2025-09-06T00:33:53.875833629Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Sep 6 00:33:53.876206 env[1222]: time="2025-09-06T00:33:53.876169045Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Sep 6 00:33:53.876512 env[1222]: time="2025-09-06T00:33:53.876471102Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Sep 6 00:33:53.876595 env[1222]: time="2025-09-06T00:33:53.876511767Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Sep 6 00:33:53.876660 env[1222]: time="2025-09-06T00:33:53.876596516Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Sep 6 00:33:53.876660 env[1222]: time="2025-09-06T00:33:53.876618354Z" level=info msg="metadata content store policy set" policy=shared Sep 6 00:33:53.885587 env[1222]: time="2025-09-06T00:33:53.885467601Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Sep 6 00:33:53.885587 env[1222]: time="2025-09-06T00:33:53.885519007Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Sep 6 00:33:53.885587 env[1222]: time="2025-09-06T00:33:53.885543037Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885601274Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885625615Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885701696Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885727638Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885751160Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885773357Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.885800 env[1222]: time="2025-09-06T00:33:53.885795917Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.886115 env[1222]: time="2025-09-06T00:33:53.885817377Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.886115 env[1222]: time="2025-09-06T00:33:53.885839129Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Sep 6 00:33:53.886115 env[1222]: time="2025-09-06T00:33:53.885978082Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Sep 6 00:33:53.886115 env[1222]: time="2025-09-06T00:33:53.886099316Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Sep 6 00:33:53.886755 env[1222]: time="2025-09-06T00:33:53.886719808Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Sep 6 00:33:53.886857 env[1222]: time="2025-09-06T00:33:53.886772476Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.886857 env[1222]: time="2025-09-06T00:33:53.886796123Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Sep 6 00:33:53.886956 env[1222]: time="2025-09-06T00:33:53.886886589Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.886956 env[1222]: time="2025-09-06T00:33:53.886912575Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887063 env[1222]: time="2025-09-06T00:33:53.887013410Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887063 env[1222]: time="2025-09-06T00:33:53.887047480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887159 env[1222]: time="2025-09-06T00:33:53.887070215Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887159 env[1222]: time="2025-09-06T00:33:53.887094100Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887159 env[1222]: time="2025-09-06T00:33:53.887115508Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887159 env[1222]: time="2025-09-06T00:33:53.887136174Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887347 env[1222]: time="2025-09-06T00:33:53.887159173Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Sep 6 00:33:53.887435 env[1222]: time="2025-09-06T00:33:53.887346679Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887435 env[1222]: time="2025-09-06T00:33:53.887374526Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887435 env[1222]: time="2025-09-06T00:33:53.887420272Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.887588 env[1222]: time="2025-09-06T00:33:53.887443849Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Sep 6 00:33:53.887588 env[1222]: time="2025-09-06T00:33:53.887468400Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Sep 6 00:33:53.887588 env[1222]: time="2025-09-06T00:33:53.887489513Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Sep 6 00:33:53.887588 env[1222]: time="2025-09-06T00:33:53.887517477Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Sep 6 00:33:53.887588 env[1222]: time="2025-09-06T00:33:53.887566755Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Sep 6 00:33:53.888840 env[1222]: time="2025-09-06T00:33:53.887892182Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Sep 6 00:33:53.888840 env[1222]: time="2025-09-06T00:33:53.888005017Z" level=info msg="Connect containerd service" Sep 6 00:33:53.888840 env[1222]: time="2025-09-06T00:33:53.888049448Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Sep 6 00:33:53.898222 polkitd[1265]: Started polkitd version 121 Sep 6 00:33:53.902512 env[1222]: time="2025-09-06T00:33:53.898920873Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 6 00:33:53.902512 env[1222]: time="2025-09-06T00:33:53.900199221Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Sep 6 00:33:53.902512 env[1222]: time="2025-09-06T00:33:53.900268138Z" level=info msg=serving... address=/run/containerd/containerd.sock Sep 6 00:33:53.902512 env[1222]: time="2025-09-06T00:33:53.901342775Z" level=info msg="containerd successfully booted in 0.160103s" Sep 6 00:33:53.900471 systemd[1]: Started containerd.service. Sep 6 00:33:53.906889 env[1222]: time="2025-09-06T00:33:53.906835206Z" level=info msg="Start subscribing containerd event" Sep 6 00:33:53.922887 polkitd[1265]: Loading rules from directory /etc/polkit-1/rules.d Sep 6 00:33:53.922984 polkitd[1265]: Loading rules from directory /usr/share/polkit-1/rules.d Sep 6 00:33:53.926373 polkitd[1265]: Finished loading, compiling and executing 2 rules Sep 6 00:33:53.926996 dbus-daemon[1190]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Sep 6 00:33:53.927537 polkitd[1265]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Sep 6 00:33:53.927228 systemd[1]: Started polkit.service. Sep 6 00:33:53.941085 env[1222]: time="2025-09-06T00:33:53.941029775Z" level=info msg="Start recovering state" Sep 6 00:33:53.941357 env[1222]: time="2025-09-06T00:33:53.941331165Z" level=info msg="Start event monitor" Sep 6 00:33:53.947289 env[1222]: time="2025-09-06T00:33:53.947247478Z" level=info msg="Start snapshots syncer" Sep 6 00:33:53.947476 env[1222]: time="2025-09-06T00:33:53.947439008Z" level=info msg="Start cni network conf syncer for default" Sep 6 00:33:53.947608 env[1222]: time="2025-09-06T00:33:53.947583971Z" level=info msg="Start streaming server" Sep 6 00:33:53.949764 systemd-hostnamed[1227]: Hostname set to (transient) Sep 6 00:33:53.953412 systemd-resolved[1154]: System hostname changed to 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6'. Sep 6 00:33:55.098693 tar[1220]: linux-amd64/LICENSE Sep 6 00:33:55.099366 tar[1220]: linux-amd64/README.md Sep 6 00:33:55.115323 systemd[1]: Finished prepare-helm.service. Sep 6 00:33:55.739466 systemd[1]: Started kubelet.service. Sep 6 00:33:56.647255 locksmithd[1256]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Sep 6 00:33:57.056333 kubelet[1277]: E0906 00:33:57.056279 1277 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 6 00:33:57.059044 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 6 00:33:57.059259 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 6 00:33:57.059640 systemd[1]: kubelet.service: Consumed 1.587s CPU time. Sep 6 00:33:59.921454 sshd_keygen[1225]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Sep 6 00:33:59.962552 systemd[1]: Finished sshd-keygen.service. Sep 6 00:33:59.974033 systemd[1]: Starting issuegen.service... Sep 6 00:33:59.985802 systemd[1]: issuegen.service: Deactivated successfully. Sep 6 00:33:59.986069 systemd[1]: Finished issuegen.service. Sep 6 00:33:59.996008 systemd[1]: Starting systemd-user-sessions.service... Sep 6 00:34:00.009166 systemd[1]: Finished systemd-user-sessions.service. Sep 6 00:34:00.019787 systemd[1]: Started getty@tty1.service. Sep 6 00:34:00.028888 systemd[1]: Started serial-getty@ttyS0.service. Sep 6 00:34:00.038900 systemd[1]: Reached target getty.target. Sep 6 00:34:00.314724 systemd[1]: var-lib-flatcar\x2doem\x2dgce.mount: Deactivated successfully. Sep 6 00:34:01.851429 systemd[1]: Created slice system-sshd.slice. Sep 6 00:34:01.863277 systemd[1]: Started sshd@0-10.128.0.94:22-139.178.89.65:51140.service. Sep 6 00:34:02.154464 kernel: loop2: detected capacity change from 0 to 2097152 Sep 6 00:34:02.172513 systemd-nspawn[1304]: Spawning container oem-gce on /var/lib/flatcar-oem-gce.img. Sep 6 00:34:02.172513 systemd-nspawn[1304]: Press ^] three times within 1s to kill container. Sep 6 00:34:02.185423 kernel: EXT4-fs (loop2): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Sep 6 00:34:02.188518 sshd[1302]: Accepted publickey for core from 139.178.89.65 port 51140 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:34:02.192646 sshd[1302]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:34:02.219596 systemd[1]: Created slice user-500.slice. Sep 6 00:34:02.219959 systemd-logind[1211]: New session 1 of user core. Sep 6 00:34:02.228698 systemd[1]: Starting user-runtime-dir@500.service... Sep 6 00:34:02.251283 systemd[1]: Finished user-runtime-dir@500.service. Sep 6 00:34:02.262739 systemd[1]: Starting user@500.service... Sep 6 00:34:02.287215 (systemd)[1310]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:34:02.305488 systemd[1]: Started oem-gce.service. Sep 6 00:34:02.314013 systemd[1]: Reached target multi-user.target. Sep 6 00:34:02.324377 systemd[1]: Starting systemd-update-utmp-runlevel.service... Sep 6 00:34:02.338758 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Sep 6 00:34:02.339008 systemd[1]: Finished systemd-update-utmp-runlevel.service. Sep 6 00:34:02.372989 systemd-nspawn[1304]: + '[' -e /etc/default/instance_configs.cfg.template ']' Sep 6 00:34:02.373335 systemd-nspawn[1304]: + echo -e '[InstanceSetup]\nset_host_keys = false' Sep 6 00:34:02.373730 systemd-nspawn[1304]: + /usr/bin/google_instance_setup Sep 6 00:34:02.431386 systemd[1310]: Queued start job for default target default.target. Sep 6 00:34:02.433113 systemd[1310]: Reached target paths.target. Sep 6 00:34:02.433161 systemd[1310]: Reached target sockets.target. Sep 6 00:34:02.433186 systemd[1310]: Reached target timers.target. Sep 6 00:34:02.433208 systemd[1310]: Reached target basic.target. Sep 6 00:34:02.433367 systemd[1]: Started user@500.service. Sep 6 00:34:02.433552 systemd[1310]: Reached target default.target. Sep 6 00:34:02.433646 systemd[1310]: Startup finished in 130ms. Sep 6 00:34:02.443429 systemd[1]: Started session-1.scope. Sep 6 00:34:02.450774 systemd[1]: Startup finished in 1.046s (kernel) + 8.375s (initrd) + 17.262s (userspace) = 26.685s. Sep 6 00:34:02.685587 systemd[1]: Started sshd@1-10.128.0.94:22-139.178.89.65:51148.service. Sep 6 00:34:02.981929 sshd[1321]: Accepted publickey for core from 139.178.89.65 port 51148 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:34:02.983540 sshd[1321]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:34:02.991543 systemd-logind[1211]: New session 2 of user core. Sep 6 00:34:02.992582 systemd[1]: Started session-2.scope. Sep 6 00:34:03.090348 instance-setup[1317]: INFO Running google_set_multiqueue. Sep 6 00:34:03.104184 instance-setup[1317]: INFO Set channels for eth0 to 2. Sep 6 00:34:03.107951 instance-setup[1317]: INFO Setting /proc/irq/31/smp_affinity_list to 0 for device virtio1. Sep 6 00:34:03.109454 instance-setup[1317]: INFO /proc/irq/31/smp_affinity_list: real affinity 0 Sep 6 00:34:03.109752 instance-setup[1317]: INFO Setting /proc/irq/32/smp_affinity_list to 0 for device virtio1. Sep 6 00:34:03.111313 instance-setup[1317]: INFO /proc/irq/32/smp_affinity_list: real affinity 0 Sep 6 00:34:03.111700 instance-setup[1317]: INFO Setting /proc/irq/33/smp_affinity_list to 1 for device virtio1. Sep 6 00:34:03.113050 instance-setup[1317]: INFO /proc/irq/33/smp_affinity_list: real affinity 1 Sep 6 00:34:03.113511 instance-setup[1317]: INFO Setting /proc/irq/34/smp_affinity_list to 1 for device virtio1. Sep 6 00:34:03.114940 instance-setup[1317]: INFO /proc/irq/34/smp_affinity_list: real affinity 1 Sep 6 00:34:03.129158 instance-setup[1317]: INFO Queue 0 XPS=1 for /sys/class/net/eth0/queues/tx-0/xps_cpus Sep 6 00:34:03.129314 instance-setup[1317]: INFO Queue 1 XPS=2 for /sys/class/net/eth0/queues/tx-1/xps_cpus Sep 6 00:34:03.176210 systemd-nspawn[1304]: + /usr/bin/google_metadata_script_runner --script-type startup Sep 6 00:34:03.201383 sshd[1321]: pam_unix(sshd:session): session closed for user core Sep 6 00:34:03.206261 systemd[1]: sshd@1-10.128.0.94:22-139.178.89.65:51148.service: Deactivated successfully. Sep 6 00:34:03.207505 systemd[1]: session-2.scope: Deactivated successfully. Sep 6 00:34:03.209947 systemd-logind[1211]: Session 2 logged out. Waiting for processes to exit. Sep 6 00:34:03.211969 systemd-logind[1211]: Removed session 2. Sep 6 00:34:03.247621 systemd[1]: Started sshd@2-10.128.0.94:22-139.178.89.65:51156.service. Sep 6 00:34:03.548565 startup-script[1355]: INFO Starting startup scripts. Sep 6 00:34:03.557039 sshd[1358]: Accepted publickey for core from 139.178.89.65 port 51156 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:34:03.558110 sshd[1358]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:34:03.565943 systemd[1]: Started session-3.scope. Sep 6 00:34:03.566825 systemd-logind[1211]: New session 3 of user core. Sep 6 00:34:03.570300 startup-script[1355]: INFO No startup scripts found in metadata. Sep 6 00:34:03.570531 startup-script[1355]: INFO Finished running startup scripts. Sep 6 00:34:03.612429 systemd-nspawn[1304]: + trap 'stopping=1 ; kill "${daemon_pids[@]}" || :' SIGTERM Sep 6 00:34:03.612429 systemd-nspawn[1304]: + daemon_pids=() Sep 6 00:34:03.612429 systemd-nspawn[1304]: + for d in accounts clock_skew network Sep 6 00:34:03.613469 systemd-nspawn[1304]: + daemon_pids+=($!) Sep 6 00:34:03.613469 systemd-nspawn[1304]: + for d in accounts clock_skew network Sep 6 00:34:03.613832 systemd-nspawn[1304]: + /usr/bin/google_accounts_daemon Sep 6 00:34:03.614329 systemd-nspawn[1304]: + daemon_pids+=($!) Sep 6 00:34:03.614738 systemd-nspawn[1304]: + for d in accounts clock_skew network Sep 6 00:34:03.615042 systemd-nspawn[1304]: + daemon_pids+=($!) Sep 6 00:34:03.615143 systemd-nspawn[1304]: + NOTIFY_SOCKET=/run/systemd/notify Sep 6 00:34:03.615143 systemd-nspawn[1304]: + /usr/bin/systemd-notify --ready Sep 6 00:34:03.615705 systemd-nspawn[1304]: + /usr/bin/google_clock_skew_daemon Sep 6 00:34:03.616189 systemd-nspawn[1304]: + /usr/bin/google_network_daemon Sep 6 00:34:03.673123 systemd-nspawn[1304]: + wait -n 36 37 38 Sep 6 00:34:03.771792 sshd[1358]: pam_unix(sshd:session): session closed for user core Sep 6 00:34:03.776383 systemd[1]: sshd@2-10.128.0.94:22-139.178.89.65:51156.service: Deactivated successfully. Sep 6 00:34:03.777563 systemd[1]: session-3.scope: Deactivated successfully. Sep 6 00:34:03.779923 systemd-logind[1211]: Session 3 logged out. Waiting for processes to exit. Sep 6 00:34:03.781460 systemd-logind[1211]: Removed session 3. Sep 6 00:34:03.817953 systemd[1]: Started sshd@3-10.128.0.94:22-139.178.89.65:51166.service. Sep 6 00:34:04.122794 sshd[1370]: Accepted publickey for core from 139.178.89.65 port 51166 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:34:04.124946 sshd[1370]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:34:04.134691 systemd[1]: Started session-4.scope. Sep 6 00:34:04.137506 systemd-logind[1211]: New session 4 of user core. Sep 6 00:34:04.330088 groupadd[1380]: group added to /etc/group: name=google-sudoers, GID=1000 Sep 6 00:34:04.332987 groupadd[1380]: group added to /etc/gshadow: name=google-sudoers Sep 6 00:34:04.345753 sshd[1370]: pam_unix(sshd:session): session closed for user core Sep 6 00:34:04.350296 systemd[1]: sshd@3-10.128.0.94:22-139.178.89.65:51166.service: Deactivated successfully. Sep 6 00:34:04.351462 systemd[1]: session-4.scope: Deactivated successfully. Sep 6 00:34:04.353239 systemd-logind[1211]: Session 4 logged out. Waiting for processes to exit. Sep 6 00:34:04.354523 systemd-logind[1211]: Removed session 4. Sep 6 00:34:04.365895 groupadd[1380]: new group: name=google-sudoers, GID=1000 Sep 6 00:34:04.370055 google-clock-skew[1364]: INFO Starting Google Clock Skew daemon. Sep 6 00:34:04.386729 google-clock-skew[1364]: INFO Clock drift token has changed: 0. Sep 6 00:34:04.391795 systemd[1]: Started sshd@4-10.128.0.94:22-139.178.89.65:51176.service. Sep 6 00:34:04.405864 google-accounts[1363]: INFO Starting Google Accounts daemon. Sep 6 00:34:04.411346 systemd-nspawn[1304]: hwclock: Cannot access the Hardware Clock via any known method. Sep 6 00:34:04.411346 systemd-nspawn[1304]: hwclock: Use the --verbose option to see the details of our search for an access method. Sep 6 00:34:04.412752 google-clock-skew[1364]: WARNING Failed to sync system time with hardware clock. Sep 6 00:34:04.445114 google-networking[1365]: INFO Starting Google Networking daemon. Sep 6 00:34:04.449264 google-accounts[1363]: WARNING OS Login not installed. Sep 6 00:34:04.450436 google-accounts[1363]: INFO Creating a new user account for 0. Sep 6 00:34:04.457061 systemd-nspawn[1304]: useradd: invalid user name '0': use --badname to ignore Sep 6 00:34:04.458025 google-accounts[1363]: WARNING Could not create user 0. Command '['useradd', '-m', '-s', '/bin/bash', '-p', '*', '0']' returned non-zero exit status 3.. Sep 6 00:34:04.693933 sshd[1390]: Accepted publickey for core from 139.178.89.65 port 51176 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:34:04.695311 sshd[1390]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:34:04.702522 systemd-logind[1211]: New session 5 of user core. Sep 6 00:34:04.702551 systemd[1]: Started session-5.scope. Sep 6 00:34:04.891281 sudo[1397]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Sep 6 00:34:04.891730 sudo[1397]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Sep 6 00:34:04.924996 systemd[1]: Starting docker.service... Sep 6 00:34:04.975886 env[1407]: time="2025-09-06T00:34:04.975349883Z" level=info msg="Starting up" Sep 6 00:34:04.977854 env[1407]: time="2025-09-06T00:34:04.977814378Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 6 00:34:04.978062 env[1407]: time="2025-09-06T00:34:04.978040675Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 6 00:34:04.978162 env[1407]: time="2025-09-06T00:34:04.978142420Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 6 00:34:04.978235 env[1407]: time="2025-09-06T00:34:04.978220350Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 6 00:34:04.981034 env[1407]: time="2025-09-06T00:34:04.980992168Z" level=info msg="parsed scheme: \"unix\"" module=grpc Sep 6 00:34:04.981034 env[1407]: time="2025-09-06T00:34:04.981015328Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Sep 6 00:34:04.981203 env[1407]: time="2025-09-06T00:34:04.981040669Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Sep 6 00:34:04.981203 env[1407]: time="2025-09-06T00:34:04.981058070Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Sep 6 00:34:05.018588 env[1407]: time="2025-09-06T00:34:05.017934698Z" level=info msg="Loading containers: start." Sep 6 00:34:05.183456 kernel: Initializing XFRM netlink socket Sep 6 00:34:05.227266 env[1407]: time="2025-09-06T00:34:05.227128075Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Sep 6 00:34:05.313022 systemd-networkd[1017]: docker0: Link UP Sep 6 00:34:05.331648 env[1407]: time="2025-09-06T00:34:05.331604551Z" level=info msg="Loading containers: done." Sep 6 00:34:05.348837 env[1407]: time="2025-09-06T00:34:05.348781184Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Sep 6 00:34:05.349094 env[1407]: time="2025-09-06T00:34:05.349046136Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Sep 6 00:34:05.349225 env[1407]: time="2025-09-06T00:34:05.349179059Z" level=info msg="Daemon has completed initialization" Sep 6 00:34:05.369146 systemd[1]: Started docker.service. Sep 6 00:34:05.377879 env[1407]: time="2025-09-06T00:34:05.377814585Z" level=info msg="API listen on /run/docker.sock" Sep 6 00:34:06.280435 env[1222]: time="2025-09-06T00:34:06.280373321Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.12\"" Sep 6 00:34:06.767300 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1202657875.mount: Deactivated successfully. Sep 6 00:34:07.226884 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Sep 6 00:34:07.227200 systemd[1]: Stopped kubelet.service. Sep 6 00:34:07.227271 systemd[1]: kubelet.service: Consumed 1.587s CPU time. Sep 6 00:34:07.229531 systemd[1]: Starting kubelet.service... Sep 6 00:34:07.638501 systemd[1]: Started kubelet.service. Sep 6 00:34:07.728941 kubelet[1534]: E0906 00:34:07.728880 1534 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 6 00:34:07.733595 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 6 00:34:07.733818 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 6 00:34:08.554615 env[1222]: time="2025-09-06T00:34:08.554532628Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:08.557362 env[1222]: time="2025-09-06T00:34:08.557302471Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b1963c5b49c1722b8f408deaf83aafca7f48f47fed0ed14e5c10e93cc55974a7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:08.560043 env[1222]: time="2025-09-06T00:34:08.560002477Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:08.563289 env[1222]: time="2025-09-06T00:34:08.563230740Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:e9011c3bee8c06ecabd7816e119dca4e448c92f7a78acd891de3d2db1dc6c234,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:08.564601 env[1222]: time="2025-09-06T00:34:08.564559849Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.12\" returns image reference \"sha256:b1963c5b49c1722b8f408deaf83aafca7f48f47fed0ed14e5c10e93cc55974a7\"" Sep 6 00:34:08.565768 env[1222]: time="2025-09-06T00:34:08.565734411Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.12\"" Sep 6 00:34:10.111298 env[1222]: time="2025-09-06T00:34:10.111222137Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:10.113953 env[1222]: time="2025-09-06T00:34:10.113906853Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:200c1a99a6f2b9d3b0a6e9b7362663513589341e0e58bc3b953a373efa735dfd,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:10.116168 env[1222]: time="2025-09-06T00:34:10.116124833Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:10.118312 env[1222]: time="2025-09-06T00:34:10.118269446Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:d2862f94d87320267fddbd55db26556a267aa802e51d6b60f25786b4c428afc8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:10.119304 env[1222]: time="2025-09-06T00:34:10.119251829Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.12\" returns image reference \"sha256:200c1a99a6f2b9d3b0a6e9b7362663513589341e0e58bc3b953a373efa735dfd\"" Sep 6 00:34:10.120116 env[1222]: time="2025-09-06T00:34:10.120083617Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.12\"" Sep 6 00:34:11.434772 env[1222]: time="2025-09-06T00:34:11.434695889Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:11.437526 env[1222]: time="2025-09-06T00:34:11.437479796Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:bcdd9599681a9460a5539177a986dbdaf880ac56eeb117ab94adb8f37889efba,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:11.439940 env[1222]: time="2025-09-06T00:34:11.439902052Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:11.442611 env[1222]: time="2025-09-06T00:34:11.442563212Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:152943b7e30244f4415fd0a5860a2dccd91660fe983d30a28a10edb0cc8f6756,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:11.444045 env[1222]: time="2025-09-06T00:34:11.443963624Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.12\" returns image reference \"sha256:bcdd9599681a9460a5539177a986dbdaf880ac56eeb117ab94adb8f37889efba\"" Sep 6 00:34:11.445204 env[1222]: time="2025-09-06T00:34:11.445156751Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.12\"" Sep 6 00:34:12.675768 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3075204504.mount: Deactivated successfully. Sep 6 00:34:13.443682 env[1222]: time="2025-09-06T00:34:13.443589702Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:13.446012 env[1222]: time="2025-09-06T00:34:13.445972678Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:507cc52f5f78c0cff25e904c76c18e6bfc90982e9cc2aa4dcb19033f21c3f679,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:13.449291 env[1222]: time="2025-09-06T00:34:13.449249932Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:13.453616 env[1222]: time="2025-09-06T00:34:13.453559743Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.12\" returns image reference \"sha256:507cc52f5f78c0cff25e904c76c18e6bfc90982e9cc2aa4dcb19033f21c3f679\"" Sep 6 00:34:13.454792 env[1222]: time="2025-09-06T00:34:13.454734664Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:90aa6b5f4065937521ff8438bc705317485d0be3f8b00a07145e697d92cc2cc6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:13.455931 env[1222]: time="2025-09-06T00:34:13.455876501Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Sep 6 00:34:13.828219 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3320725405.mount: Deactivated successfully. Sep 6 00:34:15.107574 env[1222]: time="2025-09-06T00:34:15.107389799Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.110300 env[1222]: time="2025-09-06T00:34:15.110233348Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.117316 env[1222]: time="2025-09-06T00:34:15.117259394Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.119669 env[1222]: time="2025-09-06T00:34:15.119627519Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.120751 env[1222]: time="2025-09-06T00:34:15.120700305Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" Sep 6 00:34:15.121591 env[1222]: time="2025-09-06T00:34:15.121557247Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Sep 6 00:34:15.516468 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3931742343.mount: Deactivated successfully. Sep 6 00:34:15.520483 env[1222]: time="2025-09-06T00:34:15.520368260Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.522707 env[1222]: time="2025-09-06T00:34:15.522663562Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.525040 env[1222]: time="2025-09-06T00:34:15.524988797Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.527261 env[1222]: time="2025-09-06T00:34:15.527216407Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:15.528118 env[1222]: time="2025-09-06T00:34:15.528067047Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" Sep 6 00:34:15.528800 env[1222]: time="2025-09-06T00:34:15.528755266Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" Sep 6 00:34:15.949959 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount570576007.mount: Deactivated successfully. Sep 6 00:34:17.976729 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Sep 6 00:34:17.977063 systemd[1]: Stopped kubelet.service. Sep 6 00:34:17.980091 systemd[1]: Starting kubelet.service... Sep 6 00:34:18.246547 systemd[1]: Started kubelet.service. Sep 6 00:34:18.330228 kubelet[1544]: E0906 00:34:18.330176 1544 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Sep 6 00:34:18.333085 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Sep 6 00:34:18.333308 systemd[1]: kubelet.service: Failed with result 'exit-code'. Sep 6 00:34:18.653844 env[1222]: time="2025-09-06T00:34:18.653684978Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:18.656705 env[1222]: time="2025-09-06T00:34:18.656664208Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:18.659201 env[1222]: time="2025-09-06T00:34:18.659154609Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:18.661654 env[1222]: time="2025-09-06T00:34:18.661616742Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:18.662908 env[1222]: time="2025-09-06T00:34:18.662857541Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" Sep 6 00:34:22.798312 systemd[1]: Stopped kubelet.service. Sep 6 00:34:22.802296 systemd[1]: Starting kubelet.service... Sep 6 00:34:22.840985 systemd[1]: Reloading. Sep 6 00:34:22.989233 /usr/lib/systemd/system-generators/torcx-generator[1594]: time="2025-09-06T00:34:22Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 6 00:34:22.995337 /usr/lib/systemd/system-generators/torcx-generator[1594]: time="2025-09-06T00:34:22Z" level=info msg="torcx already run" Sep 6 00:34:23.131907 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 6 00:34:23.131934 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 6 00:34:23.156819 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 6 00:34:23.307535 systemd[1]: Started kubelet.service. Sep 6 00:34:23.319971 systemd[1]: Stopping kubelet.service... Sep 6 00:34:23.321890 systemd[1]: kubelet.service: Deactivated successfully. Sep 6 00:34:23.322213 systemd[1]: Stopped kubelet.service. Sep 6 00:34:23.326512 systemd[1]: Starting kubelet.service... Sep 6 00:34:23.623263 systemd[1]: Started kubelet.service. Sep 6 00:34:23.687651 kubelet[1650]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 6 00:34:23.687651 kubelet[1650]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 6 00:34:23.687651 kubelet[1650]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 6 00:34:23.689388 kubelet[1650]: I0906 00:34:23.687748 1650 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 6 00:34:23.983828 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Sep 6 00:34:24.086909 kubelet[1650]: I0906 00:34:24.086842 1650 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 6 00:34:24.086909 kubelet[1650]: I0906 00:34:24.086885 1650 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 6 00:34:24.088001 kubelet[1650]: I0906 00:34:24.087961 1650 server.go:934] "Client rotation is on, will bootstrap in background" Sep 6 00:34:24.124054 kubelet[1650]: E0906 00:34:24.124007 1650 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.128.0.94:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:24.125314 kubelet[1650]: I0906 00:34:24.125276 1650 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 6 00:34:24.134034 kubelet[1650]: E0906 00:34:24.133982 1650 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 6 00:34:24.134034 kubelet[1650]: I0906 00:34:24.134022 1650 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 6 00:34:24.139718 kubelet[1650]: I0906 00:34:24.139660 1650 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 6 00:34:24.141129 kubelet[1650]: I0906 00:34:24.141087 1650 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 6 00:34:24.141414 kubelet[1650]: I0906 00:34:24.141347 1650 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 6 00:34:24.141671 kubelet[1650]: I0906 00:34:24.141387 1650 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 6 00:34:24.141860 kubelet[1650]: I0906 00:34:24.141680 1650 topology_manager.go:138] "Creating topology manager with none policy" Sep 6 00:34:24.141860 kubelet[1650]: I0906 00:34:24.141698 1650 container_manager_linux.go:300] "Creating device plugin manager" Sep 6 00:34:24.141860 kubelet[1650]: I0906 00:34:24.141841 1650 state_mem.go:36] "Initialized new in-memory state store" Sep 6 00:34:24.149625 kubelet[1650]: I0906 00:34:24.149592 1650 kubelet.go:408] "Attempting to sync node with API server" Sep 6 00:34:24.149733 kubelet[1650]: I0906 00:34:24.149634 1650 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 6 00:34:24.149733 kubelet[1650]: I0906 00:34:24.149686 1650 kubelet.go:314] "Adding apiserver pod source" Sep 6 00:34:24.149733 kubelet[1650]: I0906 00:34:24.149714 1650 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 6 00:34:24.179018 kubelet[1650]: W0906 00:34:24.178897 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.128.0.94:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6&limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:24.179300 kubelet[1650]: E0906 00:34:24.179264 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.128.0.94:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6&limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:24.179552 kubelet[1650]: I0906 00:34:24.179532 1650 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 6 00:34:24.180391 kubelet[1650]: I0906 00:34:24.180349 1650 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 6 00:34:24.180612 kubelet[1650]: W0906 00:34:24.180556 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.128.0.94:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:24.180699 kubelet[1650]: E0906 00:34:24.180632 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.128.0.94:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:24.180822 kubelet[1650]: W0906 00:34:24.180803 1650 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Sep 6 00:34:24.187720 kubelet[1650]: I0906 00:34:24.187694 1650 server.go:1274] "Started kubelet" Sep 6 00:34:24.188034 kubelet[1650]: I0906 00:34:24.187991 1650 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 6 00:34:24.188507 kubelet[1650]: I0906 00:34:24.188458 1650 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 6 00:34:24.189055 kubelet[1650]: I0906 00:34:24.189032 1650 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 6 00:34:24.190114 kubelet[1650]: I0906 00:34:24.190084 1650 server.go:449] "Adding debug handlers to kubelet server" Sep 6 00:34:24.213372 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Sep 6 00:34:24.214475 kubelet[1650]: I0906 00:34:24.213566 1650 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 6 00:34:24.219149 kubelet[1650]: E0906 00:34:24.217148 1650 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.128.0.94:6443/api/v1/namespaces/default/events\": dial tcp 10.128.0.94:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6.18628a4c200b1b1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,UID:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,},FirstTimestamp:2025-09-06 00:34:24.187661083 +0000 UTC m=+0.556847241,LastTimestamp:2025-09-06 00:34:24.187661083 +0000 UTC m=+0.556847241,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,}" Sep 6 00:34:24.219593 kubelet[1650]: I0906 00:34:24.219560 1650 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 6 00:34:24.221450 kubelet[1650]: I0906 00:34:24.221424 1650 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 6 00:34:24.222300 kubelet[1650]: E0906 00:34:24.222268 1650 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" not found" Sep 6 00:34:24.222878 kubelet[1650]: I0906 00:34:24.222852 1650 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 6 00:34:24.225617 kubelet[1650]: I0906 00:34:24.224353 1650 reconciler.go:26] "Reconciler: start to sync state" Sep 6 00:34:24.226275 kubelet[1650]: W0906 00:34:24.226215 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.128.0.94:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:24.226488 kubelet[1650]: E0906 00:34:24.226456 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.128.0.94:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:24.226724 kubelet[1650]: E0906 00:34:24.226684 1650 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.94:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6?timeout=10s\": dial tcp 10.128.0.94:6443: connect: connection refused" interval="200ms" Sep 6 00:34:24.227166 kubelet[1650]: I0906 00:34:24.227059 1650 factory.go:221] Registration of the systemd container factory successfully Sep 6 00:34:24.227267 kubelet[1650]: I0906 00:34:24.227240 1650 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 6 00:34:24.229536 kubelet[1650]: I0906 00:34:24.229282 1650 factory.go:221] Registration of the containerd container factory successfully Sep 6 00:34:24.254250 kubelet[1650]: I0906 00:34:24.252078 1650 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 6 00:34:24.254472 kubelet[1650]: I0906 00:34:24.253328 1650 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 6 00:34:24.254563 kubelet[1650]: I0906 00:34:24.254480 1650 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 6 00:34:24.254563 kubelet[1650]: I0906 00:34:24.254506 1650 state_mem.go:36] "Initialized new in-memory state store" Sep 6 00:34:24.256618 kubelet[1650]: I0906 00:34:24.256591 1650 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 6 00:34:24.256851 kubelet[1650]: I0906 00:34:24.256831 1650 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 6 00:34:24.256941 kubelet[1650]: I0906 00:34:24.256868 1650 kubelet.go:2321] "Starting kubelet main sync loop" Sep 6 00:34:24.256941 kubelet[1650]: E0906 00:34:24.256929 1650 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 6 00:34:24.257071 kubelet[1650]: I0906 00:34:24.256796 1650 policy_none.go:49] "None policy: Start" Sep 6 00:34:24.258644 kubelet[1650]: I0906 00:34:24.258615 1650 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 6 00:34:24.258747 kubelet[1650]: I0906 00:34:24.258654 1650 state_mem.go:35] "Initializing new in-memory state store" Sep 6 00:34:24.259882 kubelet[1650]: W0906 00:34:24.259840 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.128.0.94:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:24.260002 kubelet[1650]: E0906 00:34:24.259901 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.128.0.94:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:24.267650 systemd[1]: Created slice kubepods.slice. Sep 6 00:34:24.274965 systemd[1]: Created slice kubepods-burstable.slice. Sep 6 00:34:24.279235 systemd[1]: Created slice kubepods-besteffort.slice. Sep 6 00:34:24.286382 kubelet[1650]: I0906 00:34:24.286350 1650 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 6 00:34:24.287799 kubelet[1650]: I0906 00:34:24.287767 1650 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 6 00:34:24.288189 kubelet[1650]: I0906 00:34:24.287930 1650 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 6 00:34:24.288927 kubelet[1650]: I0906 00:34:24.288891 1650 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 6 00:34:24.291629 kubelet[1650]: E0906 00:34:24.291605 1650 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" not found" Sep 6 00:34:24.374148 systemd[1]: Created slice kubepods-burstable-poda4e0d5c204c75029efdf941e9099398e.slice. Sep 6 00:34:24.388091 systemd[1]: Created slice kubepods-burstable-podbd7d616967c1192e8094f04741fba04b.slice. Sep 6 00:34:24.394034 kubelet[1650]: I0906 00:34:24.393043 1650 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.394034 kubelet[1650]: E0906 00:34:24.393464 1650 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.128.0.94:6443/api/v1/nodes\": dial tcp 10.128.0.94:6443: connect: connection refused" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.395543 systemd[1]: Created slice kubepods-burstable-pod40b169362efbea634edbe6f60ead31e0.slice. Sep 6 00:34:24.425792 kubelet[1650]: I0906 00:34:24.425723 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a4e0d5c204c75029efdf941e9099398e-k8s-certs\") pod \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"a4e0d5c204c75029efdf941e9099398e\") " pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.425792 kubelet[1650]: I0906 00:34:24.425782 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-ca-certs\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426013 kubelet[1650]: I0906 00:34:24.425816 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-k8s-certs\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426013 kubelet[1650]: I0906 00:34:24.425845 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-kubeconfig\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426013 kubelet[1650]: I0906 00:34:24.425870 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a4e0d5c204c75029efdf941e9099398e-ca-certs\") pod \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"a4e0d5c204c75029efdf941e9099398e\") " pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426013 kubelet[1650]: I0906 00:34:24.425898 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-flexvolume-dir\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426227 kubelet[1650]: I0906 00:34:24.425927 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426227 kubelet[1650]: I0906 00:34:24.425954 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/40b169362efbea634edbe6f60ead31e0-kubeconfig\") pod \"kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"40b169362efbea634edbe6f60ead31e0\") " pod="kube-system/kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.426227 kubelet[1650]: I0906 00:34:24.425986 1650 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a4e0d5c204c75029efdf941e9099398e-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"a4e0d5c204c75029efdf941e9099398e\") " pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.428302 kubelet[1650]: E0906 00:34:24.428256 1650 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.94:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6?timeout=10s\": dial tcp 10.128.0.94:6443: connect: connection refused" interval="400ms" Sep 6 00:34:24.599158 kubelet[1650]: I0906 00:34:24.599006 1650 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.601145 kubelet[1650]: E0906 00:34:24.601057 1650 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.128.0.94:6443/api/v1/nodes\": dial tcp 10.128.0.94:6443: connect: connection refused" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:24.684157 env[1222]: time="2025-09-06T00:34:24.684082069Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,Uid:a4e0d5c204c75029efdf941e9099398e,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:24.693982 env[1222]: time="2025-09-06T00:34:24.693920888Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,Uid:bd7d616967c1192e8094f04741fba04b,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:24.699598 env[1222]: time="2025-09-06T00:34:24.699552759Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,Uid:40b169362efbea634edbe6f60ead31e0,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:24.829656 kubelet[1650]: E0906 00:34:24.829590 1650 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.94:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6?timeout=10s\": dial tcp 10.128.0.94:6443: connect: connection refused" interval="800ms" Sep 6 00:34:25.005857 kubelet[1650]: I0906 00:34:25.005809 1650 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:25.006237 kubelet[1650]: E0906 00:34:25.006187 1650 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.128.0.94:6443/api/v1/nodes\": dial tcp 10.128.0.94:6443: connect: connection refused" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:25.094316 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3848214960.mount: Deactivated successfully. Sep 6 00:34:25.103855 env[1222]: time="2025-09-06T00:34:25.103792202Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.107588 env[1222]: time="2025-09-06T00:34:25.107541514Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.108961 env[1222]: time="2025-09-06T00:34:25.108908975Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.110038 env[1222]: time="2025-09-06T00:34:25.109999697Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.112103 env[1222]: time="2025-09-06T00:34:25.112056922Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.114098 env[1222]: time="2025-09-06T00:34:25.114057901Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.115919 env[1222]: time="2025-09-06T00:34:25.115873308Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.116852 env[1222]: time="2025-09-06T00:34:25.116799249Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.117826 env[1222]: time="2025-09-06T00:34:25.117788232Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.121825 env[1222]: time="2025-09-06T00:34:25.121781189Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.125817 env[1222]: time="2025-09-06T00:34:25.125780137Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.126786 env[1222]: time="2025-09-06T00:34:25.126740883Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:25.154778 env[1222]: time="2025-09-06T00:34:25.154249906Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:34:25.154778 env[1222]: time="2025-09-06T00:34:25.154753470Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:34:25.160490 env[1222]: time="2025-09-06T00:34:25.155059682Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:34:25.163106 env[1222]: time="2025-09-06T00:34:25.162542851Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/f947d19b70682fd4379be5664ca62da747a7c24a9396a3a6e3c8cb4af09e462c pid=1692 runtime=io.containerd.runc.v2 Sep 6 00:34:25.171213 kubelet[1650]: W0906 00:34:25.170924 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.128.0.94:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6&limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:25.171213 kubelet[1650]: E0906 00:34:25.171029 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.128.0.94:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6&limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:25.194224 env[1222]: time="2025-09-06T00:34:25.194132071Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:34:25.194540 env[1222]: time="2025-09-06T00:34:25.194477637Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:34:25.194755 env[1222]: time="2025-09-06T00:34:25.194693941Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:34:25.195265 env[1222]: time="2025-09-06T00:34:25.195194466Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/bcfe5ad852657d8caa0d78539afd0e568ab5a4dea91a2a0d0c82426eea309fcd pid=1727 runtime=io.containerd.runc.v2 Sep 6 00:34:25.195819 env[1222]: time="2025-09-06T00:34:25.195721838Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:34:25.196135 env[1222]: time="2025-09-06T00:34:25.196072054Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:34:25.196354 env[1222]: time="2025-09-06T00:34:25.196291655Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:34:25.198839 env[1222]: time="2025-09-06T00:34:25.198267345Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b31c4d9e650be41a14fa4e81c7b69461a511fc42900e79f2141d99ba0cb1d22a pid=1716 runtime=io.containerd.runc.v2 Sep 6 00:34:25.202809 systemd[1]: Started cri-containerd-f947d19b70682fd4379be5664ca62da747a7c24a9396a3a6e3c8cb4af09e462c.scope. Sep 6 00:34:25.236226 systemd[1]: Started cri-containerd-b31c4d9e650be41a14fa4e81c7b69461a511fc42900e79f2141d99ba0cb1d22a.scope. Sep 6 00:34:25.254901 systemd[1]: Started cri-containerd-bcfe5ad852657d8caa0d78539afd0e568ab5a4dea91a2a0d0c82426eea309fcd.scope. Sep 6 00:34:25.319685 kubelet[1650]: W0906 00:34:25.319532 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.128.0.94:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:25.319685 kubelet[1650]: E0906 00:34:25.319621 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.128.0.94:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:25.369245 kubelet[1650]: W0906 00:34:25.369146 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.128.0.94:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:25.369472 kubelet[1650]: E0906 00:34:25.369258 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.128.0.94:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:25.376579 env[1222]: time="2025-09-06T00:34:25.376520759Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,Uid:40b169362efbea634edbe6f60ead31e0,Namespace:kube-system,Attempt:0,} returns sandbox id \"f947d19b70682fd4379be5664ca62da747a7c24a9396a3a6e3c8cb4af09e462c\"" Sep 6 00:34:25.377418 env[1222]: time="2025-09-06T00:34:25.377350270Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,Uid:bd7d616967c1192e8094f04741fba04b,Namespace:kube-system,Attempt:0,} returns sandbox id \"b31c4d9e650be41a14fa4e81c7b69461a511fc42900e79f2141d99ba0cb1d22a\"" Sep 6 00:34:25.391672 env[1222]: time="2025-09-06T00:34:25.391624333Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6,Uid:a4e0d5c204c75029efdf941e9099398e,Namespace:kube-system,Attempt:0,} returns sandbox id \"bcfe5ad852657d8caa0d78539afd0e568ab5a4dea91a2a0d0c82426eea309fcd\"" Sep 6 00:34:25.392031 kubelet[1650]: E0906 00:34:25.391977 1650 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" hostnameMaxLen=63 truncatedHostname="kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d4" Sep 6 00:34:25.393447 kubelet[1650]: E0906 00:34:25.393360 1650 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" hostnameMaxLen=63 truncatedHostname="kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2e" Sep 6 00:34:25.395529 env[1222]: time="2025-09-06T00:34:25.395471382Z" level=info msg="CreateContainer within sandbox \"f947d19b70682fd4379be5664ca62da747a7c24a9396a3a6e3c8cb4af09e462c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Sep 6 00:34:25.397307 env[1222]: time="2025-09-06T00:34:25.397263998Z" level=info msg="CreateContainer within sandbox \"b31c4d9e650be41a14fa4e81c7b69461a511fc42900e79f2141d99ba0cb1d22a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Sep 6 00:34:25.397853 kubelet[1650]: E0906 00:34:25.397813 1650 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" hostnameMaxLen=63 truncatedHostname="kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2e" Sep 6 00:34:25.400153 env[1222]: time="2025-09-06T00:34:25.400111980Z" level=info msg="CreateContainer within sandbox \"bcfe5ad852657d8caa0d78539afd0e568ab5a4dea91a2a0d0c82426eea309fcd\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Sep 6 00:34:25.409590 kubelet[1650]: W0906 00:34:25.409378 1650 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.128.0.94:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.94:6443: connect: connection refused Sep 6 00:34:25.409590 kubelet[1650]: E0906 00:34:25.409535 1650 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.128.0.94:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.128.0.94:6443: connect: connection refused" logger="UnhandledError" Sep 6 00:34:25.418808 env[1222]: time="2025-09-06T00:34:25.418755538Z" level=info msg="CreateContainer within sandbox \"f947d19b70682fd4379be5664ca62da747a7c24a9396a3a6e3c8cb4af09e462c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"13979a3a3dc9430ccca21408353c9a8e43a8bf07fb5f2c1f497c429748c418f3\"" Sep 6 00:34:25.419896 env[1222]: time="2025-09-06T00:34:25.419848490Z" level=info msg="StartContainer for \"13979a3a3dc9430ccca21408353c9a8e43a8bf07fb5f2c1f497c429748c418f3\"" Sep 6 00:34:25.425658 env[1222]: time="2025-09-06T00:34:25.425606818Z" level=info msg="CreateContainer within sandbox \"b31c4d9e650be41a14fa4e81c7b69461a511fc42900e79f2141d99ba0cb1d22a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"10119bf1d5221156a596bb7d1d857d3ed50849de5cc8d17ddc1a0bb7d4317cb6\"" Sep 6 00:34:25.426472 env[1222]: time="2025-09-06T00:34:25.426431308Z" level=info msg="StartContainer for \"10119bf1d5221156a596bb7d1d857d3ed50849de5cc8d17ddc1a0bb7d4317cb6\"" Sep 6 00:34:25.432218 env[1222]: time="2025-09-06T00:34:25.432176469Z" level=info msg="CreateContainer within sandbox \"bcfe5ad852657d8caa0d78539afd0e568ab5a4dea91a2a0d0c82426eea309fcd\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"6c4d87678a1380c6fd46525ec2de26fa2ce6eba28434beffd004d1d7cd4f269a\"" Sep 6 00:34:25.433707 env[1222]: time="2025-09-06T00:34:25.433659025Z" level=info msg="StartContainer for \"6c4d87678a1380c6fd46525ec2de26fa2ce6eba28434beffd004d1d7cd4f269a\"" Sep 6 00:34:25.467816 systemd[1]: Started cri-containerd-13979a3a3dc9430ccca21408353c9a8e43a8bf07fb5f2c1f497c429748c418f3.scope. Sep 6 00:34:25.474194 systemd[1]: Started cri-containerd-10119bf1d5221156a596bb7d1d857d3ed50849de5cc8d17ddc1a0bb7d4317cb6.scope. Sep 6 00:34:25.508067 systemd[1]: Started cri-containerd-6c4d87678a1380c6fd46525ec2de26fa2ce6eba28434beffd004d1d7cd4f269a.scope. Sep 6 00:34:25.572641 env[1222]: time="2025-09-06T00:34:25.572510800Z" level=info msg="StartContainer for \"10119bf1d5221156a596bb7d1d857d3ed50849de5cc8d17ddc1a0bb7d4317cb6\" returns successfully" Sep 6 00:34:25.621385 env[1222]: time="2025-09-06T00:34:25.621332523Z" level=info msg="StartContainer for \"6c4d87678a1380c6fd46525ec2de26fa2ce6eba28434beffd004d1d7cd4f269a\" returns successfully" Sep 6 00:34:25.630381 kubelet[1650]: E0906 00:34:25.630284 1650 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.94:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6?timeout=10s\": dial tcp 10.128.0.94:6443: connect: connection refused" interval="1.6s" Sep 6 00:34:25.662548 env[1222]: time="2025-09-06T00:34:25.662485912Z" level=info msg="StartContainer for \"13979a3a3dc9430ccca21408353c9a8e43a8bf07fb5f2c1f497c429748c418f3\" returns successfully" Sep 6 00:34:25.811212 kubelet[1650]: I0906 00:34:25.810731 1650 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:28.794909 kubelet[1650]: E0906 00:34:28.794856 1650 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" not found" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:28.908950 kubelet[1650]: I0906 00:34:28.908887 1650 kubelet_node_status.go:75] "Successfully registered node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:28.908950 kubelet[1650]: E0906 00:34:28.908954 1650 kubelet_node_status.go:535] "Error updating node status, will retry" err="error getting node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\": node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" not found" Sep 6 00:34:29.182137 kubelet[1650]: I0906 00:34:29.182002 1650 apiserver.go:52] "Watching apiserver" Sep 6 00:34:29.224893 kubelet[1650]: I0906 00:34:29.224840 1650 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 6 00:34:31.104965 systemd[1]: Reloading. Sep 6 00:34:31.170718 kubelet[1650]: W0906 00:34:31.170661 1650 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] Sep 6 00:34:31.252078 /usr/lib/systemd/system-generators/torcx-generator[1956]: time="2025-09-06T00:34:31Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.8 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.8 /var/lib/torcx/store]" Sep 6 00:34:31.252737 /usr/lib/systemd/system-generators/torcx-generator[1956]: time="2025-09-06T00:34:31Z" level=info msg="torcx already run" Sep 6 00:34:31.368725 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Sep 6 00:34:31.368764 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Sep 6 00:34:31.401411 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Sep 6 00:34:31.556926 systemd[1]: Stopping kubelet.service... Sep 6 00:34:31.580069 systemd[1]: kubelet.service: Deactivated successfully. Sep 6 00:34:31.580343 systemd[1]: Stopped kubelet.service. Sep 6 00:34:31.580443 systemd[1]: kubelet.service: Consumed 1.050s CPU time. Sep 6 00:34:31.582954 systemd[1]: Starting kubelet.service... Sep 6 00:34:31.835872 systemd[1]: Started kubelet.service. Sep 6 00:34:31.929094 kubelet[2004]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 6 00:34:31.929094 kubelet[2004]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Sep 6 00:34:31.929094 kubelet[2004]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Sep 6 00:34:31.929858 kubelet[2004]: I0906 00:34:31.929195 2004 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Sep 6 00:34:31.939089 kubelet[2004]: I0906 00:34:31.939026 2004 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" Sep 6 00:34:31.939089 kubelet[2004]: I0906 00:34:31.939065 2004 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Sep 6 00:34:31.939576 kubelet[2004]: I0906 00:34:31.939541 2004 server.go:934] "Client rotation is on, will bootstrap in background" Sep 6 00:34:31.941643 kubelet[2004]: I0906 00:34:31.941601 2004 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Sep 6 00:34:31.944306 kubelet[2004]: I0906 00:34:31.944259 2004 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Sep 6 00:34:31.949748 kubelet[2004]: E0906 00:34:31.949691 2004 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Sep 6 00:34:31.949748 kubelet[2004]: I0906 00:34:31.949732 2004 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Sep 6 00:34:31.953586 kubelet[2004]: I0906 00:34:31.953551 2004 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Sep 6 00:34:31.953736 kubelet[2004]: I0906 00:34:31.953717 2004 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Sep 6 00:34:31.953957 kubelet[2004]: I0906 00:34:31.953905 2004 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Sep 6 00:34:31.954198 kubelet[2004]: I0906 00:34:31.953944 2004 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Sep 6 00:34:31.954377 kubelet[2004]: I0906 00:34:31.954204 2004 topology_manager.go:138] "Creating topology manager with none policy" Sep 6 00:34:31.954377 kubelet[2004]: I0906 00:34:31.954225 2004 container_manager_linux.go:300] "Creating device plugin manager" Sep 6 00:34:31.954377 kubelet[2004]: I0906 00:34:31.954267 2004 state_mem.go:36] "Initialized new in-memory state store" Sep 6 00:34:31.954587 kubelet[2004]: I0906 00:34:31.954461 2004 kubelet.go:408] "Attempting to sync node with API server" Sep 6 00:34:31.954587 kubelet[2004]: I0906 00:34:31.954499 2004 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" Sep 6 00:34:31.954587 kubelet[2004]: I0906 00:34:31.954545 2004 kubelet.go:314] "Adding apiserver pod source" Sep 6 00:34:31.954587 kubelet[2004]: I0906 00:34:31.954562 2004 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Sep 6 00:34:31.972464 kubelet[2004]: I0906 00:34:31.971372 2004 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Sep 6 00:34:31.972464 kubelet[2004]: I0906 00:34:31.972133 2004 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Sep 6 00:34:31.972833 kubelet[2004]: I0906 00:34:31.972795 2004 server.go:1274] "Started kubelet" Sep 6 00:34:31.979912 kubelet[2004]: I0906 00:34:31.976610 2004 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Sep 6 00:34:31.982177 kubelet[2004]: I0906 00:34:31.982154 2004 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Sep 6 00:34:31.982552 kubelet[2004]: I0906 00:34:31.982519 2004 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Sep 6 00:34:31.985839 kubelet[2004]: I0906 00:34:31.984089 2004 server.go:449] "Adding debug handlers to kubelet server" Sep 6 00:34:31.987633 kubelet[2004]: I0906 00:34:31.987584 2004 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Sep 6 00:34:31.988018 kubelet[2004]: I0906 00:34:31.987993 2004 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Sep 6 00:34:31.992380 kubelet[2004]: I0906 00:34:31.992357 2004 volume_manager.go:289] "Starting Kubelet Volume Manager" Sep 6 00:34:31.994243 kubelet[2004]: I0906 00:34:31.994223 2004 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Sep 6 00:34:31.994779 kubelet[2004]: I0906 00:34:31.994759 2004 reconciler.go:26] "Reconciler: start to sync state" Sep 6 00:34:32.001220 kubelet[2004]: I0906 00:34:32.000563 2004 factory.go:221] Registration of the systemd container factory successfully Sep 6 00:34:32.001220 kubelet[2004]: I0906 00:34:32.000693 2004 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Sep 6 00:34:32.002603 kubelet[2004]: E0906 00:34:32.002566 2004 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Sep 6 00:34:32.004677 kubelet[2004]: I0906 00:34:32.003436 2004 factory.go:221] Registration of the containerd container factory successfully Sep 6 00:34:32.014792 kubelet[2004]: I0906 00:34:32.014756 2004 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Sep 6 00:34:32.019619 kubelet[2004]: I0906 00:34:32.019593 2004 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Sep 6 00:34:32.019743 kubelet[2004]: I0906 00:34:32.019625 2004 status_manager.go:217] "Starting to sync pod status with apiserver" Sep 6 00:34:32.019743 kubelet[2004]: I0906 00:34:32.019649 2004 kubelet.go:2321] "Starting kubelet main sync loop" Sep 6 00:34:32.019743 kubelet[2004]: E0906 00:34:32.019710 2004 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Sep 6 00:34:32.064052 kubelet[2004]: I0906 00:34:32.063999 2004 cpu_manager.go:214] "Starting CPU manager" policy="none" Sep 6 00:34:32.064052 kubelet[2004]: I0906 00:34:32.064028 2004 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Sep 6 00:34:32.064052 kubelet[2004]: I0906 00:34:32.064057 2004 state_mem.go:36] "Initialized new in-memory state store" Sep 6 00:34:32.064359 kubelet[2004]: I0906 00:34:32.064297 2004 state_mem.go:88] "Updated default CPUSet" cpuSet="" Sep 6 00:34:32.064359 kubelet[2004]: I0906 00:34:32.064317 2004 state_mem.go:96] "Updated CPUSet assignments" assignments={} Sep 6 00:34:32.064359 kubelet[2004]: I0906 00:34:32.064347 2004 policy_none.go:49] "None policy: Start" Sep 6 00:34:32.065230 kubelet[2004]: I0906 00:34:32.065194 2004 memory_manager.go:170] "Starting memorymanager" policy="None" Sep 6 00:34:32.065230 kubelet[2004]: I0906 00:34:32.065235 2004 state_mem.go:35] "Initializing new in-memory state store" Sep 6 00:34:32.065585 kubelet[2004]: I0906 00:34:32.065562 2004 state_mem.go:75] "Updated machine memory state" Sep 6 00:34:32.075314 kubelet[2004]: I0906 00:34:32.075248 2004 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Sep 6 00:34:32.076110 kubelet[2004]: I0906 00:34:32.076087 2004 eviction_manager.go:189] "Eviction manager: starting control loop" Sep 6 00:34:32.082794 kubelet[2004]: I0906 00:34:32.076471 2004 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Sep 6 00:34:32.082794 kubelet[2004]: I0906 00:34:32.077221 2004 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Sep 6 00:34:32.111737 sudo[2037]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Sep 6 00:34:32.112175 sudo[2037]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Sep 6 00:34:32.128580 kubelet[2004]: W0906 00:34:32.128551 2004 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] Sep 6 00:34:32.129118 kubelet[2004]: W0906 00:34:32.129099 2004 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] Sep 6 00:34:32.130905 kubelet[2004]: W0906 00:34:32.130879 2004 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] Sep 6 00:34:32.131113 kubelet[2004]: E0906 00:34:32.131084 2004 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" already exists" pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.196830 kubelet[2004]: I0906 00:34:32.196797 2004 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.197845 kubelet[2004]: I0906 00:34:32.197797 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/a4e0d5c204c75029efdf941e9099398e-ca-certs\") pod \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"a4e0d5c204c75029efdf941e9099398e\") " pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.198130 kubelet[2004]: I0906 00:34:32.198049 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-k8s-certs\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.198330 kubelet[2004]: I0906 00:34:32.198306 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/40b169362efbea634edbe6f60ead31e0-kubeconfig\") pod \"kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"40b169362efbea634edbe6f60ead31e0\") " pod="kube-system/kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.200571 kubelet[2004]: I0906 00:34:32.200529 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/a4e0d5c204c75029efdf941e9099398e-k8s-certs\") pod \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"a4e0d5c204c75029efdf941e9099398e\") " pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.200777 kubelet[2004]: I0906 00:34:32.200750 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/a4e0d5c204c75029efdf941e9099398e-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"a4e0d5c204c75029efdf941e9099398e\") " pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.201432 kubelet[2004]: I0906 00:34:32.200950 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-ca-certs\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.201432 kubelet[2004]: I0906 00:34:32.201023 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-flexvolume-dir\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.201432 kubelet[2004]: I0906 00:34:32.201083 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-kubeconfig\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.201432 kubelet[2004]: I0906 00:34:32.201120 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/bd7d616967c1192e8094f04741fba04b-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" (UID: \"bd7d616967c1192e8094f04741fba04b\") " pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.211886 kubelet[2004]: I0906 00:34:32.211858 2004 kubelet_node_status.go:111] "Node was previously registered" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.212128 kubelet[2004]: I0906 00:34:32.212109 2004 kubelet_node_status.go:75] "Successfully registered node" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" Sep 6 00:34:32.837661 sudo[2037]: pam_unix(sudo:session): session closed for user root Sep 6 00:34:32.956898 kubelet[2004]: I0906 00:34:32.956849 2004 apiserver.go:52] "Watching apiserver" Sep 6 00:34:32.994553 kubelet[2004]: I0906 00:34:32.994516 2004 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Sep 6 00:34:33.074588 kubelet[2004]: I0906 00:34:33.074519 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" podStartSLOduration=1.074492138 podStartE2EDuration="1.074492138s" podCreationTimestamp="2025-09-06 00:34:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:34:33.072517261 +0000 UTC m=+1.226202828" watchObservedRunningTime="2025-09-06 00:34:33.074492138 +0000 UTC m=+1.228177708" Sep 6 00:34:33.087687 kubelet[2004]: I0906 00:34:33.087611 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" podStartSLOduration=1.087585612 podStartE2EDuration="1.087585612s" podCreationTimestamp="2025-09-06 00:34:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:34:33.085915034 +0000 UTC m=+1.239600614" watchObservedRunningTime="2025-09-06 00:34:33.087585612 +0000 UTC m=+1.241271183" Sep 6 00:34:33.135391 kubelet[2004]: I0906 00:34:33.135218 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" podStartSLOduration=2.1351963 podStartE2EDuration="2.1351963s" podCreationTimestamp="2025-09-06 00:34:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:34:33.102353586 +0000 UTC m=+1.256039167" watchObservedRunningTime="2025-09-06 00:34:33.1351963 +0000 UTC m=+1.288881880" Sep 6 00:34:35.178748 sudo[1397]: pam_unix(sudo:session): session closed for user root Sep 6 00:34:35.222861 sshd[1390]: pam_unix(sshd:session): session closed for user core Sep 6 00:34:35.227819 systemd[1]: sshd@4-10.128.0.94:22-139.178.89.65:51176.service: Deactivated successfully. Sep 6 00:34:35.229111 systemd[1]: session-5.scope: Deactivated successfully. Sep 6 00:34:35.229361 systemd[1]: session-5.scope: Consumed 7.156s CPU time. Sep 6 00:34:35.230199 systemd-logind[1211]: Session 5 logged out. Waiting for processes to exit. Sep 6 00:34:35.231992 systemd-logind[1211]: Removed session 5. Sep 6 00:34:38.207192 kubelet[2004]: I0906 00:34:38.207135 2004 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Sep 6 00:34:38.207887 env[1222]: time="2025-09-06T00:34:38.207713773Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Sep 6 00:34:38.208312 kubelet[2004]: I0906 00:34:38.208030 2004 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Sep 6 00:34:38.770435 update_engine[1212]: I0906 00:34:38.769840 1212 update_attempter.cc:509] Updating boot flags... Sep 6 00:34:38.779033 systemd[1]: Created slice kubepods-besteffort-pod63ed56b9_c1bf_413a_b06b_6fa01132c387.slice. Sep 6 00:34:38.838098 systemd[1]: Created slice kubepods-burstable-pod3532905a_8f04_4418_b4ad_6a6a50e8f32b.slice. Sep 6 00:34:38.848183 kubelet[2004]: I0906 00:34:38.848125 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cni-path\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.848558 kubelet[2004]: I0906 00:34:38.848527 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-config-path\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.848781 kubelet[2004]: I0906 00:34:38.848751 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jd8z\" (UniqueName: \"kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-kube-api-access-9jd8z\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.848993 kubelet[2004]: I0906 00:34:38.848968 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/63ed56b9-c1bf-413a-b06b-6fa01132c387-lib-modules\") pod \"kube-proxy-6rglw\" (UID: \"63ed56b9-c1bf-413a-b06b-6fa01132c387\") " pod="kube-system/kube-proxy-6rglw" Sep 6 00:34:38.849200 kubelet[2004]: I0906 00:34:38.849158 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-run\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.851058 kubelet[2004]: I0906 00:34:38.851014 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hostproc\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.852649 kubelet[2004]: I0906 00:34:38.852620 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-net\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.852827 kubelet[2004]: I0906 00:34:38.852804 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-lib-modules\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.853057 kubelet[2004]: I0906 00:34:38.853032 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hubble-tls\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.853229 kubelet[2004]: I0906 00:34:38.853205 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/63ed56b9-c1bf-413a-b06b-6fa01132c387-xtables-lock\") pod \"kube-proxy-6rglw\" (UID: \"63ed56b9-c1bf-413a-b06b-6fa01132c387\") " pod="kube-system/kube-proxy-6rglw" Sep 6 00:34:38.853376 kubelet[2004]: I0906 00:34:38.853351 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv4dq\" (UniqueName: \"kubernetes.io/projected/63ed56b9-c1bf-413a-b06b-6fa01132c387-kube-api-access-gv4dq\") pod \"kube-proxy-6rglw\" (UID: \"63ed56b9-c1bf-413a-b06b-6fa01132c387\") " pod="kube-system/kube-proxy-6rglw" Sep 6 00:34:38.853563 kubelet[2004]: I0906 00:34:38.853534 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-etc-cni-netd\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.853695 kubelet[2004]: I0906 00:34:38.853673 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-xtables-lock\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.853833 kubelet[2004]: I0906 00:34:38.853811 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3532905a-8f04-4418-b4ad-6a6a50e8f32b-clustermesh-secrets\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.853960 kubelet[2004]: I0906 00:34:38.853939 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-kernel\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.854102 kubelet[2004]: I0906 00:34:38.854079 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/63ed56b9-c1bf-413a-b06b-6fa01132c387-kube-proxy\") pod \"kube-proxy-6rglw\" (UID: \"63ed56b9-c1bf-413a-b06b-6fa01132c387\") " pod="kube-system/kube-proxy-6rglw" Sep 6 00:34:38.854255 kubelet[2004]: I0906 00:34:38.854232 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-bpf-maps\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.856445 kubelet[2004]: I0906 00:34:38.854392 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-cgroup\") pod \"cilium-jwtq9\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " pod="kube-system/cilium-jwtq9" Sep 6 00:34:38.961443 kubelet[2004]: I0906 00:34:38.958322 2004 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Sep 6 00:34:39.036478 kubelet[2004]: E0906 00:34:39.035668 2004 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 6 00:34:39.036478 kubelet[2004]: E0906 00:34:39.035966 2004 projected.go:194] Error preparing data for projected volume kube-api-access-9jd8z for pod kube-system/cilium-jwtq9: configmap "kube-root-ca.crt" not found Sep 6 00:34:39.036478 kubelet[2004]: E0906 00:34:39.036053 2004 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-kube-api-access-9jd8z podName:3532905a-8f04-4418-b4ad-6a6a50e8f32b nodeName:}" failed. No retries permitted until 2025-09-06 00:34:39.536025191 +0000 UTC m=+7.689710761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9jd8z" (UniqueName: "kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-kube-api-access-9jd8z") pod "cilium-jwtq9" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b") : configmap "kube-root-ca.crt" not found Sep 6 00:34:39.036478 kubelet[2004]: E0906 00:34:39.036361 2004 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Sep 6 00:34:39.036478 kubelet[2004]: E0906 00:34:39.036385 2004 projected.go:194] Error preparing data for projected volume kube-api-access-gv4dq for pod kube-system/kube-proxy-6rglw: configmap "kube-root-ca.crt" not found Sep 6 00:34:39.036478 kubelet[2004]: E0906 00:34:39.036477 2004 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/63ed56b9-c1bf-413a-b06b-6fa01132c387-kube-api-access-gv4dq podName:63ed56b9-c1bf-413a-b06b-6fa01132c387 nodeName:}" failed. No retries permitted until 2025-09-06 00:34:39.536447241 +0000 UTC m=+7.690132808 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gv4dq" (UniqueName: "kubernetes.io/projected/63ed56b9-c1bf-413a-b06b-6fa01132c387-kube-api-access-gv4dq") pod "kube-proxy-6rglw" (UID: "63ed56b9-c1bf-413a-b06b-6fa01132c387") : configmap "kube-root-ca.crt" not found Sep 6 00:34:39.305504 systemd[1]: Created slice kubepods-besteffort-pod448ef295_16b6_4f6a_9a49_a911a225b515.slice. Sep 6 00:34:39.362668 kubelet[2004]: I0906 00:34:39.362606 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jtr9\" (UniqueName: \"kubernetes.io/projected/448ef295-16b6-4f6a-9a49-a911a225b515-kube-api-access-5jtr9\") pod \"cilium-operator-5d85765b45-bhrj4\" (UID: \"448ef295-16b6-4f6a-9a49-a911a225b515\") " pod="kube-system/cilium-operator-5d85765b45-bhrj4" Sep 6 00:34:39.363214 kubelet[2004]: I0906 00:34:39.362760 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/448ef295-16b6-4f6a-9a49-a911a225b515-cilium-config-path\") pod \"cilium-operator-5d85765b45-bhrj4\" (UID: \"448ef295-16b6-4f6a-9a49-a911a225b515\") " pod="kube-system/cilium-operator-5d85765b45-bhrj4" Sep 6 00:34:39.612929 env[1222]: time="2025-09-06T00:34:39.612756503Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-bhrj4,Uid:448ef295-16b6-4f6a-9a49-a911a225b515,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:39.638260 env[1222]: time="2025-09-06T00:34:39.638160529Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:34:39.638585 env[1222]: time="2025-09-06T00:34:39.638545822Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:34:39.638848 env[1222]: time="2025-09-06T00:34:39.638798887Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:34:39.639423 env[1222]: time="2025-09-06T00:34:39.639343159Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee pid=2103 runtime=io.containerd.runc.v2 Sep 6 00:34:39.658194 systemd[1]: Started cri-containerd-6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee.scope. Sep 6 00:34:39.723287 env[1222]: time="2025-09-06T00:34:39.721745023Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-bhrj4,Uid:448ef295-16b6-4f6a-9a49-a911a225b515,Namespace:kube-system,Attempt:0,} returns sandbox id \"6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee\"" Sep 6 00:34:39.726508 env[1222]: time="2025-09-06T00:34:39.725119181Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Sep 6 00:34:39.726508 env[1222]: time="2025-09-06T00:34:39.725912732Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6rglw,Uid:63ed56b9-c1bf-413a-b06b-6fa01132c387,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:39.745271 env[1222]: time="2025-09-06T00:34:39.745215158Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jwtq9,Uid:3532905a-8f04-4418-b4ad-6a6a50e8f32b,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:39.752267 env[1222]: time="2025-09-06T00:34:39.752172128Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:34:39.752867 env[1222]: time="2025-09-06T00:34:39.752230043Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:34:39.752867 env[1222]: time="2025-09-06T00:34:39.752248464Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:34:39.752867 env[1222]: time="2025-09-06T00:34:39.752481385Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/602645763dac86538752d704fa999201c9588da25310a46dfaf22bb597437be7 pid=2146 runtime=io.containerd.runc.v2 Sep 6 00:34:39.771849 env[1222]: time="2025-09-06T00:34:39.771569005Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:34:39.771849 env[1222]: time="2025-09-06T00:34:39.771641245Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:34:39.771849 env[1222]: time="2025-09-06T00:34:39.771662845Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:34:39.773287 env[1222]: time="2025-09-06T00:34:39.773172070Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379 pid=2169 runtime=io.containerd.runc.v2 Sep 6 00:34:39.781330 systemd[1]: Started cri-containerd-602645763dac86538752d704fa999201c9588da25310a46dfaf22bb597437be7.scope. Sep 6 00:34:39.815612 systemd[1]: Started cri-containerd-22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379.scope. Sep 6 00:34:39.842987 env[1222]: time="2025-09-06T00:34:39.842917005Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-6rglw,Uid:63ed56b9-c1bf-413a-b06b-6fa01132c387,Namespace:kube-system,Attempt:0,} returns sandbox id \"602645763dac86538752d704fa999201c9588da25310a46dfaf22bb597437be7\"" Sep 6 00:34:39.850012 env[1222]: time="2025-09-06T00:34:39.849955337Z" level=info msg="CreateContainer within sandbox \"602645763dac86538752d704fa999201c9588da25310a46dfaf22bb597437be7\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Sep 6 00:34:39.874064 env[1222]: time="2025-09-06T00:34:39.873962000Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-jwtq9,Uid:3532905a-8f04-4418-b4ad-6a6a50e8f32b,Namespace:kube-system,Attempt:0,} returns sandbox id \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\"" Sep 6 00:34:39.886302 env[1222]: time="2025-09-06T00:34:39.886261565Z" level=info msg="CreateContainer within sandbox \"602645763dac86538752d704fa999201c9588da25310a46dfaf22bb597437be7\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d42642fc983af38c4d352af4b84e26fa49724a040587f5ee0f26c6e0c21d0f64\"" Sep 6 00:34:39.888597 env[1222]: time="2025-09-06T00:34:39.888566509Z" level=info msg="StartContainer for \"d42642fc983af38c4d352af4b84e26fa49724a040587f5ee0f26c6e0c21d0f64\"" Sep 6 00:34:39.914864 systemd[1]: Started cri-containerd-d42642fc983af38c4d352af4b84e26fa49724a040587f5ee0f26c6e0c21d0f64.scope. Sep 6 00:34:39.969989 env[1222]: time="2025-09-06T00:34:39.969928378Z" level=info msg="StartContainer for \"d42642fc983af38c4d352af4b84e26fa49724a040587f5ee0f26c6e0c21d0f64\" returns successfully" Sep 6 00:34:40.516025 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount509768524.mount: Deactivated successfully. Sep 6 00:34:41.680859 env[1222]: time="2025-09-06T00:34:41.680782684Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:41.683316 env[1222]: time="2025-09-06T00:34:41.683271149Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:41.685247 env[1222]: time="2025-09-06T00:34:41.685206329Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:41.685966 env[1222]: time="2025-09-06T00:34:41.685922455Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Sep 6 00:34:41.689804 env[1222]: time="2025-09-06T00:34:41.688721327Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Sep 6 00:34:41.690438 env[1222]: time="2025-09-06T00:34:41.690365227Z" level=info msg="CreateContainer within sandbox \"6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Sep 6 00:34:41.709741 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1145632721.mount: Deactivated successfully. Sep 6 00:34:41.718222 env[1222]: time="2025-09-06T00:34:41.718159204Z" level=info msg="CreateContainer within sandbox \"6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\"" Sep 6 00:34:41.720181 env[1222]: time="2025-09-06T00:34:41.720139036Z" level=info msg="StartContainer for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\"" Sep 6 00:34:41.762167 systemd[1]: Started cri-containerd-7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89.scope. Sep 6 00:34:41.817163 env[1222]: time="2025-09-06T00:34:41.814821112Z" level=info msg="StartContainer for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" returns successfully" Sep 6 00:34:42.095617 kubelet[2004]: I0906 00:34:42.095541 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-6rglw" podStartSLOduration=4.095517056 podStartE2EDuration="4.095517056s" podCreationTimestamp="2025-09-06 00:34:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:34:40.087005142 +0000 UTC m=+8.240690722" watchObservedRunningTime="2025-09-06 00:34:42.095517056 +0000 UTC m=+10.249202889" Sep 6 00:34:42.703218 systemd[1]: run-containerd-runc-k8s.io-7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89-runc.mHORSI.mount: Deactivated successfully. Sep 6 00:34:48.272450 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1399831883.mount: Deactivated successfully. Sep 6 00:34:51.662810 env[1222]: time="2025-09-06T00:34:51.662720886Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:51.665488 env[1222]: time="2025-09-06T00:34:51.665447004Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:51.667924 env[1222]: time="2025-09-06T00:34:51.667888468Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Sep 6 00:34:51.669113 env[1222]: time="2025-09-06T00:34:51.669066778Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Sep 6 00:34:51.674522 env[1222]: time="2025-09-06T00:34:51.674486460Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 6 00:34:51.693647 env[1222]: time="2025-09-06T00:34:51.693598714Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\"" Sep 6 00:34:51.694134 env[1222]: time="2025-09-06T00:34:51.694096818Z" level=info msg="StartContainer for \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\"" Sep 6 00:34:51.737557 systemd[1]: run-containerd-runc-k8s.io-351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279-runc.twejIq.mount: Deactivated successfully. Sep 6 00:34:51.742019 systemd[1]: Started cri-containerd-351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279.scope. Sep 6 00:34:51.786128 env[1222]: time="2025-09-06T00:34:51.782730952Z" level=info msg="StartContainer for \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\" returns successfully" Sep 6 00:34:51.793181 systemd[1]: cri-containerd-351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279.scope: Deactivated successfully. Sep 6 00:34:52.130581 kubelet[2004]: I0906 00:34:52.130496 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-bhrj4" podStartSLOduration=11.167100819 podStartE2EDuration="13.130471298s" podCreationTimestamp="2025-09-06 00:34:39 +0000 UTC" firstStartedPulling="2025-09-06 00:34:39.724061627 +0000 UTC m=+7.877747197" lastFinishedPulling="2025-09-06 00:34:41.687432103 +0000 UTC m=+9.841117676" observedRunningTime="2025-09-06 00:34:42.126751956 +0000 UTC m=+10.280437536" watchObservedRunningTime="2025-09-06 00:34:52.130471298 +0000 UTC m=+20.284156877" Sep 6 00:34:52.686112 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279-rootfs.mount: Deactivated successfully. Sep 6 00:34:53.876031 env[1222]: time="2025-09-06T00:34:53.875949113Z" level=info msg="shim disconnected" id=351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279 Sep 6 00:34:53.876031 env[1222]: time="2025-09-06T00:34:53.876032842Z" level=warning msg="cleaning up after shim disconnected" id=351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279 namespace=k8s.io Sep 6 00:34:53.876031 env[1222]: time="2025-09-06T00:34:53.876050894Z" level=info msg="cleaning up dead shim" Sep 6 00:34:53.888618 env[1222]: time="2025-09-06T00:34:53.888538819Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:34:53Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2474 runtime=io.containerd.runc.v2\n" Sep 6 00:34:54.115099 env[1222]: time="2025-09-06T00:34:54.115041313Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 6 00:34:54.138376 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount12400721.mount: Deactivated successfully. Sep 6 00:34:54.151605 env[1222]: time="2025-09-06T00:34:54.151532428Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\"" Sep 6 00:34:54.153752 env[1222]: time="2025-09-06T00:34:54.153713193Z" level=info msg="StartContainer for \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\"" Sep 6 00:34:54.200013 systemd[1]: Started cri-containerd-85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c.scope. Sep 6 00:34:54.236720 env[1222]: time="2025-09-06T00:34:54.236654082Z" level=info msg="StartContainer for \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\" returns successfully" Sep 6 00:34:54.253388 systemd[1]: systemd-sysctl.service: Deactivated successfully. Sep 6 00:34:54.254160 systemd[1]: Stopped systemd-sysctl.service. Sep 6 00:34:54.254752 systemd[1]: Stopping systemd-sysctl.service... Sep 6 00:34:54.257341 systemd[1]: Starting systemd-sysctl.service... Sep 6 00:34:54.265572 systemd[1]: cri-containerd-85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c.scope: Deactivated successfully. Sep 6 00:34:54.274792 systemd[1]: Finished systemd-sysctl.service. Sep 6 00:34:54.296037 env[1222]: time="2025-09-06T00:34:54.295956805Z" level=info msg="shim disconnected" id=85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c Sep 6 00:34:54.296037 env[1222]: time="2025-09-06T00:34:54.296014361Z" level=warning msg="cleaning up after shim disconnected" id=85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c namespace=k8s.io Sep 6 00:34:54.296037 env[1222]: time="2025-09-06T00:34:54.296032455Z" level=info msg="cleaning up dead shim" Sep 6 00:34:54.307794 env[1222]: time="2025-09-06T00:34:54.307740627Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:34:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2537 runtime=io.containerd.runc.v2\n" Sep 6 00:34:55.119630 env[1222]: time="2025-09-06T00:34:55.119570020Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 6 00:34:55.130057 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c-rootfs.mount: Deactivated successfully. Sep 6 00:34:55.158347 env[1222]: time="2025-09-06T00:34:55.158277029Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\"" Sep 6 00:34:55.161278 env[1222]: time="2025-09-06T00:34:55.161236959Z" level=info msg="StartContainer for \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\"" Sep 6 00:34:55.229012 systemd[1]: Started cri-containerd-7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8.scope. Sep 6 00:34:55.275149 env[1222]: time="2025-09-06T00:34:55.275092333Z" level=info msg="StartContainer for \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\" returns successfully" Sep 6 00:34:55.279584 systemd[1]: cri-containerd-7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8.scope: Deactivated successfully. Sep 6 00:34:55.309961 env[1222]: time="2025-09-06T00:34:55.309897927Z" level=info msg="shim disconnected" id=7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8 Sep 6 00:34:55.310351 env[1222]: time="2025-09-06T00:34:55.310317032Z" level=warning msg="cleaning up after shim disconnected" id=7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8 namespace=k8s.io Sep 6 00:34:55.310573 env[1222]: time="2025-09-06T00:34:55.310504222Z" level=info msg="cleaning up dead shim" Sep 6 00:34:55.321720 env[1222]: time="2025-09-06T00:34:55.321666485Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:34:55Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2595 runtime=io.containerd.runc.v2\n" Sep 6 00:34:56.125767 env[1222]: time="2025-09-06T00:34:56.125661981Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 6 00:34:56.130930 systemd[1]: run-containerd-runc-k8s.io-7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8-runc.RIH9Vs.mount: Deactivated successfully. Sep 6 00:34:56.131115 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8-rootfs.mount: Deactivated successfully. Sep 6 00:34:56.160436 env[1222]: time="2025-09-06T00:34:56.156911741Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\"" Sep 6 00:34:56.163775 env[1222]: time="2025-09-06T00:34:56.163077940Z" level=info msg="StartContainer for \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\"" Sep 6 00:34:56.196735 systemd[1]: Started cri-containerd-04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716.scope. Sep 6 00:34:56.235012 systemd[1]: cri-containerd-04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716.scope: Deactivated successfully. Sep 6 00:34:56.236572 env[1222]: time="2025-09-06T00:34:56.236252425Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3532905a_8f04_4418_b4ad_6a6a50e8f32b.slice/cri-containerd-04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716.scope/memory.events\": no such file or directory" Sep 6 00:34:56.239535 env[1222]: time="2025-09-06T00:34:56.239490679Z" level=info msg="StartContainer for \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\" returns successfully" Sep 6 00:34:56.270075 env[1222]: time="2025-09-06T00:34:56.270010578Z" level=info msg="shim disconnected" id=04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716 Sep 6 00:34:56.270449 env[1222]: time="2025-09-06T00:34:56.270082795Z" level=warning msg="cleaning up after shim disconnected" id=04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716 namespace=k8s.io Sep 6 00:34:56.270449 env[1222]: time="2025-09-06T00:34:56.270099430Z" level=info msg="cleaning up dead shim" Sep 6 00:34:56.281923 env[1222]: time="2025-09-06T00:34:56.281873799Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:34:56Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2651 runtime=io.containerd.runc.v2\n" Sep 6 00:34:57.130456 env[1222]: time="2025-09-06T00:34:57.129447179Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 6 00:34:57.130870 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716-rootfs.mount: Deactivated successfully. Sep 6 00:34:57.158374 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount685229584.mount: Deactivated successfully. Sep 6 00:34:57.171990 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2956652388.mount: Deactivated successfully. Sep 6 00:34:57.174198 env[1222]: time="2025-09-06T00:34:57.174141055Z" level=info msg="CreateContainer within sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\"" Sep 6 00:34:57.179757 env[1222]: time="2025-09-06T00:34:57.179700085Z" level=info msg="StartContainer for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\"" Sep 6 00:34:57.210959 systemd[1]: Started cri-containerd-44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3.scope. Sep 6 00:34:57.258439 env[1222]: time="2025-09-06T00:34:57.256822036Z" level=info msg="StartContainer for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" returns successfully" Sep 6 00:34:57.483367 kubelet[2004]: I0906 00:34:57.483230 2004 kubelet_node_status.go:488] "Fast updating node status as it just became ready" Sep 6 00:34:57.538844 systemd[1]: Created slice kubepods-burstable-pod5999b4c9_2550_4912_9432_ee4d2d455edf.slice. Sep 6 00:34:57.550412 systemd[1]: Created slice kubepods-burstable-pod67f5d924_e914_4357_aecf_b9d0c736dfcb.slice. Sep 6 00:34:57.604112 kubelet[2004]: I0906 00:34:57.604054 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrpb6\" (UniqueName: \"kubernetes.io/projected/5999b4c9-2550-4912-9432-ee4d2d455edf-kube-api-access-mrpb6\") pod \"coredns-7c65d6cfc9-cn6v6\" (UID: \"5999b4c9-2550-4912-9432-ee4d2d455edf\") " pod="kube-system/coredns-7c65d6cfc9-cn6v6" Sep 6 00:34:57.604330 kubelet[2004]: I0906 00:34:57.604159 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4xxj\" (UniqueName: \"kubernetes.io/projected/67f5d924-e914-4357-aecf-b9d0c736dfcb-kube-api-access-g4xxj\") pod \"coredns-7c65d6cfc9-kmv9n\" (UID: \"67f5d924-e914-4357-aecf-b9d0c736dfcb\") " pod="kube-system/coredns-7c65d6cfc9-kmv9n" Sep 6 00:34:57.604330 kubelet[2004]: I0906 00:34:57.604205 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67f5d924-e914-4357-aecf-b9d0c736dfcb-config-volume\") pod \"coredns-7c65d6cfc9-kmv9n\" (UID: \"67f5d924-e914-4357-aecf-b9d0c736dfcb\") " pod="kube-system/coredns-7c65d6cfc9-kmv9n" Sep 6 00:34:57.604330 kubelet[2004]: I0906 00:34:57.604234 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5999b4c9-2550-4912-9432-ee4d2d455edf-config-volume\") pod \"coredns-7c65d6cfc9-cn6v6\" (UID: \"5999b4c9-2550-4912-9432-ee4d2d455edf\") " pod="kube-system/coredns-7c65d6cfc9-cn6v6" Sep 6 00:34:57.852455 env[1222]: time="2025-09-06T00:34:57.852303542Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-cn6v6,Uid:5999b4c9-2550-4912-9432-ee4d2d455edf,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:57.856935 env[1222]: time="2025-09-06T00:34:57.856847674Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-kmv9n,Uid:67f5d924-e914-4357-aecf-b9d0c736dfcb,Namespace:kube-system,Attempt:0,}" Sep 6 00:34:59.601440 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Sep 6 00:34:59.601629 systemd-networkd[1017]: cilium_host: Link UP Sep 6 00:34:59.601939 systemd-networkd[1017]: cilium_net: Link UP Sep 6 00:34:59.608507 systemd-networkd[1017]: cilium_net: Gained carrier Sep 6 00:34:59.611443 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Sep 6 00:34:59.611740 systemd-networkd[1017]: cilium_host: Gained carrier Sep 6 00:34:59.764000 systemd-networkd[1017]: cilium_vxlan: Link UP Sep 6 00:34:59.764012 systemd-networkd[1017]: cilium_vxlan: Gained carrier Sep 6 00:35:00.013646 systemd-networkd[1017]: cilium_host: Gained IPv6LL Sep 6 00:35:00.043439 kernel: NET: Registered PF_ALG protocol family Sep 6 00:35:00.509973 systemd-networkd[1017]: cilium_net: Gained IPv6LL Sep 6 00:35:00.896835 systemd-networkd[1017]: lxc_health: Link UP Sep 6 00:35:00.908426 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 6 00:35:00.917033 systemd-networkd[1017]: lxc_health: Gained carrier Sep 6 00:35:01.342201 systemd-networkd[1017]: cilium_vxlan: Gained IPv6LL Sep 6 00:35:01.427190 systemd-networkd[1017]: lxc5cc4ed987413: Link UP Sep 6 00:35:01.440428 kernel: eth0: renamed from tmp3a322 Sep 6 00:35:01.452488 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc5cc4ed987413: link becomes ready Sep 6 00:35:01.455085 systemd-networkd[1017]: lxc5cc4ed987413: Gained carrier Sep 6 00:35:01.459130 systemd-networkd[1017]: lxcdd8edcb752d0: Link UP Sep 6 00:35:01.468431 kernel: eth0: renamed from tmp0fdfe Sep 6 00:35:01.483550 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcdd8edcb752d0: link becomes ready Sep 6 00:35:01.492931 systemd-networkd[1017]: lxcdd8edcb752d0: Gained carrier Sep 6 00:35:01.785726 kubelet[2004]: I0906 00:35:01.785649 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-jwtq9" podStartSLOduration=11.991873024 podStartE2EDuration="23.785594843s" podCreationTimestamp="2025-09-06 00:34:38 +0000 UTC" firstStartedPulling="2025-09-06 00:34:39.876993026 +0000 UTC m=+8.030678578" lastFinishedPulling="2025-09-06 00:34:51.670714831 +0000 UTC m=+19.824400397" observedRunningTime="2025-09-06 00:34:58.16555534 +0000 UTC m=+26.319240920" watchObservedRunningTime="2025-09-06 00:35:01.785594843 +0000 UTC m=+29.939280423" Sep 6 00:35:02.174161 systemd-networkd[1017]: lxc_health: Gained IPv6LL Sep 6 00:35:02.558044 systemd-networkd[1017]: lxc5cc4ed987413: Gained IPv6LL Sep 6 00:35:02.685990 systemd-networkd[1017]: lxcdd8edcb752d0: Gained IPv6LL Sep 6 00:35:06.311671 env[1222]: time="2025-09-06T00:35:06.311327784Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:35:06.311671 env[1222]: time="2025-09-06T00:35:06.311382033Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:35:06.311671 env[1222]: time="2025-09-06T00:35:06.311424560Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:35:06.312363 env[1222]: time="2025-09-06T00:35:06.312233034Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0fdfee4a82d8c0cee6b0d2bbe4d877e1e8354f9d5ef28f4811c2cac7309c1385 pid=3191 runtime=io.containerd.runc.v2 Sep 6 00:35:06.351542 env[1222]: time="2025-09-06T00:35:06.351434423Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:35:06.351838 env[1222]: time="2025-09-06T00:35:06.351791608Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:35:06.352014 env[1222]: time="2025-09-06T00:35:06.351975456Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:35:06.352483 env[1222]: time="2025-09-06T00:35:06.352426886Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3a322b05dc0282583eaf00aa173b8ddb7e60327e854a2775698280e349faa37c pid=3210 runtime=io.containerd.runc.v2 Sep 6 00:35:06.353518 systemd[1]: Started cri-containerd-0fdfee4a82d8c0cee6b0d2bbe4d877e1e8354f9d5ef28f4811c2cac7309c1385.scope. Sep 6 00:35:06.383866 systemd[1]: run-containerd-runc-k8s.io-0fdfee4a82d8c0cee6b0d2bbe4d877e1e8354f9d5ef28f4811c2cac7309c1385-runc.5rTXKy.mount: Deactivated successfully. Sep 6 00:35:06.413702 systemd[1]: Started cri-containerd-3a322b05dc0282583eaf00aa173b8ddb7e60327e854a2775698280e349faa37c.scope. Sep 6 00:35:06.534673 env[1222]: time="2025-09-06T00:35:06.534618239Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-kmv9n,Uid:67f5d924-e914-4357-aecf-b9d0c736dfcb,Namespace:kube-system,Attempt:0,} returns sandbox id \"3a322b05dc0282583eaf00aa173b8ddb7e60327e854a2775698280e349faa37c\"" Sep 6 00:35:06.540789 env[1222]: time="2025-09-06T00:35:06.540742221Z" level=info msg="CreateContainer within sandbox \"3a322b05dc0282583eaf00aa173b8ddb7e60327e854a2775698280e349faa37c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 6 00:35:06.546255 env[1222]: time="2025-09-06T00:35:06.546197150Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-cn6v6,Uid:5999b4c9-2550-4912-9432-ee4d2d455edf,Namespace:kube-system,Attempt:0,} returns sandbox id \"0fdfee4a82d8c0cee6b0d2bbe4d877e1e8354f9d5ef28f4811c2cac7309c1385\"" Sep 6 00:35:06.551102 env[1222]: time="2025-09-06T00:35:06.551058631Z" level=info msg="CreateContainer within sandbox \"0fdfee4a82d8c0cee6b0d2bbe4d877e1e8354f9d5ef28f4811c2cac7309c1385\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Sep 6 00:35:06.582103 env[1222]: time="2025-09-06T00:35:06.581927445Z" level=info msg="CreateContainer within sandbox \"0fdfee4a82d8c0cee6b0d2bbe4d877e1e8354f9d5ef28f4811c2cac7309c1385\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"98b79494c37193037d7662fda3ecc7dbc04fb271befa6cebec4f844a6d4e536e\"" Sep 6 00:35:06.583454 env[1222]: time="2025-09-06T00:35:06.583415250Z" level=info msg="StartContainer for \"98b79494c37193037d7662fda3ecc7dbc04fb271befa6cebec4f844a6d4e536e\"" Sep 6 00:35:06.589597 env[1222]: time="2025-09-06T00:35:06.589552117Z" level=info msg="CreateContainer within sandbox \"3a322b05dc0282583eaf00aa173b8ddb7e60327e854a2775698280e349faa37c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"2c3c52f58ee6fb8646930402370b14fce4cf5f9328f7f243617710f53a4800fa\"" Sep 6 00:35:06.592063 env[1222]: time="2025-09-06T00:35:06.592014508Z" level=info msg="StartContainer for \"2c3c52f58ee6fb8646930402370b14fce4cf5f9328f7f243617710f53a4800fa\"" Sep 6 00:35:06.629973 systemd[1]: Started cri-containerd-98b79494c37193037d7662fda3ecc7dbc04fb271befa6cebec4f844a6d4e536e.scope. Sep 6 00:35:06.639344 systemd[1]: Started cri-containerd-2c3c52f58ee6fb8646930402370b14fce4cf5f9328f7f243617710f53a4800fa.scope. Sep 6 00:35:06.707435 env[1222]: time="2025-09-06T00:35:06.707024266Z" level=info msg="StartContainer for \"98b79494c37193037d7662fda3ecc7dbc04fb271befa6cebec4f844a6d4e536e\" returns successfully" Sep 6 00:35:06.711226 env[1222]: time="2025-09-06T00:35:06.711175905Z" level=info msg="StartContainer for \"2c3c52f58ee6fb8646930402370b14fce4cf5f9328f7f243617710f53a4800fa\" returns successfully" Sep 6 00:35:07.170215 kubelet[2004]: I0906 00:35:07.170130 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-cn6v6" podStartSLOduration=28.170106979 podStartE2EDuration="28.170106979s" podCreationTimestamp="2025-09-06 00:34:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:35:07.169058021 +0000 UTC m=+35.322743600" watchObservedRunningTime="2025-09-06 00:35:07.170106979 +0000 UTC m=+35.323792537" Sep 6 00:35:07.320236 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2312317992.mount: Deactivated successfully. Sep 6 00:35:34.548674 systemd[1]: Started sshd@5-10.128.0.94:22-139.178.89.65:42148.service. Sep 6 00:35:34.839518 sshd[3360]: Accepted publickey for core from 139.178.89.65 port 42148 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:34.841633 sshd[3360]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:34.849025 systemd[1]: Started session-6.scope. Sep 6 00:35:34.849701 systemd-logind[1211]: New session 6 of user core. Sep 6 00:35:35.152518 sshd[3360]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:35.157566 systemd[1]: sshd@5-10.128.0.94:22-139.178.89.65:42148.service: Deactivated successfully. Sep 6 00:35:35.158824 systemd[1]: session-6.scope: Deactivated successfully. Sep 6 00:35:35.160234 systemd-logind[1211]: Session 6 logged out. Waiting for processes to exit. Sep 6 00:35:35.161954 systemd-logind[1211]: Removed session 6. Sep 6 00:35:40.200856 systemd[1]: Started sshd@6-10.128.0.94:22-139.178.89.65:57580.service. Sep 6 00:35:40.488889 sshd[3374]: Accepted publickey for core from 139.178.89.65 port 57580 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:40.490992 sshd[3374]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:40.498334 systemd[1]: Started session-7.scope. Sep 6 00:35:40.499509 systemd-logind[1211]: New session 7 of user core. Sep 6 00:35:40.775895 sshd[3374]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:40.780845 systemd[1]: sshd@6-10.128.0.94:22-139.178.89.65:57580.service: Deactivated successfully. Sep 6 00:35:40.782084 systemd[1]: session-7.scope: Deactivated successfully. Sep 6 00:35:40.783484 systemd-logind[1211]: Session 7 logged out. Waiting for processes to exit. Sep 6 00:35:40.785002 systemd-logind[1211]: Removed session 7. Sep 6 00:35:45.826001 systemd[1]: Started sshd@7-10.128.0.94:22-139.178.89.65:57588.service. Sep 6 00:35:46.119528 sshd[3389]: Accepted publickey for core from 139.178.89.65 port 57588 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:46.121842 sshd[3389]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:46.129244 systemd[1]: Started session-8.scope. Sep 6 00:35:46.130011 systemd-logind[1211]: New session 8 of user core. Sep 6 00:35:46.408817 sshd[3389]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:46.413472 systemd-logind[1211]: Session 8 logged out. Waiting for processes to exit. Sep 6 00:35:46.413796 systemd[1]: sshd@7-10.128.0.94:22-139.178.89.65:57588.service: Deactivated successfully. Sep 6 00:35:46.414994 systemd[1]: session-8.scope: Deactivated successfully. Sep 6 00:35:46.416334 systemd-logind[1211]: Removed session 8. Sep 6 00:35:51.456374 systemd[1]: Started sshd@8-10.128.0.94:22-139.178.89.65:48638.service. Sep 6 00:35:51.742086 sshd[3402]: Accepted publickey for core from 139.178.89.65 port 48638 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:51.744328 sshd[3402]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:51.751808 systemd[1]: Started session-9.scope. Sep 6 00:35:51.752757 systemd-logind[1211]: New session 9 of user core. Sep 6 00:35:52.037793 sshd[3402]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:52.042584 systemd[1]: sshd@8-10.128.0.94:22-139.178.89.65:48638.service: Deactivated successfully. Sep 6 00:35:52.043761 systemd[1]: session-9.scope: Deactivated successfully. Sep 6 00:35:52.044894 systemd-logind[1211]: Session 9 logged out. Waiting for processes to exit. Sep 6 00:35:52.046278 systemd-logind[1211]: Removed session 9. Sep 6 00:35:57.085803 systemd[1]: Started sshd@9-10.128.0.94:22-139.178.89.65:48644.service. Sep 6 00:35:57.376137 sshd[3414]: Accepted publickey for core from 139.178.89.65 port 48644 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:57.377853 sshd[3414]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:57.384873 systemd[1]: Started session-10.scope. Sep 6 00:35:57.385745 systemd-logind[1211]: New session 10 of user core. Sep 6 00:35:57.667881 sshd[3414]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:57.673140 systemd[1]: sshd@9-10.128.0.94:22-139.178.89.65:48644.service: Deactivated successfully. Sep 6 00:35:57.674389 systemd[1]: session-10.scope: Deactivated successfully. Sep 6 00:35:57.676348 systemd-logind[1211]: Session 10 logged out. Waiting for processes to exit. Sep 6 00:35:57.677693 systemd-logind[1211]: Removed session 10. Sep 6 00:35:57.716431 systemd[1]: Started sshd@10-10.128.0.94:22-139.178.89.65:48654.service. Sep 6 00:35:58.007881 sshd[3426]: Accepted publickey for core from 139.178.89.65 port 48654 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:58.009986 sshd[3426]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:58.017220 systemd[1]: Started session-11.scope. Sep 6 00:35:58.017914 systemd-logind[1211]: New session 11 of user core. Sep 6 00:35:58.352454 sshd[3426]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:58.357683 systemd-logind[1211]: Session 11 logged out. Waiting for processes to exit. Sep 6 00:35:58.358336 systemd[1]: sshd@10-10.128.0.94:22-139.178.89.65:48654.service: Deactivated successfully. Sep 6 00:35:58.359545 systemd[1]: session-11.scope: Deactivated successfully. Sep 6 00:35:58.361996 systemd-logind[1211]: Removed session 11. Sep 6 00:35:58.397900 systemd[1]: Started sshd@11-10.128.0.94:22-139.178.89.65:48670.service. Sep 6 00:35:58.690412 sshd[3436]: Accepted publickey for core from 139.178.89.65 port 48670 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:35:58.692009 sshd[3436]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:35:58.698484 systemd-logind[1211]: New session 12 of user core. Sep 6 00:35:58.699543 systemd[1]: Started session-12.scope. Sep 6 00:35:58.990158 sshd[3436]: pam_unix(sshd:session): session closed for user core Sep 6 00:35:58.995130 systemd[1]: sshd@11-10.128.0.94:22-139.178.89.65:48670.service: Deactivated successfully. Sep 6 00:35:58.996222 systemd[1]: session-12.scope: Deactivated successfully. Sep 6 00:35:58.996901 systemd-logind[1211]: Session 12 logged out. Waiting for processes to exit. Sep 6 00:35:58.998310 systemd-logind[1211]: Removed session 12. Sep 6 00:36:04.036925 systemd[1]: Started sshd@12-10.128.0.94:22-139.178.89.65:43480.service. Sep 6 00:36:04.325841 sshd[3448]: Accepted publickey for core from 139.178.89.65 port 43480 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:04.327896 sshd[3448]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:04.334261 systemd-logind[1211]: New session 13 of user core. Sep 6 00:36:04.335125 systemd[1]: Started session-13.scope. Sep 6 00:36:04.616143 sshd[3448]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:04.620995 systemd[1]: sshd@12-10.128.0.94:22-139.178.89.65:43480.service: Deactivated successfully. Sep 6 00:36:04.622208 systemd[1]: session-13.scope: Deactivated successfully. Sep 6 00:36:04.623346 systemd-logind[1211]: Session 13 logged out. Waiting for processes to exit. Sep 6 00:36:04.624920 systemd-logind[1211]: Removed session 13. Sep 6 00:36:09.663705 systemd[1]: Started sshd@13-10.128.0.94:22-139.178.89.65:43488.service. Sep 6 00:36:09.952108 sshd[3461]: Accepted publickey for core from 139.178.89.65 port 43488 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:09.954485 sshd[3461]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:09.961532 systemd[1]: Started session-14.scope. Sep 6 00:36:09.962194 systemd-logind[1211]: New session 14 of user core. Sep 6 00:36:10.244366 sshd[3461]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:10.248751 systemd[1]: sshd@13-10.128.0.94:22-139.178.89.65:43488.service: Deactivated successfully. Sep 6 00:36:10.249997 systemd[1]: session-14.scope: Deactivated successfully. Sep 6 00:36:10.251121 systemd-logind[1211]: Session 14 logged out. Waiting for processes to exit. Sep 6 00:36:10.252485 systemd-logind[1211]: Removed session 14. Sep 6 00:36:10.290957 systemd[1]: Started sshd@14-10.128.0.94:22-139.178.89.65:44196.service. Sep 6 00:36:10.579528 sshd[3475]: Accepted publickey for core from 139.178.89.65 port 44196 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:10.582076 sshd[3475]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:10.588604 systemd-logind[1211]: New session 15 of user core. Sep 6 00:36:10.589010 systemd[1]: Started session-15.scope. Sep 6 00:36:10.942566 sshd[3475]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:10.947565 systemd[1]: sshd@14-10.128.0.94:22-139.178.89.65:44196.service: Deactivated successfully. Sep 6 00:36:10.948649 systemd[1]: session-15.scope: Deactivated successfully. Sep 6 00:36:10.949575 systemd-logind[1211]: Session 15 logged out. Waiting for processes to exit. Sep 6 00:36:10.950930 systemd-logind[1211]: Removed session 15. Sep 6 00:36:10.990129 systemd[1]: Started sshd@15-10.128.0.94:22-139.178.89.65:44200.service. Sep 6 00:36:11.279441 sshd[3484]: Accepted publickey for core from 139.178.89.65 port 44200 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:11.281452 sshd[3484]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:11.288660 systemd[1]: Started session-16.scope. Sep 6 00:36:11.289894 systemd-logind[1211]: New session 16 of user core. Sep 6 00:36:13.122184 sshd[3484]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:13.129517 systemd-logind[1211]: Session 16 logged out. Waiting for processes to exit. Sep 6 00:36:13.130729 systemd[1]: sshd@15-10.128.0.94:22-139.178.89.65:44200.service: Deactivated successfully. Sep 6 00:36:13.132014 systemd[1]: session-16.scope: Deactivated successfully. Sep 6 00:36:13.133830 systemd-logind[1211]: Removed session 16. Sep 6 00:36:13.170673 systemd[1]: Started sshd@16-10.128.0.94:22-139.178.89.65:44210.service. Sep 6 00:36:13.461000 sshd[3501]: Accepted publickey for core from 139.178.89.65 port 44210 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:13.463192 sshd[3501]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:13.471283 systemd[1]: Started session-17.scope. Sep 6 00:36:13.472472 systemd-logind[1211]: New session 17 of user core. Sep 6 00:36:13.890142 sshd[3501]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:13.895235 systemd[1]: sshd@16-10.128.0.94:22-139.178.89.65:44210.service: Deactivated successfully. Sep 6 00:36:13.896509 systemd[1]: session-17.scope: Deactivated successfully. Sep 6 00:36:13.897393 systemd-logind[1211]: Session 17 logged out. Waiting for processes to exit. Sep 6 00:36:13.898773 systemd-logind[1211]: Removed session 17. Sep 6 00:36:13.937590 systemd[1]: Started sshd@17-10.128.0.94:22-139.178.89.65:44224.service. Sep 6 00:36:14.227048 sshd[3511]: Accepted publickey for core from 139.178.89.65 port 44224 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:14.229305 sshd[3511]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:14.236495 systemd-logind[1211]: New session 18 of user core. Sep 6 00:36:14.236566 systemd[1]: Started session-18.scope. Sep 6 00:36:14.512274 sshd[3511]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:14.516839 systemd[1]: sshd@17-10.128.0.94:22-139.178.89.65:44224.service: Deactivated successfully. Sep 6 00:36:14.518038 systemd[1]: session-18.scope: Deactivated successfully. Sep 6 00:36:14.518985 systemd-logind[1211]: Session 18 logged out. Waiting for processes to exit. Sep 6 00:36:14.520307 systemd-logind[1211]: Removed session 18. Sep 6 00:36:19.561030 systemd[1]: Started sshd@18-10.128.0.94:22-139.178.89.65:44240.service. Sep 6 00:36:19.853850 sshd[3523]: Accepted publickey for core from 139.178.89.65 port 44240 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:19.855848 sshd[3523]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:19.863076 systemd[1]: Started session-19.scope. Sep 6 00:36:19.863937 systemd-logind[1211]: New session 19 of user core. Sep 6 00:36:20.160019 sshd[3523]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:20.165430 systemd-logind[1211]: Session 19 logged out. Waiting for processes to exit. Sep 6 00:36:20.166054 systemd[1]: sshd@18-10.128.0.94:22-139.178.89.65:44240.service: Deactivated successfully. Sep 6 00:36:20.167228 systemd[1]: session-19.scope: Deactivated successfully. Sep 6 00:36:20.169106 systemd-logind[1211]: Removed session 19. Sep 6 00:36:25.209303 systemd[1]: Started sshd@19-10.128.0.94:22-139.178.89.65:33496.service. Sep 6 00:36:25.506539 sshd[3538]: Accepted publickey for core from 139.178.89.65 port 33496 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:25.508880 sshd[3538]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:25.517032 systemd[1]: Started session-20.scope. Sep 6 00:36:25.517809 systemd-logind[1211]: New session 20 of user core. Sep 6 00:36:25.803088 sshd[3538]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:25.808351 systemd[1]: sshd@19-10.128.0.94:22-139.178.89.65:33496.service: Deactivated successfully. Sep 6 00:36:25.809618 systemd[1]: session-20.scope: Deactivated successfully. Sep 6 00:36:25.810598 systemd-logind[1211]: Session 20 logged out. Waiting for processes to exit. Sep 6 00:36:25.811949 systemd-logind[1211]: Removed session 20. Sep 6 00:36:30.851504 systemd[1]: Started sshd@20-10.128.0.94:22-139.178.89.65:46618.service. Sep 6 00:36:31.145158 sshd[3550]: Accepted publickey for core from 139.178.89.65 port 46618 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:31.148423 sshd[3550]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:31.158698 systemd[1]: Started session-21.scope. Sep 6 00:36:31.159842 systemd-logind[1211]: New session 21 of user core. Sep 6 00:36:31.442229 sshd[3550]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:31.447375 systemd[1]: sshd@20-10.128.0.94:22-139.178.89.65:46618.service: Deactivated successfully. Sep 6 00:36:31.448626 systemd[1]: session-21.scope: Deactivated successfully. Sep 6 00:36:31.449477 systemd-logind[1211]: Session 21 logged out. Waiting for processes to exit. Sep 6 00:36:31.451061 systemd-logind[1211]: Removed session 21. Sep 6 00:36:31.491754 systemd[1]: Started sshd@21-10.128.0.94:22-139.178.89.65:46626.service. Sep 6 00:36:31.784563 sshd[3562]: Accepted publickey for core from 139.178.89.65 port 46626 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:31.786576 sshd[3562]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:31.794216 systemd[1]: Started session-22.scope. Sep 6 00:36:31.795177 systemd-logind[1211]: New session 22 of user core. Sep 6 00:36:33.572872 kubelet[2004]: I0906 00:36:33.572772 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-kmv9n" podStartSLOduration=114.572721802 podStartE2EDuration="1m54.572721802s" podCreationTimestamp="2025-09-06 00:34:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:35:07.206967495 +0000 UTC m=+35.360653079" watchObservedRunningTime="2025-09-06 00:36:33.572721802 +0000 UTC m=+121.726407382" Sep 6 00:36:33.587668 env[1222]: time="2025-09-06T00:36:33.587610533Z" level=info msg="StopContainer for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" with timeout 30 (s)" Sep 6 00:36:33.588812 env[1222]: time="2025-09-06T00:36:33.588764536Z" level=info msg="Stop container \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" with signal terminated" Sep 6 00:36:33.618018 systemd[1]: run-containerd-runc-k8s.io-44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3-runc.cwPkK1.mount: Deactivated successfully. Sep 6 00:36:33.629555 systemd[1]: cri-containerd-7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89.scope: Deactivated successfully. Sep 6 00:36:33.664431 env[1222]: time="2025-09-06T00:36:33.664315573Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Sep 6 00:36:33.678169 env[1222]: time="2025-09-06T00:36:33.678119327Z" level=info msg="StopContainer for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" with timeout 2 (s)" Sep 6 00:36:33.681923 env[1222]: time="2025-09-06T00:36:33.681871552Z" level=info msg="Stop container \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" with signal terminated" Sep 6 00:36:33.682083 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89-rootfs.mount: Deactivated successfully. Sep 6 00:36:33.695082 systemd-networkd[1017]: lxc_health: Link DOWN Sep 6 00:36:33.695095 systemd-networkd[1017]: lxc_health: Lost carrier Sep 6 00:36:33.704847 env[1222]: time="2025-09-06T00:36:33.702331545Z" level=info msg="shim disconnected" id=7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89 Sep 6 00:36:33.704847 env[1222]: time="2025-09-06T00:36:33.702563748Z" level=warning msg="cleaning up after shim disconnected" id=7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89 namespace=k8s.io Sep 6 00:36:33.704847 env[1222]: time="2025-09-06T00:36:33.702591485Z" level=info msg="cleaning up dead shim" Sep 6 00:36:33.729112 env[1222]: time="2025-09-06T00:36:33.729069451Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:33Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3617 runtime=io.containerd.runc.v2\n" Sep 6 00:36:33.732136 systemd[1]: cri-containerd-44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3.scope: Deactivated successfully. Sep 6 00:36:33.732614 systemd[1]: cri-containerd-44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3.scope: Consumed 9.005s CPU time. Sep 6 00:36:33.734069 env[1222]: time="2025-09-06T00:36:33.733926388Z" level=info msg="StopContainer for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" returns successfully" Sep 6 00:36:33.735688 env[1222]: time="2025-09-06T00:36:33.735641589Z" level=info msg="StopPodSandbox for \"6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee\"" Sep 6 00:36:33.735924 env[1222]: time="2025-09-06T00:36:33.735888731Z" level=info msg="Container to stop \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:33.739635 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee-shm.mount: Deactivated successfully. Sep 6 00:36:33.754502 systemd[1]: cri-containerd-6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee.scope: Deactivated successfully. Sep 6 00:36:33.788295 env[1222]: time="2025-09-06T00:36:33.788226809Z" level=info msg="shim disconnected" id=44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3 Sep 6 00:36:33.788295 env[1222]: time="2025-09-06T00:36:33.788294895Z" level=warning msg="cleaning up after shim disconnected" id=44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3 namespace=k8s.io Sep 6 00:36:33.788295 env[1222]: time="2025-09-06T00:36:33.788311800Z" level=info msg="cleaning up dead shim" Sep 6 00:36:33.803623 env[1222]: time="2025-09-06T00:36:33.803522684Z" level=info msg="shim disconnected" id=6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee Sep 6 00:36:33.803623 env[1222]: time="2025-09-06T00:36:33.803594980Z" level=warning msg="cleaning up after shim disconnected" id=6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee namespace=k8s.io Sep 6 00:36:33.803623 env[1222]: time="2025-09-06T00:36:33.803615414Z" level=info msg="cleaning up dead shim" Sep 6 00:36:33.806916 env[1222]: time="2025-09-06T00:36:33.806829999Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:33Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3665 runtime=io.containerd.runc.v2\n" Sep 6 00:36:33.809934 env[1222]: time="2025-09-06T00:36:33.809881574Z" level=info msg="StopContainer for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" returns successfully" Sep 6 00:36:33.810928 env[1222]: time="2025-09-06T00:36:33.810853626Z" level=info msg="StopPodSandbox for \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\"" Sep 6 00:36:33.811120 env[1222]: time="2025-09-06T00:36:33.810936154Z" level=info msg="Container to stop \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:33.811120 env[1222]: time="2025-09-06T00:36:33.810961897Z" level=info msg="Container to stop \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:33.811120 env[1222]: time="2025-09-06T00:36:33.810981704Z" level=info msg="Container to stop \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:33.811120 env[1222]: time="2025-09-06T00:36:33.811003432Z" level=info msg="Container to stop \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:33.811120 env[1222]: time="2025-09-06T00:36:33.811025593Z" level=info msg="Container to stop \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:33.820590 env[1222]: time="2025-09-06T00:36:33.820532205Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:33Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3678 runtime=io.containerd.runc.v2\n" Sep 6 00:36:33.821036 env[1222]: time="2025-09-06T00:36:33.820991924Z" level=info msg="TearDown network for sandbox \"6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee\" successfully" Sep 6 00:36:33.821161 env[1222]: time="2025-09-06T00:36:33.821033367Z" level=info msg="StopPodSandbox for \"6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee\" returns successfully" Sep 6 00:36:33.828305 systemd[1]: cri-containerd-22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379.scope: Deactivated successfully. Sep 6 00:36:33.874171 env[1222]: time="2025-09-06T00:36:33.874105666Z" level=info msg="shim disconnected" id=22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379 Sep 6 00:36:33.875488 env[1222]: time="2025-09-06T00:36:33.875445090Z" level=warning msg="cleaning up after shim disconnected" id=22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379 namespace=k8s.io Sep 6 00:36:33.875684 env[1222]: time="2025-09-06T00:36:33.875643697Z" level=info msg="cleaning up dead shim" Sep 6 00:36:33.890641 env[1222]: time="2025-09-06T00:36:33.890580657Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:33Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3710 runtime=io.containerd.runc.v2\n" Sep 6 00:36:33.891390 env[1222]: time="2025-09-06T00:36:33.891342945Z" level=info msg="TearDown network for sandbox \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" successfully" Sep 6 00:36:33.891876 env[1222]: time="2025-09-06T00:36:33.891801559Z" level=info msg="StopPodSandbox for \"22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379\" returns successfully" Sep 6 00:36:33.968797 kubelet[2004]: I0906 00:36:33.968693 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cni-path\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.968797 kubelet[2004]: I0906 00:36:33.968798 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-config-path\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969505 kubelet[2004]: I0906 00:36:33.968839 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jd8z\" (UniqueName: \"kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-kube-api-access-9jd8z\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969505 kubelet[2004]: I0906 00:36:33.968872 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hostproc\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969505 kubelet[2004]: I0906 00:36:33.968903 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jtr9\" (UniqueName: \"kubernetes.io/projected/448ef295-16b6-4f6a-9a49-a911a225b515-kube-api-access-5jtr9\") pod \"448ef295-16b6-4f6a-9a49-a911a225b515\" (UID: \"448ef295-16b6-4f6a-9a49-a911a225b515\") " Sep 6 00:36:33.969505 kubelet[2004]: I0906 00:36:33.968934 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-net\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969505 kubelet[2004]: I0906 00:36:33.968959 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-etc-cni-netd\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969505 kubelet[2004]: I0906 00:36:33.968988 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/448ef295-16b6-4f6a-9a49-a911a225b515-cilium-config-path\") pod \"448ef295-16b6-4f6a-9a49-a911a225b515\" (UID: \"448ef295-16b6-4f6a-9a49-a911a225b515\") " Sep 6 00:36:33.969871 kubelet[2004]: I0906 00:36:33.969012 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-kernel\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969871 kubelet[2004]: I0906 00:36:33.969036 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-cgroup\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969871 kubelet[2004]: I0906 00:36:33.969063 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-xtables-lock\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969871 kubelet[2004]: I0906 00:36:33.969198 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hubble-tls\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969871 kubelet[2004]: I0906 00:36:33.969317 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3532905a-8f04-4418-b4ad-6a6a50e8f32b-clustermesh-secrets\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.969871 kubelet[2004]: I0906 00:36:33.969355 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-run\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.970210 kubelet[2004]: I0906 00:36:33.969383 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-lib-modules\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.970210 kubelet[2004]: I0906 00:36:33.969428 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-bpf-maps\") pod \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\" (UID: \"3532905a-8f04-4418-b4ad-6a6a50e8f32b\") " Sep 6 00:36:33.970210 kubelet[2004]: I0906 00:36:33.969565 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.970210 kubelet[2004]: I0906 00:36:33.969627 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cni-path" (OuterVolumeSpecName: "cni-path") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.973100 kubelet[2004]: I0906 00:36:33.973029 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 6 00:36:33.973616 kubelet[2004]: I0906 00:36:33.973571 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.973756 kubelet[2004]: I0906 00:36:33.973635 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hostproc" (OuterVolumeSpecName: "hostproc") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.976021 kubelet[2004]: I0906 00:36:33.975982 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.976208 kubelet[2004]: I0906 00:36:33.975982 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.976341 kubelet[2004]: I0906 00:36:33.976025 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.976667 kubelet[2004]: I0906 00:36:33.976635 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.979352 kubelet[2004]: I0906 00:36:33.979306 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/448ef295-16b6-4f6a-9a49-a911a225b515-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "448ef295-16b6-4f6a-9a49-a911a225b515" (UID: "448ef295-16b6-4f6a-9a49-a911a225b515"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 6 00:36:33.979519 kubelet[2004]: I0906 00:36:33.979391 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.980276 kubelet[2004]: I0906 00:36:33.980225 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:33.980459 kubelet[2004]: I0906 00:36:33.980385 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/448ef295-16b6-4f6a-9a49-a911a225b515-kube-api-access-5jtr9" (OuterVolumeSpecName: "kube-api-access-5jtr9") pod "448ef295-16b6-4f6a-9a49-a911a225b515" (UID: "448ef295-16b6-4f6a-9a49-a911a225b515"). InnerVolumeSpecName "kube-api-access-5jtr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 6 00:36:33.983503 kubelet[2004]: I0906 00:36:33.983464 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-kube-api-access-9jd8z" (OuterVolumeSpecName: "kube-api-access-9jd8z") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "kube-api-access-9jd8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 6 00:36:33.984640 kubelet[2004]: I0906 00:36:33.984600 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 6 00:36:33.988043 kubelet[2004]: I0906 00:36:33.988000 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3532905a-8f04-4418-b4ad-6a6a50e8f32b-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "3532905a-8f04-4418-b4ad-6a6a50e8f32b" (UID: "3532905a-8f04-4418-b4ad-6a6a50e8f32b"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 6 00:36:34.030070 systemd[1]: Removed slice kubepods-burstable-pod3532905a_8f04_4418_b4ad_6a6a50e8f32b.slice. Sep 6 00:36:34.030283 systemd[1]: kubepods-burstable-pod3532905a_8f04_4418_b4ad_6a6a50e8f32b.slice: Consumed 9.154s CPU time. Sep 6 00:36:34.035012 systemd[1]: Removed slice kubepods-besteffort-pod448ef295_16b6_4f6a_9a49_a911a225b515.slice. Sep 6 00:36:34.070382 kubelet[2004]: I0906 00:36:34.070333 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-run\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.070676 kubelet[2004]: I0906 00:36:34.070653 2004 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-lib-modules\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.070821 kubelet[2004]: I0906 00:36:34.070802 2004 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-bpf-maps\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.071622 kubelet[2004]: I0906 00:36:34.071590 2004 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jd8z\" (UniqueName: \"kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-kube-api-access-9jd8z\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.071950 kubelet[2004]: I0906 00:36:34.071925 2004 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cni-path\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072291 kubelet[2004]: I0906 00:36:34.072128 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-config-path\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072544 kubelet[2004]: I0906 00:36:34.072475 2004 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hostproc\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072544 kubelet[2004]: I0906 00:36:34.072533 2004 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jtr9\" (UniqueName: \"kubernetes.io/projected/448ef295-16b6-4f6a-9a49-a911a225b515-kube-api-access-5jtr9\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072719 kubelet[2004]: I0906 00:36:34.072555 2004 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-etc-cni-netd\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072719 kubelet[2004]: I0906 00:36:34.072573 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/448ef295-16b6-4f6a-9a49-a911a225b515-cilium-config-path\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072719 kubelet[2004]: I0906 00:36:34.072590 2004 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-net\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072719 kubelet[2004]: I0906 00:36:34.072672 2004 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-host-proc-sys-kernel\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072719 kubelet[2004]: I0906 00:36:34.072692 2004 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3532905a-8f04-4418-b4ad-6a6a50e8f32b-hubble-tls\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072973 kubelet[2004]: I0906 00:36:34.072732 2004 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3532905a-8f04-4418-b4ad-6a6a50e8f32b-clustermesh-secrets\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072973 kubelet[2004]: I0906 00:36:34.072752 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-cilium-cgroup\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.072973 kubelet[2004]: I0906 00:36:34.072771 2004 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3532905a-8f04-4418-b4ad-6a6a50e8f32b-xtables-lock\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:34.369533 kubelet[2004]: I0906 00:36:34.369490 2004 scope.go:117] "RemoveContainer" containerID="7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89" Sep 6 00:36:34.382744 env[1222]: time="2025-09-06T00:36:34.382167302Z" level=info msg="RemoveContainer for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\"" Sep 6 00:36:34.391585 env[1222]: time="2025-09-06T00:36:34.391532182Z" level=info msg="RemoveContainer for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" returns successfully" Sep 6 00:36:34.392820 kubelet[2004]: I0906 00:36:34.392780 2004 scope.go:117] "RemoveContainer" containerID="7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89" Sep 6 00:36:34.395209 env[1222]: time="2025-09-06T00:36:34.394262185Z" level=error msg="ContainerStatus for \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\": not found" Sep 6 00:36:34.399453 kubelet[2004]: E0906 00:36:34.399365 2004 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\": not found" containerID="7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89" Sep 6 00:36:34.399638 kubelet[2004]: I0906 00:36:34.399455 2004 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89"} err="failed to get container status \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\": rpc error: code = NotFound desc = an error occurred when try to find container \"7ac4f810e27b76d3f02ffcfc68e62278f4cdea37bd8a8052f4bbca940f0bbd89\": not found" Sep 6 00:36:34.399638 kubelet[2004]: I0906 00:36:34.399591 2004 scope.go:117] "RemoveContainer" containerID="44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3" Sep 6 00:36:34.405701 env[1222]: time="2025-09-06T00:36:34.405183026Z" level=info msg="RemoveContainer for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\"" Sep 6 00:36:34.410810 env[1222]: time="2025-09-06T00:36:34.410752259Z" level=info msg="RemoveContainer for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" returns successfully" Sep 6 00:36:34.411075 kubelet[2004]: I0906 00:36:34.411043 2004 scope.go:117] "RemoveContainer" containerID="04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716" Sep 6 00:36:34.412569 env[1222]: time="2025-09-06T00:36:34.412522527Z" level=info msg="RemoveContainer for \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\"" Sep 6 00:36:34.419480 env[1222]: time="2025-09-06T00:36:34.418720410Z" level=info msg="RemoveContainer for \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\" returns successfully" Sep 6 00:36:34.419760 kubelet[2004]: I0906 00:36:34.419011 2004 scope.go:117] "RemoveContainer" containerID="7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8" Sep 6 00:36:34.424843 env[1222]: time="2025-09-06T00:36:34.424740545Z" level=info msg="RemoveContainer for \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\"" Sep 6 00:36:34.429971 env[1222]: time="2025-09-06T00:36:34.429889431Z" level=info msg="RemoveContainer for \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\" returns successfully" Sep 6 00:36:34.430387 kubelet[2004]: I0906 00:36:34.430357 2004 scope.go:117] "RemoveContainer" containerID="85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c" Sep 6 00:36:34.432330 env[1222]: time="2025-09-06T00:36:34.432260983Z" level=info msg="RemoveContainer for \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\"" Sep 6 00:36:34.437210 env[1222]: time="2025-09-06T00:36:34.437143989Z" level=info msg="RemoveContainer for \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\" returns successfully" Sep 6 00:36:34.437552 kubelet[2004]: I0906 00:36:34.437518 2004 scope.go:117] "RemoveContainer" containerID="351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279" Sep 6 00:36:34.439141 env[1222]: time="2025-09-06T00:36:34.439079400Z" level=info msg="RemoveContainer for \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\"" Sep 6 00:36:34.444381 env[1222]: time="2025-09-06T00:36:34.444303248Z" level=info msg="RemoveContainer for \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\" returns successfully" Sep 6 00:36:34.444660 kubelet[2004]: I0906 00:36:34.444627 2004 scope.go:117] "RemoveContainer" containerID="44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3" Sep 6 00:36:34.445066 env[1222]: time="2025-09-06T00:36:34.444958065Z" level=error msg="ContainerStatus for \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\": not found" Sep 6 00:36:34.445259 kubelet[2004]: E0906 00:36:34.445222 2004 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\": not found" containerID="44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3" Sep 6 00:36:34.445380 kubelet[2004]: I0906 00:36:34.445271 2004 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3"} err="failed to get container status \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\": rpc error: code = NotFound desc = an error occurred when try to find container \"44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3\": not found" Sep 6 00:36:34.445380 kubelet[2004]: I0906 00:36:34.445307 2004 scope.go:117] "RemoveContainer" containerID="04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716" Sep 6 00:36:34.445852 env[1222]: time="2025-09-06T00:36:34.445749795Z" level=error msg="ContainerStatus for \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\": not found" Sep 6 00:36:34.446056 kubelet[2004]: E0906 00:36:34.446020 2004 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\": not found" containerID="04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716" Sep 6 00:36:34.446173 kubelet[2004]: I0906 00:36:34.446065 2004 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716"} err="failed to get container status \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\": rpc error: code = NotFound desc = an error occurred when try to find container \"04024bd8515a5dd5efe21b615f9f8aaef1c8f8b1f2673598f748e094da412716\": not found" Sep 6 00:36:34.446173 kubelet[2004]: I0906 00:36:34.446095 2004 scope.go:117] "RemoveContainer" containerID="7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8" Sep 6 00:36:34.446532 env[1222]: time="2025-09-06T00:36:34.446438213Z" level=error msg="ContainerStatus for \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\": not found" Sep 6 00:36:34.446673 kubelet[2004]: E0906 00:36:34.446637 2004 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\": not found" containerID="7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8" Sep 6 00:36:34.446775 kubelet[2004]: I0906 00:36:34.446682 2004 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8"} err="failed to get container status \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\": rpc error: code = NotFound desc = an error occurred when try to find container \"7c9a5dc00dcac247e6c0e0c7d4884f34d5929b362d53eab81a7738b37a960ac8\": not found" Sep 6 00:36:34.446775 kubelet[2004]: I0906 00:36:34.446709 2004 scope.go:117] "RemoveContainer" containerID="85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c" Sep 6 00:36:34.447224 env[1222]: time="2025-09-06T00:36:34.447062745Z" level=error msg="ContainerStatus for \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\": not found" Sep 6 00:36:34.447538 kubelet[2004]: E0906 00:36:34.447507 2004 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\": not found" containerID="85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c" Sep 6 00:36:34.447711 kubelet[2004]: I0906 00:36:34.447543 2004 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c"} err="failed to get container status \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\": rpc error: code = NotFound desc = an error occurred when try to find container \"85b2f91f6f97276284d292932581e47b6b1d89e0881d2afd9edaf2eee52d242c\": not found" Sep 6 00:36:34.447711 kubelet[2004]: I0906 00:36:34.447570 2004 scope.go:117] "RemoveContainer" containerID="351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279" Sep 6 00:36:34.448433 env[1222]: time="2025-09-06T00:36:34.448097823Z" level=error msg="ContainerStatus for \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\": not found" Sep 6 00:36:34.448579 kubelet[2004]: E0906 00:36:34.448554 2004 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\": not found" containerID="351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279" Sep 6 00:36:34.448663 kubelet[2004]: I0906 00:36:34.448587 2004 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279"} err="failed to get container status \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\": rpc error: code = NotFound desc = an error occurred when try to find container \"351c080b3add09f59a19f479dee8031ee0eb7818c629b87f448fb11de1186279\": not found" Sep 6 00:36:34.601721 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-44a302ba0dd31734b9817615ab5519dcf04c02f148c279749d12e7bca15789a3-rootfs.mount: Deactivated successfully. Sep 6 00:36:34.601891 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379-rootfs.mount: Deactivated successfully. Sep 6 00:36:34.601999 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-22868236cd82a52c344efc7ccd24f274ea3a559e19af1bf2db4738148de6e379-shm.mount: Deactivated successfully. Sep 6 00:36:34.602097 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6c3875da974ce88b16da113893fcc589b97d5e5ba37a8a7f1b4b7c7ec7290fee-rootfs.mount: Deactivated successfully. Sep 6 00:36:34.602192 systemd[1]: var-lib-kubelet-pods-3532905a\x2d8f04\x2d4418\x2db4ad\x2d6a6a50e8f32b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d9jd8z.mount: Deactivated successfully. Sep 6 00:36:34.602307 systemd[1]: var-lib-kubelet-pods-448ef295\x2d16b6\x2d4f6a\x2d9a49\x2da911a225b515-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d5jtr9.mount: Deactivated successfully. Sep 6 00:36:34.602519 systemd[1]: var-lib-kubelet-pods-3532905a\x2d8f04\x2d4418\x2db4ad\x2d6a6a50e8f32b-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 6 00:36:34.602752 systemd[1]: var-lib-kubelet-pods-3532905a\x2d8f04\x2d4418\x2db4ad\x2d6a6a50e8f32b-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 6 00:36:35.553030 sshd[3562]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:35.560060 systemd[1]: sshd@21-10.128.0.94:22-139.178.89.65:46626.service: Deactivated successfully. Sep 6 00:36:35.561764 systemd[1]: session-22.scope: Deactivated successfully. Sep 6 00:36:35.561799 systemd-logind[1211]: Session 22 logged out. Waiting for processes to exit. Sep 6 00:36:35.563592 systemd-logind[1211]: Removed session 22. Sep 6 00:36:35.600020 systemd[1]: Started sshd@22-10.128.0.94:22-139.178.89.65:46634.service. Sep 6 00:36:35.889496 sshd[3731]: Accepted publickey for core from 139.178.89.65 port 46634 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:35.891573 sshd[3731]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:35.899998 systemd[1]: Started session-23.scope. Sep 6 00:36:35.900686 systemd-logind[1211]: New session 23 of user core. Sep 6 00:36:36.025447 kubelet[2004]: I0906 00:36:36.024599 2004 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" path="/var/lib/kubelet/pods/3532905a-8f04-4418-b4ad-6a6a50e8f32b/volumes" Sep 6 00:36:36.026999 kubelet[2004]: I0906 00:36:36.026944 2004 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="448ef295-16b6-4f6a-9a49-a911a225b515" path="/var/lib/kubelet/pods/448ef295-16b6-4f6a-9a49-a911a225b515/volumes" Sep 6 00:36:37.124984 kubelet[2004]: E0906 00:36:37.124926 2004 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 6 00:36:37.230658 sshd[3731]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:37.238672 systemd-logind[1211]: Session 23 logged out. Waiting for processes to exit. Sep 6 00:36:37.241596 systemd[1]: sshd@22-10.128.0.94:22-139.178.89.65:46634.service: Deactivated successfully. Sep 6 00:36:37.242758 systemd[1]: session-23.scope: Deactivated successfully. Sep 6 00:36:37.242988 systemd[1]: session-23.scope: Consumed 1.073s CPU time. Sep 6 00:36:37.247864 kubelet[2004]: E0906 00:36:37.247822 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="448ef295-16b6-4f6a-9a49-a911a225b515" containerName="cilium-operator" Sep 6 00:36:37.247864 kubelet[2004]: E0906 00:36:37.247863 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" containerName="clean-cilium-state" Sep 6 00:36:37.248102 kubelet[2004]: E0906 00:36:37.247876 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" containerName="cilium-agent" Sep 6 00:36:37.248102 kubelet[2004]: E0906 00:36:37.247887 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" containerName="apply-sysctl-overwrites" Sep 6 00:36:37.248102 kubelet[2004]: E0906 00:36:37.247897 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" containerName="mount-bpf-fs" Sep 6 00:36:37.248102 kubelet[2004]: E0906 00:36:37.247908 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" containerName="mount-cgroup" Sep 6 00:36:37.248102 kubelet[2004]: I0906 00:36:37.247947 2004 memory_manager.go:354] "RemoveStaleState removing state" podUID="448ef295-16b6-4f6a-9a49-a911a225b515" containerName="cilium-operator" Sep 6 00:36:37.248102 kubelet[2004]: I0906 00:36:37.247960 2004 memory_manager.go:354] "RemoveStaleState removing state" podUID="3532905a-8f04-4418-b4ad-6a6a50e8f32b" containerName="cilium-agent" Sep 6 00:36:37.251753 systemd-logind[1211]: Removed session 23. Sep 6 00:36:37.259185 systemd[1]: Created slice kubepods-burstable-podf4dfe5cd_ac72_4c84_89b7_7f963241e978.slice. Sep 6 00:36:37.282858 systemd[1]: Started sshd@23-10.128.0.94:22-139.178.89.65:46638.service. Sep 6 00:36:37.293699 kubelet[2004]: I0906 00:36:37.293646 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hostproc\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.293911 kubelet[2004]: I0906 00:36:37.293719 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-bpf-maps\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.293911 kubelet[2004]: I0906 00:36:37.293747 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-xtables-lock\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.293911 kubelet[2004]: I0906 00:36:37.293793 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-clustermesh-secrets\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.293911 kubelet[2004]: I0906 00:36:37.293819 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb26f\" (UniqueName: \"kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-kube-api-access-bb26f\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.293911 kubelet[2004]: I0906 00:36:37.293867 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-run\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.293911 kubelet[2004]: I0906 00:36:37.293890 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-cgroup\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294255 kubelet[2004]: I0906 00:36:37.293916 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-kernel\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294255 kubelet[2004]: I0906 00:36:37.293965 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-lib-modules\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294255 kubelet[2004]: I0906 00:36:37.293992 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-net\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294255 kubelet[2004]: I0906 00:36:37.294043 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hubble-tls\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294255 kubelet[2004]: I0906 00:36:37.294074 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cni-path\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294255 kubelet[2004]: I0906 00:36:37.294122 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-ipsec-secrets\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294617 kubelet[2004]: I0906 00:36:37.294151 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-etc-cni-netd\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.294617 kubelet[2004]: I0906 00:36:37.294209 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-config-path\") pod \"cilium-rmfmr\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " pod="kube-system/cilium-rmfmr" Sep 6 00:36:37.568440 env[1222]: time="2025-09-06T00:36:37.566764314Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-rmfmr,Uid:f4dfe5cd-ac72-4c84-89b7-7f963241e978,Namespace:kube-system,Attempt:0,}" Sep 6 00:36:37.592757 sshd[3743]: Accepted publickey for core from 139.178.89.65 port 46638 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:37.594372 sshd[3743]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:37.608830 systemd[1]: Started session-24.scope. Sep 6 00:36:37.616158 systemd-logind[1211]: New session 24 of user core. Sep 6 00:36:37.623595 env[1222]: time="2025-09-06T00:36:37.601799832Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:36:37.623595 env[1222]: time="2025-09-06T00:36:37.601855656Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:36:37.623595 env[1222]: time="2025-09-06T00:36:37.601874922Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:36:37.623595 env[1222]: time="2025-09-06T00:36:37.602124013Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd pid=3757 runtime=io.containerd.runc.v2 Sep 6 00:36:37.636629 systemd[1]: Started cri-containerd-7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd.scope. Sep 6 00:36:37.673367 env[1222]: time="2025-09-06T00:36:37.673300972Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-rmfmr,Uid:f4dfe5cd-ac72-4c84-89b7-7f963241e978,Namespace:kube-system,Attempt:0,} returns sandbox id \"7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd\"" Sep 6 00:36:37.679029 env[1222]: time="2025-09-06T00:36:37.678876557Z" level=info msg="CreateContainer within sandbox \"7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 6 00:36:37.694566 env[1222]: time="2025-09-06T00:36:37.694487752Z" level=info msg="CreateContainer within sandbox \"7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\"" Sep 6 00:36:37.697077 env[1222]: time="2025-09-06T00:36:37.696559588Z" level=info msg="StartContainer for \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\"" Sep 6 00:36:37.723205 systemd[1]: Started cri-containerd-51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6.scope. Sep 6 00:36:37.740582 systemd[1]: cri-containerd-51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6.scope: Deactivated successfully. Sep 6 00:36:37.764195 env[1222]: time="2025-09-06T00:36:37.762425774Z" level=info msg="shim disconnected" id=51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6 Sep 6 00:36:37.764195 env[1222]: time="2025-09-06T00:36:37.762495706Z" level=warning msg="cleaning up after shim disconnected" id=51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6 namespace=k8s.io Sep 6 00:36:37.764195 env[1222]: time="2025-09-06T00:36:37.762511631Z" level=info msg="cleaning up dead shim" Sep 6 00:36:37.783581 env[1222]: time="2025-09-06T00:36:37.783516027Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:37Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3817 runtime=io.containerd.runc.v2\ntime=\"2025-09-06T00:36:37Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Sep 6 00:36:37.784280 env[1222]: time="2025-09-06T00:36:37.784139847Z" level=error msg="copy shim log" error="read /proc/self/fd/30: file already closed" Sep 6 00:36:37.785558 env[1222]: time="2025-09-06T00:36:37.785498549Z" level=error msg="Failed to pipe stdout of container \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\"" error="reading from a closed fifo" Sep 6 00:36:37.786089 env[1222]: time="2025-09-06T00:36:37.785737155Z" level=error msg="Failed to pipe stderr of container \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\"" error="reading from a closed fifo" Sep 6 00:36:37.789190 env[1222]: time="2025-09-06T00:36:37.789109627Z" level=error msg="StartContainer for \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Sep 6 00:36:37.789916 kubelet[2004]: E0906 00:36:37.789645 2004 log.go:32] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6" Sep 6 00:36:37.789916 kubelet[2004]: E0906 00:36:37.789865 2004 kuberuntime_manager.go:1274] "Unhandled Error" err=< Sep 6 00:36:37.789916 kubelet[2004]: init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Sep 6 00:36:37.789916 kubelet[2004]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Sep 6 00:36:37.789916 kubelet[2004]: rm /hostbin/cilium-mount Sep 6 00:36:37.790705 kubelet[2004]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bb26f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:&AppArmorProfile{Type:Unconfined,LocalhostProfile:nil,},},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-rmfmr_kube-system(f4dfe5cd-ac72-4c84-89b7-7f963241e978): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Sep 6 00:36:37.790705 kubelet[2004]: > logger="UnhandledError" Sep 6 00:36:37.792089 kubelet[2004]: E0906 00:36:37.792004 2004 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-rmfmr" podUID="f4dfe5cd-ac72-4c84-89b7-7f963241e978" Sep 6 00:36:37.922453 sshd[3743]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:37.928639 systemd[1]: sshd@23-10.128.0.94:22-139.178.89.65:46638.service: Deactivated successfully. Sep 6 00:36:37.929718 systemd[1]: session-24.scope: Deactivated successfully. Sep 6 00:36:37.930627 systemd-logind[1211]: Session 24 logged out. Waiting for processes to exit. Sep 6 00:36:37.932026 systemd-logind[1211]: Removed session 24. Sep 6 00:36:37.969615 systemd[1]: Started sshd@24-10.128.0.94:22-139.178.89.65:46654.service. Sep 6 00:36:38.265256 sshd[3838]: Accepted publickey for core from 139.178.89.65 port 46654 ssh2: RSA SHA256:O4b1lx+UphQ1XQCPwsrjL8IoqrnWSgynNYcpg4eKVRo Sep 6 00:36:38.267981 sshd[3838]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Sep 6 00:36:38.277586 systemd[1]: Started session-25.scope. Sep 6 00:36:38.279478 systemd-logind[1211]: New session 25 of user core. Sep 6 00:36:38.434606 env[1222]: time="2025-09-06T00:36:38.429949855Z" level=info msg="StopPodSandbox for \"7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd\"" Sep 6 00:36:38.434606 env[1222]: time="2025-09-06T00:36:38.430036339Z" level=info msg="Container to stop \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Sep 6 00:36:38.433532 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd-shm.mount: Deactivated successfully. Sep 6 00:36:38.460156 systemd[1]: cri-containerd-7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd.scope: Deactivated successfully. Sep 6 00:36:38.509624 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd-rootfs.mount: Deactivated successfully. Sep 6 00:36:38.519647 env[1222]: time="2025-09-06T00:36:38.519516432Z" level=info msg="shim disconnected" id=7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd Sep 6 00:36:38.521530 env[1222]: time="2025-09-06T00:36:38.521479611Z" level=warning msg="cleaning up after shim disconnected" id=7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd namespace=k8s.io Sep 6 00:36:38.522453 env[1222]: time="2025-09-06T00:36:38.522420971Z" level=info msg="cleaning up dead shim" Sep 6 00:36:38.537856 env[1222]: time="2025-09-06T00:36:38.537792205Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:38Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3864 runtime=io.containerd.runc.v2\n" Sep 6 00:36:38.538518 env[1222]: time="2025-09-06T00:36:38.538481879Z" level=info msg="TearDown network for sandbox \"7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd\" successfully" Sep 6 00:36:38.538687 env[1222]: time="2025-09-06T00:36:38.538660911Z" level=info msg="StopPodSandbox for \"7b5652fdac6423f6ee752622f0c4694a5ad34828a383a7bc334c2abf65ca25dd\" returns successfully" Sep 6 00:36:38.604887 kubelet[2004]: I0906 00:36:38.604830 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-etc-cni-netd\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.605824 kubelet[2004]: I0906 00:36:38.605794 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hostproc\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.606057 kubelet[2004]: I0906 00:36:38.606022 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-kernel\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.606241 kubelet[2004]: I0906 00:36:38.606220 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-config-path\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.606416 kubelet[2004]: I0906 00:36:38.606380 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-run\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.606580 kubelet[2004]: I0906 00:36:38.606560 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-clustermesh-secrets\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.606733 kubelet[2004]: I0906 00:36:38.606713 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-bpf-maps\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.606894 kubelet[2004]: I0906 00:36:38.606869 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-xtables-lock\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.607049 kubelet[2004]: I0906 00:36:38.607028 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cni-path\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.607228 kubelet[2004]: I0906 00:36:38.607207 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb26f\" (UniqueName: \"kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-kube-api-access-bb26f\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.607387 kubelet[2004]: I0906 00:36:38.607360 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-lib-modules\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.607638 kubelet[2004]: I0906 00:36:38.607613 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hubble-tls\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.607851 kubelet[2004]: I0906 00:36:38.607814 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-cgroup\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.608013 kubelet[2004]: I0906 00:36:38.607990 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-net\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.608232 kubelet[2004]: I0906 00:36:38.608198 2004 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-ipsec-secrets\") pod \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\" (UID: \"f4dfe5cd-ac72-4c84-89b7-7f963241e978\") " Sep 6 00:36:38.608626 kubelet[2004]: I0906 00:36:38.608474 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.611510 kubelet[2004]: I0906 00:36:38.608787 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hostproc" (OuterVolumeSpecName: "hostproc") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.611510 kubelet[2004]: I0906 00:36:38.608812 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.611510 kubelet[2004]: I0906 00:36:38.611429 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Sep 6 00:36:38.611755 kubelet[2004]: I0906 00:36:38.611484 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.612310 kubelet[2004]: I0906 00:36:38.612285 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.612501 kubelet[2004]: I0906 00:36:38.612481 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.612631 kubelet[2004]: I0906 00:36:38.612615 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cni-path" (OuterVolumeSpecName: "cni-path") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.620452 systemd[1]: var-lib-kubelet-pods-f4dfe5cd\x2dac72\x2d4c84\x2d89b7\x2d7f963241e978-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Sep 6 00:36:38.622666 kubelet[2004]: I0906 00:36:38.621849 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.622666 kubelet[2004]: I0906 00:36:38.621914 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.622666 kubelet[2004]: I0906 00:36:38.622443 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 6 00:36:38.622932 kubelet[2004]: I0906 00:36:38.622626 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Sep 6 00:36:38.626478 systemd[1]: var-lib-kubelet-pods-f4dfe5cd\x2dac72\x2d4c84\x2d89b7\x2d7f963241e978-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbb26f.mount: Deactivated successfully. Sep 6 00:36:38.629826 kubelet[2004]: I0906 00:36:38.629784 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-kube-api-access-bb26f" (OuterVolumeSpecName: "kube-api-access-bb26f") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "kube-api-access-bb26f". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 6 00:36:38.631873 kubelet[2004]: I0906 00:36:38.631833 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Sep 6 00:36:38.633244 kubelet[2004]: I0906 00:36:38.633208 2004 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "f4dfe5cd-ac72-4c84-89b7-7f963241e978" (UID: "f4dfe5cd-ac72-4c84-89b7-7f963241e978"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Sep 6 00:36:38.709747 kubelet[2004]: I0906 00:36:38.709687 2004 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb26f\" (UniqueName: \"kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-kube-api-access-bb26f\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.709747 kubelet[2004]: I0906 00:36:38.709745 2004 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-lib-modules\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.709747 kubelet[2004]: I0906 00:36:38.709766 2004 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hubble-tls\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709782 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-cgroup\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709798 2004 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-net\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709817 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-ipsec-secrets\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709832 2004 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-etc-cni-netd\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709845 2004 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-hostproc\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709861 2004 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-host-proc-sys-kernel\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710145 kubelet[2004]: I0906 00:36:38.709879 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-config-path\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710423 kubelet[2004]: I0906 00:36:38.709896 2004 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cilium-run\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710423 kubelet[2004]: I0906 00:36:38.709910 2004 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f4dfe5cd-ac72-4c84-89b7-7f963241e978-clustermesh-secrets\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710423 kubelet[2004]: I0906 00:36:38.709925 2004 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-xtables-lock\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710423 kubelet[2004]: I0906 00:36:38.709942 2004 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-cni-path\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:38.710423 kubelet[2004]: I0906 00:36:38.709957 2004 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f4dfe5cd-ac72-4c84-89b7-7f963241e978-bpf-maps\") on node \"ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" DevicePath \"\"" Sep 6 00:36:39.409857 systemd[1]: var-lib-kubelet-pods-f4dfe5cd\x2dac72\x2d4c84\x2d89b7\x2d7f963241e978-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Sep 6 00:36:39.410028 systemd[1]: var-lib-kubelet-pods-f4dfe5cd\x2dac72\x2d4c84\x2d89b7\x2d7f963241e978-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Sep 6 00:36:39.432626 kubelet[2004]: I0906 00:36:39.432589 2004 scope.go:117] "RemoveContainer" containerID="51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6" Sep 6 00:36:39.438963 systemd[1]: Removed slice kubepods-burstable-podf4dfe5cd_ac72_4c84_89b7_7f963241e978.slice. Sep 6 00:36:39.440488 env[1222]: time="2025-09-06T00:36:39.439995539Z" level=info msg="RemoveContainer for \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\"" Sep 6 00:36:39.444722 env[1222]: time="2025-09-06T00:36:39.444665674Z" level=info msg="RemoveContainer for \"51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6\" returns successfully" Sep 6 00:36:39.538255 kubelet[2004]: E0906 00:36:39.538188 2004 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f4dfe5cd-ac72-4c84-89b7-7f963241e978" containerName="mount-cgroup" Sep 6 00:36:39.538631 kubelet[2004]: I0906 00:36:39.538587 2004 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4dfe5cd-ac72-4c84-89b7-7f963241e978" containerName="mount-cgroup" Sep 6 00:36:39.548241 systemd[1]: Created slice kubepods-burstable-pod464b2206_1322_46bc_a8c7_380ca10d6599.slice. Sep 6 00:36:39.561740 kubelet[2004]: W0906 00:36:39.561690 2004 reflector.go:561] object-"kube-system"/"cilium-clustermesh": failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object Sep 6 00:36:39.562104 kubelet[2004]: E0906 00:36:39.562050 2004 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-clustermesh\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-clustermesh\" is forbidden: User \"system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object" logger="UnhandledError" Sep 6 00:36:39.562785 kubelet[2004]: W0906 00:36:39.562760 2004 reflector.go:561] object-"kube-system"/"cilium-ipsec-keys": failed to list *v1.Secret: secrets "cilium-ipsec-keys" is forbidden: User "system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object Sep 6 00:36:39.562991 kubelet[2004]: E0906 00:36:39.562961 2004 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-ipsec-keys\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cilium-ipsec-keys\" is forbidden: User \"system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object" logger="UnhandledError" Sep 6 00:36:39.563495 kubelet[2004]: W0906 00:36:39.563471 2004 reflector.go:561] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object Sep 6 00:36:39.563670 kubelet[2004]: E0906 00:36:39.563643 2004 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"hubble-server-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"hubble-server-certs\" is forbidden: User \"system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object" logger="UnhandledError" Sep 6 00:36:39.564121 kubelet[2004]: W0906 00:36:39.564097 2004 reflector.go:561] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object Sep 6 00:36:39.564309 kubelet[2004]: E0906 00:36:39.564283 2004 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cilium-config\" is forbidden: User \"system:node:ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6' and this object" logger="UnhandledError" Sep 6 00:36:39.616630 kubelet[2004]: I0906 00:36:39.616568 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-cgroup\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.617366 kubelet[2004]: I0906 00:36:39.617315 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-config-path\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.617575 kubelet[2004]: I0906 00:36:39.617540 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqgv\" (UniqueName: \"kubernetes.io/projected/464b2206-1322-46bc-a8c7-380ca10d6599-kube-api-access-nrqgv\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.617825 kubelet[2004]: I0906 00:36:39.617802 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-ipsec-secrets\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.618015 kubelet[2004]: I0906 00:36:39.617991 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-host-proc-sys-kernel\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.618199 kubelet[2004]: I0906 00:36:39.618173 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-hostproc\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.618374 kubelet[2004]: I0906 00:36:39.618351 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/464b2206-1322-46bc-a8c7-380ca10d6599-clustermesh-secrets\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.618553 kubelet[2004]: I0906 00:36:39.618528 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-xtables-lock\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.618714 kubelet[2004]: I0906 00:36:39.618695 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-etc-cni-netd\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.618942 kubelet[2004]: I0906 00:36:39.618920 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-lib-modules\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.619139 kubelet[2004]: I0906 00:36:39.619079 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/464b2206-1322-46bc-a8c7-380ca10d6599-hubble-tls\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.619295 kubelet[2004]: I0906 00:36:39.619275 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-run\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.619457 kubelet[2004]: I0906 00:36:39.619436 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-bpf-maps\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.619615 kubelet[2004]: I0906 00:36:39.619595 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-cni-path\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:39.619773 kubelet[2004]: I0906 00:36:39.619750 2004 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/464b2206-1322-46bc-a8c7-380ca10d6599-host-proc-sys-net\") pod \"cilium-h49vl\" (UID: \"464b2206-1322-46bc-a8c7-380ca10d6599\") " pod="kube-system/cilium-h49vl" Sep 6 00:36:40.024260 kubelet[2004]: I0906 00:36:40.024193 2004 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4dfe5cd-ac72-4c84-89b7-7f963241e978" path="/var/lib/kubelet/pods/f4dfe5cd-ac72-4c84-89b7-7f963241e978/volumes" Sep 6 00:36:40.720991 kubelet[2004]: E0906 00:36:40.720928 2004 configmap.go:193] Couldn't get configMap kube-system/cilium-config: failed to sync configmap cache: timed out waiting for the condition Sep 6 00:36:40.721682 kubelet[2004]: E0906 00:36:40.721083 2004 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-config-path podName:464b2206-1322-46bc-a8c7-380ca10d6599 nodeName:}" failed. No retries permitted until 2025-09-06 00:36:41.221052739 +0000 UTC m=+129.374738292 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-config-path" (UniqueName: "kubernetes.io/configmap/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-config-path") pod "cilium-h49vl" (UID: "464b2206-1322-46bc-a8c7-380ca10d6599") : failed to sync configmap cache: timed out waiting for the condition Sep 6 00:36:40.721682 kubelet[2004]: E0906 00:36:40.721124 2004 secret.go:189] Couldn't get secret kube-system/cilium-ipsec-keys: failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.721682 kubelet[2004]: E0906 00:36:40.721179 2004 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-ipsec-secrets podName:464b2206-1322-46bc-a8c7-380ca10d6599 nodeName:}" failed. No retries permitted until 2025-09-06 00:36:41.221160106 +0000 UTC m=+129.374845660 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-ipsec-secrets" (UniqueName: "kubernetes.io/secret/464b2206-1322-46bc-a8c7-380ca10d6599-cilium-ipsec-secrets") pod "cilium-h49vl" (UID: "464b2206-1322-46bc-a8c7-380ca10d6599") : failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.722010 kubelet[2004]: E0906 00:36:40.721985 2004 projected.go:263] Couldn't get secret kube-system/hubble-server-certs: failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.722125 kubelet[2004]: E0906 00:36:40.722107 2004 projected.go:194] Error preparing data for projected volume hubble-tls for pod kube-system/cilium-h49vl: failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.722312 kubelet[2004]: E0906 00:36:40.722265 2004 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/464b2206-1322-46bc-a8c7-380ca10d6599-hubble-tls podName:464b2206-1322-46bc-a8c7-380ca10d6599 nodeName:}" failed. No retries permitted until 2025-09-06 00:36:41.222242809 +0000 UTC m=+129.375928378 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "hubble-tls" (UniqueName: "kubernetes.io/projected/464b2206-1322-46bc-a8c7-380ca10d6599-hubble-tls") pod "cilium-h49vl" (UID: "464b2206-1322-46bc-a8c7-380ca10d6599") : failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.722312 kubelet[2004]: E0906 00:36:40.721986 2004 secret.go:189] Couldn't get secret kube-system/cilium-clustermesh: failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.722560 kubelet[2004]: E0906 00:36:40.722328 2004 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/464b2206-1322-46bc-a8c7-380ca10d6599-clustermesh-secrets podName:464b2206-1322-46bc-a8c7-380ca10d6599 nodeName:}" failed. No retries permitted until 2025-09-06 00:36:41.222315318 +0000 UTC m=+129.376000887 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "clustermesh-secrets" (UniqueName: "kubernetes.io/secret/464b2206-1322-46bc-a8c7-380ca10d6599-clustermesh-secrets") pod "cilium-h49vl" (UID: "464b2206-1322-46bc-a8c7-380ca10d6599") : failed to sync secret cache: timed out waiting for the condition Sep 6 00:36:40.869481 kubelet[2004]: W0906 00:36:40.869385 2004 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4dfe5cd_ac72_4c84_89b7_7f963241e978.slice/cri-containerd-51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6.scope WatchSource:0}: container "51da126dd3df598d74b7bab017bfcbfdd09d6da8f8f49c19774b18289a83d1c6" in namespace "k8s.io": not found Sep 6 00:36:41.353959 env[1222]: time="2025-09-06T00:36:41.353899959Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-h49vl,Uid:464b2206-1322-46bc-a8c7-380ca10d6599,Namespace:kube-system,Attempt:0,}" Sep 6 00:36:41.380535 env[1222]: time="2025-09-06T00:36:41.380390401Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Sep 6 00:36:41.380827 env[1222]: time="2025-09-06T00:36:41.380482835Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Sep 6 00:36:41.380827 env[1222]: time="2025-09-06T00:36:41.380802815Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Sep 6 00:36:41.381480 env[1222]: time="2025-09-06T00:36:41.381327927Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9 pid=3897 runtime=io.containerd.runc.v2 Sep 6 00:36:41.414844 systemd[1]: Started cri-containerd-eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9.scope. Sep 6 00:36:41.451111 env[1222]: time="2025-09-06T00:36:41.451042441Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-h49vl,Uid:464b2206-1322-46bc-a8c7-380ca10d6599,Namespace:kube-system,Attempt:0,} returns sandbox id \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\"" Sep 6 00:36:41.458232 env[1222]: time="2025-09-06T00:36:41.458160475Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Sep 6 00:36:41.472666 env[1222]: time="2025-09-06T00:36:41.472577313Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e\"" Sep 6 00:36:41.473466 env[1222]: time="2025-09-06T00:36:41.473377430Z" level=info msg="StartContainer for \"081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e\"" Sep 6 00:36:41.499995 systemd[1]: Started cri-containerd-081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e.scope. Sep 6 00:36:41.541940 env[1222]: time="2025-09-06T00:36:41.541880835Z" level=info msg="StartContainer for \"081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e\" returns successfully" Sep 6 00:36:41.555845 systemd[1]: cri-containerd-081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e.scope: Deactivated successfully. Sep 6 00:36:41.596299 env[1222]: time="2025-09-06T00:36:41.596217934Z" level=info msg="shim disconnected" id=081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e Sep 6 00:36:41.596299 env[1222]: time="2025-09-06T00:36:41.596297410Z" level=warning msg="cleaning up after shim disconnected" id=081a8effa7831b8fb4eadc07ec8b145a17aa6ea3506145ba07a674398b9f545e namespace=k8s.io Sep 6 00:36:41.596739 env[1222]: time="2025-09-06T00:36:41.596314722Z" level=info msg="cleaning up dead shim" Sep 6 00:36:41.609035 env[1222]: time="2025-09-06T00:36:41.608870905Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:41Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3980 runtime=io.containerd.runc.v2\n" Sep 6 00:36:42.127294 kubelet[2004]: E0906 00:36:42.127219 2004 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Sep 6 00:36:42.449092 env[1222]: time="2025-09-06T00:36:42.448676631Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Sep 6 00:36:42.469733 env[1222]: time="2025-09-06T00:36:42.469673818Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498\"" Sep 6 00:36:42.473554 env[1222]: time="2025-09-06T00:36:42.473508916Z" level=info msg="StartContainer for \"37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498\"" Sep 6 00:36:42.525078 systemd[1]: Started cri-containerd-37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498.scope. Sep 6 00:36:42.590306 env[1222]: time="2025-09-06T00:36:42.589388619Z" level=info msg="StartContainer for \"37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498\" returns successfully" Sep 6 00:36:42.597320 systemd[1]: cri-containerd-37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498.scope: Deactivated successfully. Sep 6 00:36:42.632023 env[1222]: time="2025-09-06T00:36:42.631951799Z" level=info msg="shim disconnected" id=37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498 Sep 6 00:36:42.632023 env[1222]: time="2025-09-06T00:36:42.632023146Z" level=warning msg="cleaning up after shim disconnected" id=37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498 namespace=k8s.io Sep 6 00:36:42.632383 env[1222]: time="2025-09-06T00:36:42.632039289Z" level=info msg="cleaning up dead shim" Sep 6 00:36:42.647928 env[1222]: time="2025-09-06T00:36:42.647852001Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:42Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4046 runtime=io.containerd.runc.v2\n" Sep 6 00:36:43.248952 systemd[1]: run-containerd-runc-k8s.io-37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498-runc.fWc7Dw.mount: Deactivated successfully. Sep 6 00:36:43.249104 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-37d9cc53d38de3cf69d88d7c417466ce4578ccca744ddc26b2ec355ebcf68498-rootfs.mount: Deactivated successfully. Sep 6 00:36:43.452288 env[1222]: time="2025-09-06T00:36:43.452204659Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Sep 6 00:36:43.494966 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2404033184.mount: Deactivated successfully. Sep 6 00:36:43.501752 env[1222]: time="2025-09-06T00:36:43.501459295Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a\"" Sep 6 00:36:43.503304 env[1222]: time="2025-09-06T00:36:43.503262353Z" level=info msg="StartContainer for \"9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a\"" Sep 6 00:36:43.544071 systemd[1]: Started cri-containerd-9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a.scope. Sep 6 00:36:43.596354 env[1222]: time="2025-09-06T00:36:43.596292516Z" level=info msg="StartContainer for \"9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a\" returns successfully" Sep 6 00:36:43.602590 systemd[1]: cri-containerd-9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a.scope: Deactivated successfully. Sep 6 00:36:43.636687 env[1222]: time="2025-09-06T00:36:43.636627043Z" level=info msg="shim disconnected" id=9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a Sep 6 00:36:43.637342 env[1222]: time="2025-09-06T00:36:43.637249750Z" level=warning msg="cleaning up after shim disconnected" id=9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a namespace=k8s.io Sep 6 00:36:43.637342 env[1222]: time="2025-09-06T00:36:43.637331891Z" level=info msg="cleaning up dead shim" Sep 6 00:36:43.650680 env[1222]: time="2025-09-06T00:36:43.650611236Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:43Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4104 runtime=io.containerd.runc.v2\n" Sep 6 00:36:44.248964 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9fbc6e6f8a70dd93294e081489bf04e8179eb98404e448eaa6453eefb724406a-rootfs.mount: Deactivated successfully. Sep 6 00:36:44.460894 env[1222]: time="2025-09-06T00:36:44.460830273Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Sep 6 00:36:44.489897 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2495851450.mount: Deactivated successfully. Sep 6 00:36:44.496684 env[1222]: time="2025-09-06T00:36:44.496506597Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb\"" Sep 6 00:36:44.500097 env[1222]: time="2025-09-06T00:36:44.497640223Z" level=info msg="StartContainer for \"47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb\"" Sep 6 00:36:44.534925 systemd[1]: Started cri-containerd-47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb.scope. Sep 6 00:36:44.582041 systemd[1]: cri-containerd-47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb.scope: Deactivated successfully. Sep 6 00:36:44.584819 env[1222]: time="2025-09-06T00:36:44.584769761Z" level=info msg="StartContainer for \"47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb\" returns successfully" Sep 6 00:36:44.615881 env[1222]: time="2025-09-06T00:36:44.615799570Z" level=info msg="shim disconnected" id=47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb Sep 6 00:36:44.615881 env[1222]: time="2025-09-06T00:36:44.615883110Z" level=warning msg="cleaning up after shim disconnected" id=47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb namespace=k8s.io Sep 6 00:36:44.616314 env[1222]: time="2025-09-06T00:36:44.615900243Z" level=info msg="cleaning up dead shim" Sep 6 00:36:44.629285 env[1222]: time="2025-09-06T00:36:44.629216073Z" level=warning msg="cleanup warnings time=\"2025-09-06T00:36:44Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4160 runtime=io.containerd.runc.v2\n" Sep 6 00:36:45.249092 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-47d636ae89f2829502791ac236adee3eef2164ce0477bb8365bff0c197b6daeb-rootfs.mount: Deactivated successfully. Sep 6 00:36:45.443827 kubelet[2004]: I0906 00:36:45.443759 2004 setters.go:600] "Node became not ready" node="ci-3510-3-8-nightly-20250905-2100-d06d46191def2efd1fb6" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-09-06T00:36:45Z","lastTransitionTime":"2025-09-06T00:36:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Sep 6 00:36:45.464231 env[1222]: time="2025-09-06T00:36:45.464179643Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Sep 6 00:36:45.495511 env[1222]: time="2025-09-06T00:36:45.495352247Z" level=info msg="CreateContainer within sandbox \"eeafde551030cdf37f1385eb8562b86251ddcb913f6448a674450264e164f7d9\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4\"" Sep 6 00:36:45.496866 env[1222]: time="2025-09-06T00:36:45.496819467Z" level=info msg="StartContainer for \"b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4\"" Sep 6 00:36:45.549661 systemd[1]: Started cri-containerd-b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4.scope. Sep 6 00:36:45.601440 env[1222]: time="2025-09-06T00:36:45.601049078Z" level=info msg="StartContainer for \"b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4\" returns successfully" Sep 6 00:36:46.105469 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Sep 6 00:36:46.505101 kubelet[2004]: I0906 00:36:46.505005 2004 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-h49vl" podStartSLOduration=7.504977488 podStartE2EDuration="7.504977488s" podCreationTimestamp="2025-09-06 00:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-09-06 00:36:46.504830741 +0000 UTC m=+134.658516321" watchObservedRunningTime="2025-09-06 00:36:46.504977488 +0000 UTC m=+134.658663070" Sep 6 00:36:46.932820 systemd[1]: run-containerd-runc-k8s.io-b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4-runc.MXkrhj.mount: Deactivated successfully. Sep 6 00:36:49.309579 systemd[1]: run-containerd-runc-k8s.io-b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4-runc.m7C0fg.mount: Deactivated successfully. Sep 6 00:36:49.510573 systemd-networkd[1017]: lxc_health: Link UP Sep 6 00:36:49.520540 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Sep 6 00:36:49.523536 systemd-networkd[1017]: lxc_health: Gained carrier Sep 6 00:36:50.910237 systemd-networkd[1017]: lxc_health: Gained IPv6LL Sep 6 00:36:51.607847 systemd[1]: run-containerd-runc-k8s.io-b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4-runc.4iDtFA.mount: Deactivated successfully. Sep 6 00:36:53.893074 systemd[1]: run-containerd-runc-k8s.io-b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4-runc.fYfZrg.mount: Deactivated successfully. Sep 6 00:36:56.085362 systemd[1]: run-containerd-runc-k8s.io-b142932728ea14e0b41f5de8a16c8791cf1b596c9740dc0968cc3c61a04799c4-runc.0RRYs5.mount: Deactivated successfully. Sep 6 00:36:56.267257 sshd[3838]: pam_unix(sshd:session): session closed for user core Sep 6 00:36:56.272377 systemd[1]: sshd@24-10.128.0.94:22-139.178.89.65:46654.service: Deactivated successfully. Sep 6 00:36:56.273560 systemd[1]: session-25.scope: Deactivated successfully. Sep 6 00:36:56.274574 systemd-logind[1211]: Session 25 logged out. Waiting for processes to exit. Sep 6 00:36:56.275879 systemd-logind[1211]: Removed session 25.