May 17 00:42:26.166913 kernel: Linux version 5.15.182-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri May 16 23:09:52 -00 2025 May 17 00:42:26.166955 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=4aad7caeadb0359f379975532748a0b4ae6bb9b229507353e0f5ae84cb9335a0 May 17 00:42:26.166975 kernel: BIOS-provided physical RAM map: May 17 00:42:26.166990 kernel: BIOS-e820: [mem 0x0000000000000000-0x0000000000000fff] reserved May 17 00:42:26.167004 kernel: BIOS-e820: [mem 0x0000000000001000-0x0000000000054fff] usable May 17 00:42:26.167018 kernel: BIOS-e820: [mem 0x0000000000055000-0x000000000005ffff] reserved May 17 00:42:26.167040 kernel: BIOS-e820: [mem 0x0000000000060000-0x0000000000097fff] usable May 17 00:42:26.167055 kernel: BIOS-e820: [mem 0x0000000000098000-0x000000000009ffff] reserved May 17 00:42:26.167070 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000bd277fff] usable May 17 00:42:26.167085 kernel: BIOS-e820: [mem 0x00000000bd278000-0x00000000bd281fff] ACPI data May 17 00:42:26.167100 kernel: BIOS-e820: [mem 0x00000000bd282000-0x00000000bf8ecfff] usable May 17 00:42:26.167115 kernel: BIOS-e820: [mem 0x00000000bf8ed000-0x00000000bfb6cfff] reserved May 17 00:42:26.167131 kernel: BIOS-e820: [mem 0x00000000bfb6d000-0x00000000bfb7efff] ACPI data May 17 00:42:26.167146 kernel: BIOS-e820: [mem 0x00000000bfb7f000-0x00000000bfbfefff] ACPI NVS May 17 00:42:26.167169 kernel: BIOS-e820: [mem 0x00000000bfbff000-0x00000000bffdffff] usable May 17 00:42:26.167186 kernel: BIOS-e820: [mem 0x00000000bffe0000-0x00000000bfffffff] reserved May 17 00:42:26.167202 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000021fffffff] usable May 17 00:42:26.167218 kernel: NX (Execute Disable) protection: active May 17 00:42:26.167234 kernel: efi: EFI v2.70 by EDK II May 17 00:42:26.167275 kernel: efi: TPMFinalLog=0xbfbf7000 ACPI=0xbfb7e000 ACPI 2.0=0xbfb7e014 SMBIOS=0xbf9e8000 RNG=0xbfb73018 TPMEventLog=0xbd278018 May 17 00:42:26.167295 kernel: random: crng init done May 17 00:42:26.167314 kernel: SMBIOS 2.4 present. May 17 00:42:26.167351 kernel: DMI: Google Google Compute Engine/Google Compute Engine, BIOS Google 02/12/2025 May 17 00:42:26.167369 kernel: Hypervisor detected: KVM May 17 00:42:26.167387 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 May 17 00:42:26.167407 kernel: kvm-clock: cpu 0, msr 1619a001, primary cpu clock May 17 00:42:26.167426 kernel: kvm-clock: using sched offset of 13896487916 cycles May 17 00:42:26.167445 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns May 17 00:42:26.167461 kernel: tsc: Detected 2299.998 MHz processor May 17 00:42:26.167476 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved May 17 00:42:26.167494 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable May 17 00:42:26.167509 kernel: last_pfn = 0x220000 max_arch_pfn = 0x400000000 May 17 00:42:26.167531 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT May 17 00:42:26.167551 kernel: last_pfn = 0xbffe0 max_arch_pfn = 0x400000000 May 17 00:42:26.167571 kernel: Using GB pages for direct mapping May 17 00:42:26.167589 kernel: Secure boot disabled May 17 00:42:26.167608 kernel: ACPI: Early table checksum verification disabled May 17 00:42:26.167625 kernel: ACPI: RSDP 0x00000000BFB7E014 000024 (v02 Google) May 17 00:42:26.167645 kernel: ACPI: XSDT 0x00000000BFB7D0E8 00005C (v01 Google GOOGFACP 00000001 01000013) May 17 00:42:26.167665 kernel: ACPI: FACP 0x00000000BFB78000 0000F4 (v02 Google GOOGFACP 00000001 GOOG 00000001) May 17 00:42:26.167695 kernel: ACPI: DSDT 0x00000000BFB79000 001A64 (v01 Google GOOGDSDT 00000001 GOOG 00000001) May 17 00:42:26.167716 kernel: ACPI: FACS 0x00000000BFBF2000 000040 May 17 00:42:26.167736 kernel: ACPI: SSDT 0x00000000BFB7C000 000316 (v02 GOOGLE Tpm2Tabl 00001000 INTL 20240322) May 17 00:42:26.167755 kernel: ACPI: TPM2 0x00000000BFB7B000 000034 (v04 GOOGLE 00000001 GOOG 00000001) May 17 00:42:26.167777 kernel: ACPI: SRAT 0x00000000BFB77000 0000C8 (v03 Google GOOGSRAT 00000001 GOOG 00000001) May 17 00:42:26.167796 kernel: ACPI: APIC 0x00000000BFB76000 000076 (v05 Google GOOGAPIC 00000001 GOOG 00000001) May 17 00:42:26.167821 kernel: ACPI: SSDT 0x00000000BFB75000 000980 (v01 Google GOOGSSDT 00000001 GOOG 00000001) May 17 00:42:26.167841 kernel: ACPI: WAET 0x00000000BFB74000 000028 (v01 Google GOOGWAET 00000001 GOOG 00000001) May 17 00:42:26.167860 kernel: ACPI: Reserving FACP table memory at [mem 0xbfb78000-0xbfb780f3] May 17 00:42:26.167880 kernel: ACPI: Reserving DSDT table memory at [mem 0xbfb79000-0xbfb7aa63] May 17 00:42:26.167901 kernel: ACPI: Reserving FACS table memory at [mem 0xbfbf2000-0xbfbf203f] May 17 00:42:26.167922 kernel: ACPI: Reserving SSDT table memory at [mem 0xbfb7c000-0xbfb7c315] May 17 00:42:26.167938 kernel: ACPI: Reserving TPM2 table memory at [mem 0xbfb7b000-0xbfb7b033] May 17 00:42:26.167958 kernel: ACPI: Reserving SRAT table memory at [mem 0xbfb77000-0xbfb770c7] May 17 00:42:26.167979 kernel: ACPI: Reserving APIC table memory at [mem 0xbfb76000-0xbfb76075] May 17 00:42:26.168003 kernel: ACPI: Reserving SSDT table memory at [mem 0xbfb75000-0xbfb7597f] May 17 00:42:26.168023 kernel: ACPI: Reserving WAET table memory at [mem 0xbfb74000-0xbfb74027] May 17 00:42:26.168043 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 May 17 00:42:26.168062 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 May 17 00:42:26.168083 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] May 17 00:42:26.168102 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0xbfffffff] May 17 00:42:26.168122 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x21fffffff] May 17 00:42:26.168143 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0xbfffffff] -> [mem 0x00000000-0xbfffffff] May 17 00:42:26.168164 kernel: NUMA: Node 0 [mem 0x00000000-0xbfffffff] + [mem 0x100000000-0x21fffffff] -> [mem 0x00000000-0x21fffffff] May 17 00:42:26.168188 kernel: NODE_DATA(0) allocated [mem 0x21fff8000-0x21fffdfff] May 17 00:42:26.168208 kernel: Zone ranges: May 17 00:42:26.168228 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] May 17 00:42:26.168264 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] May 17 00:42:26.168286 kernel: Normal [mem 0x0000000100000000-0x000000021fffffff] May 17 00:42:26.168306 kernel: Movable zone start for each node May 17 00:42:26.168326 kernel: Early memory node ranges May 17 00:42:26.168352 kernel: node 0: [mem 0x0000000000001000-0x0000000000054fff] May 17 00:42:26.168372 kernel: node 0: [mem 0x0000000000060000-0x0000000000097fff] May 17 00:42:26.168397 kernel: node 0: [mem 0x0000000000100000-0x00000000bd277fff] May 17 00:42:26.168418 kernel: node 0: [mem 0x00000000bd282000-0x00000000bf8ecfff] May 17 00:42:26.168438 kernel: node 0: [mem 0x00000000bfbff000-0x00000000bffdffff] May 17 00:42:26.168459 kernel: node 0: [mem 0x0000000100000000-0x000000021fffffff] May 17 00:42:26.168477 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000021fffffff] May 17 00:42:26.168496 kernel: On node 0, zone DMA: 1 pages in unavailable ranges May 17 00:42:26.168518 kernel: On node 0, zone DMA: 11 pages in unavailable ranges May 17 00:42:26.168537 kernel: On node 0, zone DMA: 104 pages in unavailable ranges May 17 00:42:26.168557 kernel: On node 0, zone DMA32: 10 pages in unavailable ranges May 17 00:42:26.168582 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges May 17 00:42:26.168603 kernel: On node 0, zone Normal: 32 pages in unavailable ranges May 17 00:42:26.168621 kernel: ACPI: PM-Timer IO Port: 0xb008 May 17 00:42:26.168642 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) May 17 00:42:26.168663 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 May 17 00:42:26.168680 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) May 17 00:42:26.168700 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) May 17 00:42:26.168720 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) May 17 00:42:26.168740 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) May 17 00:42:26.168763 kernel: ACPI: Using ACPI (MADT) for SMP configuration information May 17 00:42:26.168784 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs May 17 00:42:26.168804 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices May 17 00:42:26.168824 kernel: Booting paravirtualized kernel on KVM May 17 00:42:26.168843 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns May 17 00:42:26.168863 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 May 17 00:42:26.168884 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 May 17 00:42:26.168905 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 May 17 00:42:26.168924 kernel: pcpu-alloc: [0] 0 1 May 17 00:42:26.168948 kernel: kvm-guest: PV spinlocks enabled May 17 00:42:26.168968 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) May 17 00:42:26.168989 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1932270 May 17 00:42:26.169009 kernel: Policy zone: Normal May 17 00:42:26.169030 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=4aad7caeadb0359f379975532748a0b4ae6bb9b229507353e0f5ae84cb9335a0 May 17 00:42:26.169050 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. May 17 00:42:26.169071 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) May 17 00:42:26.169091 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) May 17 00:42:26.169110 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off May 17 00:42:26.169134 kernel: Memory: 7515404K/7860544K available (12294K kernel code, 2276K rwdata, 13724K rodata, 47472K init, 4108K bss, 344880K reserved, 0K cma-reserved) May 17 00:42:26.169154 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 May 17 00:42:26.169173 kernel: Kernel/User page tables isolation: enabled May 17 00:42:26.169192 kernel: ftrace: allocating 34585 entries in 136 pages May 17 00:42:26.169211 kernel: ftrace: allocated 136 pages with 2 groups May 17 00:42:26.169232 kernel: rcu: Hierarchical RCU implementation. May 17 00:42:26.169266 kernel: rcu: RCU event tracing is enabled. May 17 00:42:26.169287 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. May 17 00:42:26.169311 kernel: Rude variant of Tasks RCU enabled. May 17 00:42:26.169356 kernel: Tracing variant of Tasks RCU enabled. May 17 00:42:26.169378 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. May 17 00:42:26.169402 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 May 17 00:42:26.169423 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 May 17 00:42:26.169444 kernel: Console: colour dummy device 80x25 May 17 00:42:26.169465 kernel: printk: console [ttyS0] enabled May 17 00:42:26.169487 kernel: ACPI: Core revision 20210730 May 17 00:42:26.169507 kernel: APIC: Switch to symmetric I/O mode setup May 17 00:42:26.169529 kernel: x2apic enabled May 17 00:42:26.169554 kernel: Switched APIC routing to physical x2apic. May 17 00:42:26.169576 kernel: ..TIMER: vector=0x30 apic1=0 pin1=0 apic2=-1 pin2=-1 May 17 00:42:26.169598 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x212733415c7, max_idle_ns: 440795236380 ns May 17 00:42:26.169619 kernel: Calibrating delay loop (skipped) preset value.. 4599.99 BogoMIPS (lpj=2299998) May 17 00:42:26.169640 kernel: Last level iTLB entries: 4KB 1024, 2MB 1024, 4MB 1024 May 17 00:42:26.169660 kernel: Last level dTLB entries: 4KB 1024, 2MB 1024, 4MB 1024, 1GB 4 May 17 00:42:26.169683 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization May 17 00:42:26.169710 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit May 17 00:42:26.169742 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall May 17 00:42:26.169765 kernel: Spectre V2 : Mitigation: IBRS May 17 00:42:26.169788 kernel: Spectre V2 : Spectre v2 / SpectreRSB: Filling RSB on context switch and VMEXIT May 17 00:42:26.169816 kernel: RETBleed: Mitigation: IBRS May 17 00:42:26.169842 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier May 17 00:42:26.169862 kernel: Spectre V2 : User space: Mitigation: STIBP via seccomp and prctl May 17 00:42:26.169882 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp May 17 00:42:26.169901 kernel: MDS: Mitigation: Clear CPU buffers May 17 00:42:26.169923 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode May 17 00:42:26.169943 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' May 17 00:42:26.169961 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' May 17 00:42:26.169986 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' May 17 00:42:26.170005 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 May 17 00:42:26.170024 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. May 17 00:42:26.170043 kernel: Freeing SMP alternatives memory: 32K May 17 00:42:26.170063 kernel: pid_max: default: 32768 minimum: 301 May 17 00:42:26.170081 kernel: LSM: Security Framework initializing May 17 00:42:26.170104 kernel: SELinux: Initializing. May 17 00:42:26.170123 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) May 17 00:42:26.170143 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) May 17 00:42:26.170162 kernel: smpboot: CPU0: Intel(R) Xeon(R) CPU @ 2.30GHz (family: 0x6, model: 0x3f, stepping: 0x0) May 17 00:42:26.170182 kernel: Performance Events: unsupported p6 CPU model 63 no PMU driver, software events only. May 17 00:42:26.170207 kernel: signal: max sigframe size: 1776 May 17 00:42:26.170226 kernel: rcu: Hierarchical SRCU implementation. May 17 00:42:26.170259 kernel: NMI watchdog: Perf NMI watchdog permanently disabled May 17 00:42:26.170279 kernel: smp: Bringing up secondary CPUs ... May 17 00:42:26.170302 kernel: x86: Booting SMP configuration: May 17 00:42:26.170321 kernel: .... node #0, CPUs: #1 May 17 00:42:26.170346 kernel: kvm-clock: cpu 1, msr 1619a041, secondary cpu clock May 17 00:42:26.170364 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. May 17 00:42:26.170383 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. May 17 00:42:26.170406 kernel: smp: Brought up 1 node, 2 CPUs May 17 00:42:26.170429 kernel: smpboot: Max logical packages: 1 May 17 00:42:26.170450 kernel: smpboot: Total of 2 processors activated (9199.99 BogoMIPS) May 17 00:42:26.170481 kernel: devtmpfs: initialized May 17 00:42:26.170507 kernel: x86/mm: Memory block size: 128MB May 17 00:42:26.170526 kernel: ACPI: PM: Registering ACPI NVS region [mem 0xbfb7f000-0xbfbfefff] (524288 bytes) May 17 00:42:26.170552 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns May 17 00:42:26.170572 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) May 17 00:42:26.170596 kernel: pinctrl core: initialized pinctrl subsystem May 17 00:42:26.170616 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family May 17 00:42:26.170642 kernel: audit: initializing netlink subsys (disabled) May 17 00:42:26.170663 kernel: audit: type=2000 audit(1747442545.169:1): state=initialized audit_enabled=0 res=1 May 17 00:42:26.170686 kernel: thermal_sys: Registered thermal governor 'step_wise' May 17 00:42:26.170706 kernel: thermal_sys: Registered thermal governor 'user_space' May 17 00:42:26.170727 kernel: cpuidle: using governor menu May 17 00:42:26.170754 kernel: ACPI: bus type PCI registered May 17 00:42:26.170774 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 May 17 00:42:26.170796 kernel: dca service started, version 1.12.1 May 17 00:42:26.170817 kernel: PCI: Using configuration type 1 for base access May 17 00:42:26.170839 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. May 17 00:42:26.170862 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages May 17 00:42:26.170888 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages May 17 00:42:26.170909 kernel: ACPI: Added _OSI(Module Device) May 17 00:42:26.170934 kernel: ACPI: Added _OSI(Processor Device) May 17 00:42:26.170954 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) May 17 00:42:26.170974 kernel: ACPI: Added _OSI(Processor Aggregator Device) May 17 00:42:26.170993 kernel: ACPI: Added _OSI(Linux-Dell-Video) May 17 00:42:26.171015 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) May 17 00:42:26.171041 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) May 17 00:42:26.171061 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded May 17 00:42:26.171087 kernel: ACPI: Interpreter enabled May 17 00:42:26.171110 kernel: ACPI: PM: (supports S0 S3 S5) May 17 00:42:26.171131 kernel: ACPI: Using IOAPIC for interrupt routing May 17 00:42:26.171154 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug May 17 00:42:26.171175 kernel: ACPI: Enabled 16 GPEs in block 00 to 0F May 17 00:42:26.171198 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) May 17 00:42:26.171496 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] May 17 00:42:26.171701 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. May 17 00:42:26.171734 kernel: PCI host bridge to bus 0000:00 May 17 00:42:26.171933 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] May 17 00:42:26.172117 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] May 17 00:42:26.172313 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] May 17 00:42:26.172502 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfefff window] May 17 00:42:26.172675 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] May 17 00:42:26.172907 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 May 17 00:42:26.173124 kernel: pci 0000:00:01.0: [8086:7110] type 00 class 0x060100 May 17 00:42:26.181751 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 May 17 00:42:26.182015 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI May 17 00:42:26.182242 kernel: pci 0000:00:03.0: [1af4:1004] type 00 class 0x000000 May 17 00:42:26.182495 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc040-0xc07f] May 17 00:42:26.182716 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc0001000-0xc000107f] May 17 00:42:26.182949 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 May 17 00:42:26.183164 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc03f] May 17 00:42:26.183413 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc0000000-0xc000007f] May 17 00:42:26.183632 kernel: pci 0000:00:05.0: [1af4:1005] type 00 class 0x00ff00 May 17 00:42:26.183835 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc080-0xc09f] May 17 00:42:26.184035 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xc0002000-0xc000203f] May 17 00:42:26.184064 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 May 17 00:42:26.184096 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 May 17 00:42:26.184117 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 May 17 00:42:26.184138 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 May 17 00:42:26.184157 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 May 17 00:42:26.184181 kernel: iommu: Default domain type: Translated May 17 00:42:26.184202 kernel: iommu: DMA domain TLB invalidation policy: lazy mode May 17 00:42:26.184228 kernel: vgaarb: loaded May 17 00:42:26.184935 kernel: pps_core: LinuxPPS API ver. 1 registered May 17 00:42:26.184970 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti May 17 00:42:26.185000 kernel: PTP clock support registered May 17 00:42:26.185026 kernel: Registered efivars operations May 17 00:42:26.185047 kernel: PCI: Using ACPI for IRQ routing May 17 00:42:26.185067 kernel: PCI: pci_cache_line_size set to 64 bytes May 17 00:42:26.185085 kernel: e820: reserve RAM buffer [mem 0x00055000-0x0005ffff] May 17 00:42:26.185105 kernel: e820: reserve RAM buffer [mem 0x00098000-0x0009ffff] May 17 00:42:26.185127 kernel: e820: reserve RAM buffer [mem 0xbd278000-0xbfffffff] May 17 00:42:26.185149 kernel: e820: reserve RAM buffer [mem 0xbf8ed000-0xbfffffff] May 17 00:42:26.185174 kernel: e820: reserve RAM buffer [mem 0xbffe0000-0xbfffffff] May 17 00:42:26.185199 kernel: clocksource: Switched to clocksource kvm-clock May 17 00:42:26.185223 kernel: VFS: Disk quotas dquot_6.6.0 May 17 00:42:26.185244 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) May 17 00:42:26.185285 kernel: pnp: PnP ACPI init May 17 00:42:26.185309 kernel: pnp: PnP ACPI: found 7 devices May 17 00:42:26.185339 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns May 17 00:42:26.185359 kernel: NET: Registered PF_INET protocol family May 17 00:42:26.185383 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) May 17 00:42:26.185403 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) May 17 00:42:26.185438 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) May 17 00:42:26.185460 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) May 17 00:42:26.185479 kernel: TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear) May 17 00:42:26.185499 kernel: TCP: Hash tables configured (established 65536 bind 65536) May 17 00:42:26.185525 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) May 17 00:42:26.185548 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) May 17 00:42:26.185568 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family May 17 00:42:26.185594 kernel: NET: Registered PF_XDP protocol family May 17 00:42:26.185809 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] May 17 00:42:26.186023 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] May 17 00:42:26.186208 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] May 17 00:42:26.189968 kernel: pci_bus 0000:00: resource 7 [mem 0xc0000000-0xfebfefff window] May 17 00:42:26.190202 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers May 17 00:42:26.190233 kernel: PCI: CLS 0 bytes, default 64 May 17 00:42:26.190290 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) May 17 00:42:26.190322 kernel: software IO TLB: mapped [mem 0x00000000b7f7f000-0x00000000bbf7f000] (64MB) May 17 00:42:26.190363 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer May 17 00:42:26.190386 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x212733415c7, max_idle_ns: 440795236380 ns May 17 00:42:26.190406 kernel: clocksource: Switched to clocksource tsc May 17 00:42:26.190425 kernel: Initialise system trusted keyrings May 17 00:42:26.190450 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 May 17 00:42:26.190473 kernel: Key type asymmetric registered May 17 00:42:26.190493 kernel: Asymmetric key parser 'x509' registered May 17 00:42:26.190519 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) May 17 00:42:26.190544 kernel: io scheduler mq-deadline registered May 17 00:42:26.190569 kernel: io scheduler kyber registered May 17 00:42:26.190589 kernel: io scheduler bfq registered May 17 00:42:26.190614 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 May 17 00:42:26.190636 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 May 17 00:42:26.190856 kernel: virtio-pci 0000:00:03.0: virtio_pci: leaving for legacy driver May 17 00:42:26.190884 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 10 May 17 00:42:26.191081 kernel: virtio-pci 0000:00:04.0: virtio_pci: leaving for legacy driver May 17 00:42:26.191111 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 May 17 00:42:26.191348 kernel: virtio-pci 0000:00:05.0: virtio_pci: leaving for legacy driver May 17 00:42:26.191379 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled May 17 00:42:26.191397 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A May 17 00:42:26.191416 kernel: 00:04: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A May 17 00:42:26.191434 kernel: 00:05: ttyS2 at I/O 0x3e8 (irq = 6, base_baud = 115200) is a 16550A May 17 00:42:26.191456 kernel: 00:06: ttyS3 at I/O 0x2e8 (irq = 7, base_baud = 115200) is a 16550A May 17 00:42:26.191675 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x9009, rev-id 0) May 17 00:42:26.191708 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 May 17 00:42:26.191734 kernel: i8042: Warning: Keylock active May 17 00:42:26.191756 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 May 17 00:42:26.191775 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 May 17 00:42:26.191995 kernel: rtc_cmos 00:00: RTC can wake from S4 May 17 00:42:26.192201 kernel: rtc_cmos 00:00: registered as rtc0 May 17 00:42:26.192481 kernel: rtc_cmos 00:00: setting system clock to 2025-05-17T00:42:25 UTC (1747442545) May 17 00:42:26.192691 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram May 17 00:42:26.192723 kernel: intel_pstate: CPU model not supported May 17 00:42:26.192755 kernel: pstore: Registered efi as persistent store backend May 17 00:42:26.192781 kernel: NET: Registered PF_INET6 protocol family May 17 00:42:26.192806 kernel: Segment Routing with IPv6 May 17 00:42:26.192831 kernel: In-situ OAM (IOAM) with IPv6 May 17 00:42:26.192856 kernel: NET: Registered PF_PACKET protocol family May 17 00:42:26.192881 kernel: Key type dns_resolver registered May 17 00:42:26.192906 kernel: IPI shorthand broadcast: enabled May 17 00:42:26.192931 kernel: sched_clock: Marking stable (835006574, 175731056)->(1051217719, -40480089) May 17 00:42:26.192956 kernel: registered taskstats version 1 May 17 00:42:26.192985 kernel: Loading compiled-in X.509 certificates May 17 00:42:26.193009 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 May 17 00:42:26.193035 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.182-flatcar: 01ca23caa8e5879327538f9287e5164b3e97ac0c' May 17 00:42:26.193060 kernel: Key type .fscrypt registered May 17 00:42:26.193093 kernel: Key type fscrypt-provisioning registered May 17 00:42:26.193119 kernel: pstore: Using crash dump compression: deflate May 17 00:42:26.193144 kernel: ima: Allocated hash algorithm: sha1 May 17 00:42:26.193168 kernel: ima: No architecture policies found May 17 00:42:26.193193 kernel: clk: Disabling unused clocks May 17 00:42:26.193222 kernel: Freeing unused kernel image (initmem) memory: 47472K May 17 00:42:26.193245 kernel: Write protecting the kernel read-only data: 28672k May 17 00:42:26.193290 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K May 17 00:42:26.193315 kernel: Freeing unused kernel image (rodata/data gap) memory: 612K May 17 00:42:26.193340 kernel: Run /init as init process May 17 00:42:26.193365 kernel: with arguments: May 17 00:42:26.193390 kernel: /init May 17 00:42:26.193415 kernel: with environment: May 17 00:42:26.193438 kernel: HOME=/ May 17 00:42:26.193468 kernel: TERM=linux May 17 00:42:26.193493 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a May 17 00:42:26.193528 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 17 00:42:26.193583 systemd[1]: Detected virtualization kvm. May 17 00:42:26.193611 systemd[1]: Detected architecture x86-64. May 17 00:42:26.193639 systemd[1]: Running in initrd. May 17 00:42:26.193666 systemd[1]: No hostname configured, using default hostname. May 17 00:42:26.193698 systemd[1]: Hostname set to . May 17 00:42:26.193727 systemd[1]: Initializing machine ID from VM UUID. May 17 00:42:26.193755 systemd[1]: Queued start job for default target initrd.target. May 17 00:42:26.193783 systemd[1]: Started systemd-ask-password-console.path. May 17 00:42:26.193811 systemd[1]: Reached target cryptsetup.target. May 17 00:42:26.193837 systemd[1]: Reached target paths.target. May 17 00:42:26.193864 systemd[1]: Reached target slices.target. May 17 00:42:26.193891 systemd[1]: Reached target swap.target. May 17 00:42:26.193931 systemd[1]: Reached target timers.target. May 17 00:42:26.193952 systemd[1]: Listening on iscsid.socket. May 17 00:42:26.193974 systemd[1]: Listening on iscsiuio.socket. May 17 00:42:26.193994 systemd[1]: Listening on systemd-journald-audit.socket. May 17 00:42:26.194013 systemd[1]: Listening on systemd-journald-dev-log.socket. May 17 00:42:26.194033 systemd[1]: Listening on systemd-journald.socket. May 17 00:42:26.194053 systemd[1]: Listening on systemd-networkd.socket. May 17 00:42:26.194073 systemd[1]: Listening on systemd-udevd-control.socket. May 17 00:42:26.194107 systemd[1]: Listening on systemd-udevd-kernel.socket. May 17 00:42:26.194128 systemd[1]: Reached target sockets.target. May 17 00:42:26.194168 systemd[1]: Starting kmod-static-nodes.service... May 17 00:42:26.194192 systemd[1]: Finished network-cleanup.service. May 17 00:42:26.194213 systemd[1]: Starting systemd-fsck-usr.service... May 17 00:42:26.194234 systemd[1]: Starting systemd-journald.service... May 17 00:42:26.200216 systemd[1]: Starting systemd-modules-load.service... May 17 00:42:26.200303 systemd[1]: Starting systemd-resolved.service... May 17 00:42:26.200340 systemd[1]: Starting systemd-vconsole-setup.service... May 17 00:42:26.200370 systemd[1]: Finished kmod-static-nodes.service. May 17 00:42:26.200405 systemd-journald[189]: Journal started May 17 00:42:26.200554 systemd-journald[189]: Runtime Journal (/run/log/journal/1478b8602b6203bc8936d3dde1dce9ce) is 8.0M, max 148.8M, 140.8M free. May 17 00:42:26.186341 systemd-modules-load[190]: Inserted module 'overlay' May 17 00:42:26.244751 systemd[1]: Started systemd-journald.service. May 17 00:42:26.244846 kernel: audit: type=1130 audit(1747442546.217:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.217000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.264279 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. May 17 00:42:26.277292 kernel: Bridge firewalling registered May 17 00:42:26.277331 systemd-modules-load[190]: Inserted module 'br_netfilter' May 17 00:42:26.313429 kernel: audit: type=1130 audit(1747442546.282:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.282000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.284515 systemd[1]: Finished systemd-fsck-usr.service. May 17 00:42:26.294585 systemd-resolved[191]: Positive Trust Anchors: May 17 00:42:26.331412 kernel: SCSI subsystem initialized May 17 00:42:26.294604 systemd-resolved[191]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 17 00:42:26.373016 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. May 17 00:42:26.373068 kernel: device-mapper: uevent: version 1.0.3 May 17 00:42:26.373099 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com May 17 00:42:26.294691 systemd-resolved[191]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 17 00:42:26.446489 kernel: audit: type=1130 audit(1747442546.380:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.380000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.300838 systemd-resolved[191]: Defaulting to hostname 'linux'. May 17 00:42:26.482416 kernel: audit: type=1130 audit(1747442546.453:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.453000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.373363 systemd-modules-load[190]: Inserted module 'dm_multipath' May 17 00:42:26.520447 kernel: audit: type=1130 audit(1747442546.489:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.489000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.381903 systemd[1]: Started systemd-resolved.service. May 17 00:42:26.557473 kernel: audit: type=1130 audit(1747442546.528:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.528000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.455007 systemd[1]: Finished systemd-modules-load.service. May 17 00:42:26.490962 systemd[1]: Finished systemd-vconsole-setup.service. May 17 00:42:26.529919 systemd[1]: Reached target nss-lookup.target. May 17 00:42:26.566605 systemd[1]: Starting dracut-cmdline-ask.service... May 17 00:42:26.574858 systemd[1]: Starting systemd-sysctl.service... May 17 00:42:26.577545 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 17 00:42:26.595000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.596597 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 17 00:42:26.618391 kernel: audit: type=1130 audit(1747442546.595:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.619744 systemd[1]: Finished systemd-sysctl.service. May 17 00:42:26.618000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.641301 kernel: audit: type=1130 audit(1747442546.618:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.646685 systemd[1]: Finished dracut-cmdline-ask.service. May 17 00:42:26.685602 kernel: audit: type=1130 audit(1747442546.654:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.654000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.657423 systemd[1]: Starting dracut-cmdline.service... May 17 00:42:26.699422 dracut-cmdline[210]: dracut-dracut-053 May 17 00:42:26.699422 dracut-cmdline[210]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LA May 17 00:42:26.699422 dracut-cmdline[210]: BEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=4aad7caeadb0359f379975532748a0b4ae6bb9b229507353e0f5ae84cb9335a0 May 17 00:42:26.777294 kernel: Loading iSCSI transport class v2.0-870. May 17 00:42:26.803301 kernel: iscsi: registered transport (tcp) May 17 00:42:26.839582 kernel: iscsi: registered transport (qla4xxx) May 17 00:42:26.839671 kernel: QLogic iSCSI HBA Driver May 17 00:42:26.893622 systemd[1]: Finished dracut-cmdline.service. May 17 00:42:26.901000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:26.904141 systemd[1]: Starting dracut-pre-udev.service... May 17 00:42:26.969343 kernel: raid6: avx2x4 gen() 17872 MB/s May 17 00:42:26.990319 kernel: raid6: avx2x4 xor() 7843 MB/s May 17 00:42:27.011332 kernel: raid6: avx2x2 gen() 17726 MB/s May 17 00:42:27.032301 kernel: raid6: avx2x2 xor() 17283 MB/s May 17 00:42:27.053327 kernel: raid6: avx2x1 gen() 13664 MB/s May 17 00:42:27.074300 kernel: raid6: avx2x1 xor() 15561 MB/s May 17 00:42:27.095295 kernel: raid6: sse2x4 gen() 10891 MB/s May 17 00:42:27.116303 kernel: raid6: sse2x4 xor() 6618 MB/s May 17 00:42:27.137317 kernel: raid6: sse2x2 gen() 11917 MB/s May 17 00:42:27.158321 kernel: raid6: sse2x2 xor() 7290 MB/s May 17 00:42:27.179318 kernel: raid6: sse2x1 gen() 10371 MB/s May 17 00:42:27.205325 kernel: raid6: sse2x1 xor() 5135 MB/s May 17 00:42:27.205411 kernel: raid6: using algorithm avx2x4 gen() 17872 MB/s May 17 00:42:27.205441 kernel: raid6: .... xor() 7843 MB/s, rmw enabled May 17 00:42:27.210419 kernel: raid6: using avx2x2 recovery algorithm May 17 00:42:27.237304 kernel: xor: automatically using best checksumming function avx May 17 00:42:27.365307 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no May 17 00:42:27.378532 systemd[1]: Finished dracut-pre-udev.service. May 17 00:42:27.386000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:27.386000 audit: BPF prog-id=7 op=LOAD May 17 00:42:27.386000 audit: BPF prog-id=8 op=LOAD May 17 00:42:27.389110 systemd[1]: Starting systemd-udevd.service... May 17 00:42:27.407203 systemd-udevd[387]: Using default interface naming scheme 'v252'. May 17 00:42:27.416077 systemd[1]: Started systemd-udevd.service. May 17 00:42:27.434000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:27.436656 systemd[1]: Starting dracut-pre-trigger.service... May 17 00:42:27.454302 dracut-pre-trigger[393]: rd.md=0: removing MD RAID activation May 17 00:42:27.496446 systemd[1]: Finished dracut-pre-trigger.service. May 17 00:42:27.495000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:27.497804 systemd[1]: Starting systemd-udev-trigger.service... May 17 00:42:27.577331 systemd[1]: Finished systemd-udev-trigger.service. May 17 00:42:27.584000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:27.679277 kernel: cryptd: max_cpu_qlen set to 1000 May 17 00:42:27.830572 kernel: AVX2 version of gcm_enc/dec engaged. May 17 00:42:27.836285 kernel: AES CTR mode by8 optimization enabled May 17 00:42:27.865725 kernel: scsi host0: Virtio SCSI HBA May 17 00:42:27.888281 kernel: scsi 0:0:1:0: Direct-Access Google PersistentDisk 1 PQ: 0 ANSI: 6 May 17 00:42:27.953106 kernel: sd 0:0:1:0: [sda] 25165824 512-byte logical blocks: (12.9 GB/12.0 GiB) May 17 00:42:28.015930 kernel: sd 0:0:1:0: [sda] 4096-byte physical blocks May 17 00:42:28.016289 kernel: sd 0:0:1:0: [sda] Write Protect is off May 17 00:42:28.016583 kernel: sd 0:0:1:0: [sda] Mode Sense: 1f 00 00 08 May 17 00:42:28.016883 kernel: sd 0:0:1:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA May 17 00:42:28.017166 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. May 17 00:42:28.017210 kernel: GPT:17805311 != 25165823 May 17 00:42:28.017263 kernel: GPT:Alternate GPT header not at the end of the disk. May 17 00:42:28.017306 kernel: GPT:17805311 != 25165823 May 17 00:42:28.017344 kernel: GPT: Use GNU Parted to correct GPT errors. May 17 00:42:28.017371 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 17 00:42:28.017410 kernel: sd 0:0:1:0: [sda] Attached SCSI disk May 17 00:42:28.092283 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by (udev-worker) (436) May 17 00:42:28.104018 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. May 17 00:42:28.126453 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. May 17 00:42:28.133510 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. May 17 00:42:28.165970 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. May 17 00:42:28.176422 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 17 00:42:28.192712 systemd[1]: Starting disk-uuid.service... May 17 00:42:28.214616 disk-uuid[505]: Primary Header is updated. May 17 00:42:28.214616 disk-uuid[505]: Secondary Entries is updated. May 17 00:42:28.214616 disk-uuid[505]: Secondary Header is updated. May 17 00:42:28.246678 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 17 00:42:28.246762 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 17 00:42:28.272282 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 17 00:42:29.283754 disk-uuid[506]: The operation has completed successfully. May 17 00:42:29.292425 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 May 17 00:42:29.368410 systemd[1]: disk-uuid.service: Deactivated successfully. May 17 00:42:29.367000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.367000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.368587 systemd[1]: Finished disk-uuid.service. May 17 00:42:29.381091 systemd[1]: Starting verity-setup.service... May 17 00:42:29.418450 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" May 17 00:42:29.503995 systemd[1]: Found device dev-mapper-usr.device. May 17 00:42:29.507395 systemd[1]: Mounting sysusr-usr.mount... May 17 00:42:29.519066 systemd[1]: Finished verity-setup.service. May 17 00:42:29.530000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.627307 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. May 17 00:42:29.627514 systemd[1]: Mounted sysusr-usr.mount. May 17 00:42:29.627954 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. May 17 00:42:29.671107 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm May 17 00:42:29.671161 kernel: BTRFS info (device sda6): using free space tree May 17 00:42:29.671192 kernel: BTRFS info (device sda6): has skinny extents May 17 00:42:29.629053 systemd[1]: Starting ignition-setup.service... May 17 00:42:29.691452 kernel: BTRFS info (device sda6): enabling ssd optimizations May 17 00:42:29.680279 systemd[1]: Starting parse-ip-for-networkd.service... May 17 00:42:29.731798 systemd[1]: Finished ignition-setup.service. May 17 00:42:29.739000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.742702 systemd[1]: Starting ignition-fetch-offline.service... May 17 00:42:29.811944 systemd[1]: Finished parse-ip-for-networkd.service. May 17 00:42:29.819000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.820000 audit: BPF prog-id=9 op=LOAD May 17 00:42:29.822771 systemd[1]: Starting systemd-networkd.service... May 17 00:42:29.863870 systemd-networkd[680]: lo: Link UP May 17 00:42:29.863890 systemd-networkd[680]: lo: Gained carrier May 17 00:42:29.871000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.865165 systemd-networkd[680]: Enumeration completed May 17 00:42:29.865366 systemd[1]: Started systemd-networkd.service. May 17 00:42:29.866052 systemd-networkd[680]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 17 00:42:29.923000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.868591 systemd-networkd[680]: eth0: Link UP May 17 00:42:29.951567 iscsid[691]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi May 17 00:42:29.951567 iscsid[691]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log May 17 00:42:29.951567 iscsid[691]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. May 17 00:42:29.951567 iscsid[691]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. May 17 00:42:29.951567 iscsid[691]: If using hardware iscsi like qla4xxx this message can be ignored. May 17 00:42:29.951567 iscsid[691]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi May 17 00:42:29.951567 iscsid[691]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf May 17 00:42:30.119601 kernel: kauditd_printk_skb: 16 callbacks suppressed May 17 00:42:30.119655 kernel: audit: type=1130 audit(1747442550.011:27): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.957000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.011000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.868601 systemd-networkd[680]: eth0: Gained carrier May 17 00:42:30.118738 ignition[620]: Ignition 2.14.0 May 17 00:42:30.179513 kernel: audit: type=1130 audit(1747442550.151:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.151000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.872897 systemd[1]: Reached target network.target. May 17 00:42:30.216508 kernel: audit: type=1130 audit(1747442550.187:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.187000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.118752 ignition[620]: Stage: fetch-offline May 17 00:42:29.882558 systemd-networkd[680]: eth0: Overlong DHCP hostname received, shortened from 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9.c.flatcar-212911.internal' to 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9' May 17 00:42:30.118831 ignition[620]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:29.882577 systemd-networkd[680]: eth0: DHCPv4 address 10.128.0.10/32, gateway 10.128.0.1 acquired from 169.254.169.254 May 17 00:42:30.305459 kernel: audit: type=1130 audit(1747442550.271:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.271000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.118870 ignition[620]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:29.896226 systemd[1]: Starting iscsiuio.service... May 17 00:42:30.137838 ignition[620]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:29.916796 systemd[1]: Started iscsiuio.service. May 17 00:42:30.138163 ignition[620]: parsed url from cmdline: "" May 17 00:42:29.926365 systemd[1]: Starting iscsid.service... May 17 00:42:30.138171 ignition[620]: no config URL provided May 17 00:42:30.378597 kernel: audit: type=1130 audit(1747442550.350:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.350000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.944646 systemd[1]: Started iscsid.service. May 17 00:42:30.138179 ignition[620]: reading system config file "/usr/lib/ignition/user.ign" May 17 00:42:30.431457 kernel: audit: type=1130 audit(1747442550.400:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.400000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:29.960267 systemd[1]: Starting dracut-initqueue.service... May 17 00:42:30.138192 ignition[620]: no config at "/usr/lib/ignition/user.ign" May 17 00:42:29.980750 systemd[1]: Finished dracut-initqueue.service. May 17 00:42:30.138202 ignition[620]: failed to fetch config: resource requires networking May 17 00:42:30.012678 systemd[1]: Reached target remote-fs-pre.target. May 17 00:42:30.138409 ignition[620]: Ignition finished successfully May 17 00:42:30.069439 systemd[1]: Reached target remote-cryptsetup.target. May 17 00:42:30.208387 ignition[705]: Ignition 2.14.0 May 17 00:42:30.069598 systemd[1]: Reached target remote-fs.target. May 17 00:42:30.208401 ignition[705]: Stage: fetch May 17 00:42:30.095781 systemd[1]: Starting dracut-pre-mount.service... May 17 00:42:30.208720 ignition[705]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:30.137084 systemd[1]: Finished dracut-pre-mount.service. May 17 00:42:30.208801 ignition[705]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:30.153087 systemd[1]: Finished ignition-fetch-offline.service. May 17 00:42:30.219135 ignition[705]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:30.190534 systemd[1]: Starting ignition-fetch.service... May 17 00:42:30.219378 ignition[705]: parsed url from cmdline: "" May 17 00:42:30.230187 unknown[705]: fetched base config from "system" May 17 00:42:30.219386 ignition[705]: no config URL provided May 17 00:42:30.230203 unknown[705]: fetched base config from "system" May 17 00:42:30.219394 ignition[705]: reading system config file "/usr/lib/ignition/user.ign" May 17 00:42:30.230217 unknown[705]: fetched user config from "gcp" May 17 00:42:30.219406 ignition[705]: no config at "/usr/lib/ignition/user.ign" May 17 00:42:30.254980 systemd[1]: Finished ignition-fetch.service. May 17 00:42:30.219446 ignition[705]: GET http://169.254.169.254/computeMetadata/v1/instance/attributes/user-data: attempt #1 May 17 00:42:30.274375 systemd[1]: Starting ignition-kargs.service... May 17 00:42:30.225082 ignition[705]: GET result: OK May 17 00:42:30.335961 systemd[1]: Finished ignition-kargs.service. May 17 00:42:30.225150 ignition[705]: parsing config with SHA512: e183ef97e48321d1f75ab18f7db04d6150a08fba2c3f08519bbf72c668dc2bc17d764b0423a9cc35d004ee5a58577a468648103c11285e5ef61d7287103543a0 May 17 00:42:30.353474 systemd[1]: Starting ignition-disks.service... May 17 00:42:30.233667 ignition[705]: fetch: fetch complete May 17 00:42:30.387053 systemd[1]: Finished ignition-disks.service. May 17 00:42:30.233685 ignition[705]: fetch: fetch passed May 17 00:42:30.401829 systemd[1]: Reached target initrd-root-device.target. May 17 00:42:30.233778 ignition[705]: Ignition finished successfully May 17 00:42:30.439678 systemd[1]: Reached target local-fs-pre.target. May 17 00:42:30.308298 ignition[711]: Ignition 2.14.0 May 17 00:42:30.456454 systemd[1]: Reached target local-fs.target. May 17 00:42:30.308309 ignition[711]: Stage: kargs May 17 00:42:30.470446 systemd[1]: Reached target sysinit.target. May 17 00:42:30.308458 ignition[711]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:30.483426 systemd[1]: Reached target basic.target. May 17 00:42:30.308496 ignition[711]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:30.499835 systemd[1]: Starting systemd-fsck-root.service... May 17 00:42:30.316584 ignition[711]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:30.318111 ignition[711]: kargs: kargs passed May 17 00:42:30.318167 ignition[711]: Ignition finished successfully May 17 00:42:30.372185 ignition[717]: Ignition 2.14.0 May 17 00:42:30.372198 ignition[717]: Stage: disks May 17 00:42:30.372450 ignition[717]: reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:30.372534 ignition[717]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:30.381899 ignition[717]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:30.383488 ignition[717]: disks: disks passed May 17 00:42:30.383549 ignition[717]: Ignition finished successfully May 17 00:42:30.550745 systemd-fsck[725]: ROOT: clean, 619/1628000 files, 124060/1617920 blocks May 17 00:42:30.781384 systemd[1]: Finished systemd-fsck-root.service. May 17 00:42:30.816614 kernel: audit: type=1130 audit(1747442550.780:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.780000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:30.783308 systemd[1]: Mounting sysroot.mount... May 17 00:42:30.839462 kernel: EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. May 17 00:42:30.833815 systemd[1]: Mounted sysroot.mount. May 17 00:42:30.846872 systemd[1]: Reached target initrd-root-fs.target. May 17 00:42:30.865090 systemd[1]: Mounting sysroot-usr.mount... May 17 00:42:30.882947 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. May 17 00:42:30.883044 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). May 17 00:42:30.883107 systemd[1]: Reached target ignition-diskful.target. May 17 00:42:30.905219 systemd[1]: Mounted sysroot-usr.mount. May 17 00:42:30.939879 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 17 00:42:30.964436 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (731) May 17 00:42:30.948759 systemd[1]: Starting initrd-setup-root.service... May 17 00:42:30.989963 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm May 17 00:42:30.990031 kernel: BTRFS info (device sda6): using free space tree May 17 00:42:30.990062 kernel: BTRFS info (device sda6): has skinny extents May 17 00:42:30.990403 initrd-setup-root[736]: cut: /sysroot/etc/passwd: No such file or directory May 17 00:42:31.014427 kernel: BTRFS info (device sda6): enabling ssd optimizations May 17 00:42:31.014474 initrd-setup-root[760]: cut: /sysroot/etc/group: No such file or directory May 17 00:42:31.025494 initrd-setup-root[770]: cut: /sysroot/etc/shadow: No such file or directory May 17 00:42:31.018916 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 17 00:42:31.053445 initrd-setup-root[778]: cut: /sysroot/etc/gshadow: No such file or directory May 17 00:42:31.195302 systemd[1]: Finished initrd-setup-root.service. May 17 00:42:31.229616 kernel: audit: type=1130 audit(1747442551.194:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:31.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:31.197172 systemd[1]: Starting ignition-mount.service... May 17 00:42:31.225412 systemd-networkd[680]: eth0: Gained IPv6LL May 17 00:42:31.237678 systemd[1]: Starting sysroot-boot.service... May 17 00:42:31.252152 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. May 17 00:42:31.275417 ignition[797]: INFO : Ignition 2.14.0 May 17 00:42:31.275417 ignition[797]: INFO : Stage: mount May 17 00:42:31.275417 ignition[797]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:31.275417 ignition[797]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:31.326701 kernel: audit: type=1130 audit(1747442551.288:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:31.288000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:31.252285 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. May 17 00:42:31.381476 kernel: audit: type=1130 audit(1747442551.350:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:31.350000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:31.381623 ignition[797]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:31.381623 ignition[797]: INFO : mount: mount passed May 17 00:42:31.381623 ignition[797]: INFO : Ignition finished successfully May 17 00:42:31.442423 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (806) May 17 00:42:31.442459 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm May 17 00:42:31.442476 kernel: BTRFS info (device sda6): using free space tree May 17 00:42:31.442517 kernel: BTRFS info (device sda6): has skinny extents May 17 00:42:31.286235 systemd[1]: Finished sysroot-boot.service. May 17 00:42:31.456463 kernel: BTRFS info (device sda6): enabling ssd optimizations May 17 00:42:31.292921 systemd[1]: Finished ignition-mount.service. May 17 00:42:31.353446 systemd[1]: Starting ignition-files.service... May 17 00:42:31.392698 systemd[1]: Mounting sysroot-usr-share-oem.mount... May 17 00:42:31.481154 systemd[1]: Mounted sysroot-usr-share-oem.mount. May 17 00:42:31.509733 ignition[825]: INFO : Ignition 2.14.0 May 17 00:42:31.509733 ignition[825]: INFO : Stage: files May 17 00:42:31.525383 ignition[825]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:31.525383 ignition[825]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:31.525383 ignition[825]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:31.525383 ignition[825]: DEBUG : files: compiled without relabeling support, skipping May 17 00:42:31.577456 ignition[825]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" May 17 00:42:31.577456 ignition[825]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" May 17 00:42:31.577456 ignition[825]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" May 17 00:42:31.577456 ignition[825]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" May 17 00:42:31.577456 ignition[825]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" May 17 00:42:31.577456 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" May 17 00:42:31.577456 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" May 17 00:42:31.577456 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 17 00:42:31.577456 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 May 17 00:42:31.538017 unknown[825]: wrote ssh authorized keys file for user: core May 17 00:42:31.713468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK May 17 00:42:31.918572 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/hosts" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): oem config not found in "/usr/share/oem", looking on oem partition May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): op(6): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem443735354" May 17 00:42:31.935422 ignition[825]: CRITICAL : files: createFilesystemsFiles: createFiles: op(5): op(6): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem443735354": device or resource busy May 17 00:42:31.935422 ignition[825]: ERROR : files: createFilesystemsFiles: createFiles: op(5): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem443735354", trying btrfs: device or resource busy May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): op(7): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem443735354" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): op(7): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem443735354" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): op(8): [started] unmounting "/mnt/oem443735354" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): op(8): [finished] unmounting "/mnt/oem443735354" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/hosts" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/home/core/nginx.yaml" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nginx.yaml" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/profile.d/google-cloud-sdk.sh" May 17 00:42:31.935422 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): oem config not found in "/usr/share/oem", looking on oem partition May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): op(b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2962573534" May 17 00:42:32.163414 ignition[825]: CRITICAL : files: createFilesystemsFiles: createFiles: op(a): op(b): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2962573534": device or resource busy May 17 00:42:32.163414 ignition[825]: ERROR : files: createFilesystemsFiles: createFiles: op(a): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem2962573534", trying btrfs: device or resource busy May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): op(c): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2962573534" May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): op(c): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem2962573534" May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): op(d): [started] unmounting "/mnt/oem2962573534" May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): op(d): [finished] unmounting "/mnt/oem2962573534" May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/profile.d/google-cloud-sdk.sh" May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" May 17 00:42:32.163414 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(e): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 May 17 00:42:31.936487 systemd[1]: mnt-oem443735354.mount: Deactivated successfully. May 17 00:42:32.344445 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(e): GET result: OK May 17 00:42:31.957370 systemd[1]: mnt-oem2962573534.mount: Deactivated successfully. May 17 00:42:32.392774 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/home/core/install.sh" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/home/core/install.sh" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(10): [started] writing file "/sysroot/home/core/nfs-pod.yaml" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(10): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(11): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(11): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(12): [started] writing file "/sysroot/etc/flatcar/update.conf" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(12): [finished] writing file "/sysroot/etc/flatcar/update.conf" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(13): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(13): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): [started] writing file "/sysroot/etc/systemd/system/oem-gce-enable-oslogin.service" May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): oem config not found in "/usr/share/oem", looking on oem partition May 17 00:42:32.408468 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): op(15): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem23434667" May 17 00:42:32.408468 ignition[825]: CRITICAL : files: createFilesystemsFiles: createFiles: op(14): op(15): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem23434667": device or resource busy May 17 00:42:32.650437 ignition[825]: ERROR : files: createFilesystemsFiles: createFiles: op(14): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem23434667", trying btrfs: device or resource busy May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): op(16): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem23434667" May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): op(16): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem23434667" May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): op(17): [started] unmounting "/mnt/oem23434667" May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): op(17): [finished] unmounting "/mnt/oem23434667" May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(14): [finished] writing file "/sysroot/etc/systemd/system/oem-gce-enable-oslogin.service" May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(18): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:42:32.650437 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(18): GET https://extensions.flatcar.org/extensions/kubernetes-v1.31.8-x86-64.raw: attempt #1 May 17 00:42:32.805761 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(18): GET result: OK May 17 00:42:33.221345 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(18): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.31.8-x86-64.raw" May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): [started] writing file "/sysroot/etc/systemd/system/oem-gce.service" May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): oem config not found in "/usr/share/oem", looking on oem partition May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): op(1a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1856799168" May 17 00:42:33.240455 ignition[825]: CRITICAL : files: createFilesystemsFiles: createFiles: op(19): op(1a): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1856799168": device or resource busy May 17 00:42:33.240455 ignition[825]: ERROR : files: createFilesystemsFiles: createFiles: op(19): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1856799168", trying btrfs: device or resource busy May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): op(1b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1856799168" May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): op(1b): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1856799168" May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): op(1c): [started] unmounting "/mnt/oem1856799168" May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): op(1c): [finished] unmounting "/mnt/oem1856799168" May 17 00:42:33.240455 ignition[825]: INFO : files: createFilesystemsFiles: createFiles: op(19): [finished] writing file "/sysroot/etc/systemd/system/oem-gce.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(1d): [started] processing unit "coreos-metadata-sshkeys@.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(1d): [finished] processing unit "coreos-metadata-sshkeys@.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(1e): [started] processing unit "oem-gce.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(1e): [finished] processing unit "oem-gce.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(1f): [started] processing unit "oem-gce-enable-oslogin.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(1f): [finished] processing unit "oem-gce-enable-oslogin.service" May 17 00:42:33.240455 ignition[825]: INFO : files: op(20): [started] processing unit "containerd.service" May 17 00:42:33.281000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.379000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.404000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.404000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.466000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.466000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.558000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.245843 systemd[1]: mnt-oem1856799168.mount: Deactivated successfully. May 17 00:42:33.600454 ignition[825]: INFO : files: op(20): op(21): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 17 00:42:33.600454 ignition[825]: INFO : files: op(20): op(21): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" May 17 00:42:33.600454 ignition[825]: INFO : files: op(20): [finished] processing unit "containerd.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(22): [started] processing unit "prepare-helm.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(22): op(23): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(22): op(23): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(22): [finished] processing unit "prepare-helm.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(24): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " May 17 00:42:33.600454 ignition[825]: INFO : files: op(24): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " May 17 00:42:33.600454 ignition[825]: INFO : files: op(25): [started] setting preset to enabled for "oem-gce.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(25): [finished] setting preset to enabled for "oem-gce.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(26): [started] setting preset to enabled for "oem-gce-enable-oslogin.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(26): [finished] setting preset to enabled for "oem-gce-enable-oslogin.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(27): [started] setting preset to enabled for "prepare-helm.service" May 17 00:42:33.600454 ignition[825]: INFO : files: op(27): [finished] setting preset to enabled for "prepare-helm.service" May 17 00:42:33.600454 ignition[825]: INFO : files: createResultFile: createFiles: op(28): [started] writing file "/sysroot/etc/.ignition-result.json" May 17 00:42:33.600454 ignition[825]: INFO : files: createResultFile: createFiles: op(28): [finished] writing file "/sysroot/etc/.ignition-result.json" May 17 00:42:33.600454 ignition[825]: INFO : files: files passed May 17 00:42:33.600454 ignition[825]: INFO : Ignition finished successfully May 17 00:42:33.690000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.261719 systemd[1]: Finished ignition-files.service. May 17 00:42:33.963000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.290209 systemd[1]: Starting initrd-setup-root-after-ignition.service... May 17 00:42:33.991681 initrd-setup-root-after-ignition[848]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory May 17 00:42:33.998000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.319681 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). May 17 00:42:34.022000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.321067 systemd[1]: Starting ignition-quench.service... May 17 00:42:34.048000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.353872 systemd[1]: Finished initrd-setup-root-after-ignition.service. May 17 00:42:33.380887 systemd[1]: ignition-quench.service: Deactivated successfully. May 17 00:42:33.381021 systemd[1]: Finished ignition-quench.service. May 17 00:42:33.405717 systemd[1]: Reached target ignition-complete.target. May 17 00:42:34.106611 iscsid[691]: iscsid shutting down. May 17 00:42:34.114000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.430788 systemd[1]: Starting initrd-parse-etc.service... May 17 00:42:34.131562 ignition[863]: INFO : Ignition 2.14.0 May 17 00:42:34.131562 ignition[863]: INFO : Stage: umount May 17 00:42:34.131562 ignition[863]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" May 17 00:42:34.131562 ignition[863]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 May 17 00:42:34.131562 ignition[863]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" May 17 00:42:34.131562 ignition[863]: INFO : umount: umount passed May 17 00:42:34.131562 ignition[863]: INFO : Ignition finished successfully May 17 00:42:34.155000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.162000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.179000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.221000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.228000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.462693 systemd[1]: initrd-parse-etc.service: Deactivated successfully. May 17 00:42:34.250000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.250000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.462838 systemd[1]: Finished initrd-parse-etc.service. May 17 00:42:34.266000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.467863 systemd[1]: Reached target initrd-fs.target. May 17 00:42:34.281000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.485688 systemd[1]: Reached target initrd.target. May 17 00:42:34.296000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.503735 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. May 17 00:42:34.311000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.505186 systemd[1]: Starting dracut-pre-pivot.service... May 17 00:42:33.538813 systemd[1]: Finished dracut-pre-pivot.service. May 17 00:42:33.560924 systemd[1]: Starting initrd-cleanup.service... May 17 00:42:33.600461 systemd[1]: Stopped target nss-lookup.target. May 17 00:42:33.625742 systemd[1]: Stopped target remote-cryptsetup.target. May 17 00:42:33.654841 systemd[1]: Stopped target timers.target. May 17 00:42:34.401000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.672807 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. May 17 00:42:34.417000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.673046 systemd[1]: Stopped dracut-pre-pivot.service. May 17 00:42:33.691955 systemd[1]: Stopped target initrd.target. May 17 00:42:33.730841 systemd[1]: Stopped target basic.target. May 17 00:42:34.453000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.741933 systemd[1]: Stopped target ignition-complete.target. May 17 00:42:33.761934 systemd[1]: Stopped target ignition-diskful.target. May 17 00:42:33.782910 systemd[1]: Stopped target initrd-root-device.target. May 17 00:42:33.818850 systemd[1]: Stopped target remote-fs.target. May 17 00:42:33.832882 systemd[1]: Stopped target remote-fs-pre.target. May 17 00:42:33.852889 systemd[1]: Stopped target sysinit.target. May 17 00:42:33.871933 systemd[1]: Stopped target local-fs.target. May 17 00:42:33.890886 systemd[1]: Stopped target local-fs-pre.target. May 17 00:42:34.547000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.932779 systemd[1]: Stopped target swap.target. May 17 00:42:34.562000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.562000 audit: BPF prog-id=6 op=UNLOAD May 17 00:42:33.946686 systemd[1]: dracut-pre-mount.service: Deactivated successfully. May 17 00:42:33.946917 systemd[1]: Stopped dracut-pre-mount.service. May 17 00:42:33.965002 systemd[1]: Stopped target cryptsetup.target. May 17 00:42:34.609000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.981744 systemd[1]: dracut-initqueue.service: Deactivated successfully. May 17 00:42:34.624000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.981963 systemd[1]: Stopped dracut-initqueue.service. May 17 00:42:34.640000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:33.999902 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. May 17 00:42:34.000126 systemd[1]: Stopped initrd-setup-root-after-ignition.service. May 17 00:42:34.672000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.023874 systemd[1]: ignition-files.service: Deactivated successfully. May 17 00:42:34.024081 systemd[1]: Stopped ignition-files.service. May 17 00:42:34.051532 systemd[1]: Stopping ignition-mount.service... May 17 00:42:34.722000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.084037 systemd[1]: Stopping iscsid.service... May 17 00:42:34.737000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.098438 systemd[1]: kmod-static-nodes.service: Deactivated successfully. May 17 00:42:34.753000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.098774 systemd[1]: Stopped kmod-static-nodes.service. May 17 00:42:34.117453 systemd[1]: Stopping sysroot-boot.service... May 17 00:42:34.779000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.147487 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. May 17 00:42:34.802000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.147946 systemd[1]: Stopped systemd-udev-trigger.service. May 17 00:42:34.820000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.820000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:34.156782 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. May 17 00:42:34.156980 systemd[1]: Stopped dracut-pre-trigger.service. May 17 00:42:34.168569 systemd[1]: sysroot-boot.mount: Deactivated successfully. May 17 00:42:34.865000 audit: BPF prog-id=8 op=UNLOAD May 17 00:42:34.865000 audit: BPF prog-id=7 op=UNLOAD May 17 00:42:34.868000 audit: BPF prog-id=5 op=UNLOAD May 17 00:42:34.868000 audit: BPF prog-id=4 op=UNLOAD May 17 00:42:34.868000 audit: BPF prog-id=3 op=UNLOAD May 17 00:42:34.169975 systemd[1]: iscsid.service: Deactivated successfully. May 17 00:42:34.170117 systemd[1]: Stopped iscsid.service. May 17 00:42:34.181609 systemd[1]: ignition-mount.service: Deactivated successfully. May 17 00:42:34.899583 systemd-journald[189]: Received SIGTERM from PID 1 (n/a). May 17 00:42:34.181743 systemd[1]: Stopped ignition-mount.service. May 17 00:42:34.223209 systemd[1]: sysroot-boot.service: Deactivated successfully. May 17 00:42:34.223385 systemd[1]: Stopped sysroot-boot.service. May 17 00:42:34.231978 systemd[1]: initrd-cleanup.service: Deactivated successfully. May 17 00:42:34.232109 systemd[1]: Finished initrd-cleanup.service. May 17 00:42:34.251902 systemd[1]: ignition-disks.service: Deactivated successfully. May 17 00:42:34.251966 systemd[1]: Stopped ignition-disks.service. May 17 00:42:34.267641 systemd[1]: ignition-kargs.service: Deactivated successfully. May 17 00:42:34.267714 systemd[1]: Stopped ignition-kargs.service. May 17 00:42:34.282629 systemd[1]: ignition-fetch.service: Deactivated successfully. May 17 00:42:34.282716 systemd[1]: Stopped ignition-fetch.service. May 17 00:42:34.297616 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. May 17 00:42:34.297716 systemd[1]: Stopped ignition-fetch-offline.service. May 17 00:42:34.312622 systemd[1]: Stopped target paths.target. May 17 00:42:34.330422 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. May 17 00:42:34.334354 systemd[1]: Stopped systemd-ask-password-console.path. May 17 00:42:34.346395 systemd[1]: Stopped target slices.target. May 17 00:42:34.359428 systemd[1]: Stopped target sockets.target. May 17 00:42:34.372488 systemd[1]: iscsid.socket: Deactivated successfully. May 17 00:42:34.372587 systemd[1]: Closed iscsid.socket. May 17 00:42:34.386480 systemd[1]: ignition-setup.service: Deactivated successfully. May 17 00:42:34.386596 systemd[1]: Stopped ignition-setup.service. May 17 00:42:34.402550 systemd[1]: initrd-setup-root.service: Deactivated successfully. May 17 00:42:34.402633 systemd[1]: Stopped initrd-setup-root.service. May 17 00:42:34.418720 systemd[1]: Stopping iscsiuio.service... May 17 00:42:34.434922 systemd[1]: iscsiuio.service: Deactivated successfully. May 17 00:42:34.435062 systemd[1]: Stopped iscsiuio.service. May 17 00:42:34.454797 systemd[1]: Stopped target network.target. May 17 00:42:34.469603 systemd[1]: iscsiuio.socket: Deactivated successfully. May 17 00:42:34.469671 systemd[1]: Closed iscsiuio.socket. May 17 00:42:34.491806 systemd[1]: Stopping systemd-networkd.service... May 17 00:42:34.495334 systemd-networkd[680]: eth0: DHCPv6 lease lost May 17 00:42:34.907000 audit: BPF prog-id=9 op=UNLOAD May 17 00:42:34.506745 systemd[1]: Stopping systemd-resolved.service... May 17 00:42:34.532865 systemd[1]: systemd-resolved.service: Deactivated successfully. May 17 00:42:34.533022 systemd[1]: Stopped systemd-resolved.service. May 17 00:42:34.549463 systemd[1]: systemd-networkd.service: Deactivated successfully. May 17 00:42:34.549614 systemd[1]: Stopped systemd-networkd.service. May 17 00:42:34.564303 systemd[1]: systemd-networkd.socket: Deactivated successfully. May 17 00:42:34.564365 systemd[1]: Closed systemd-networkd.socket. May 17 00:42:34.579643 systemd[1]: Stopping network-cleanup.service... May 17 00:42:34.594459 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. May 17 00:42:34.594723 systemd[1]: Stopped parse-ip-for-networkd.service. May 17 00:42:34.610747 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 17 00:42:34.610827 systemd[1]: Stopped systemd-sysctl.service. May 17 00:42:34.625754 systemd[1]: systemd-modules-load.service: Deactivated successfully. May 17 00:42:34.625826 systemd[1]: Stopped systemd-modules-load.service. May 17 00:42:34.641819 systemd[1]: Stopping systemd-udevd.service... May 17 00:42:34.657234 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. May 17 00:42:34.658066 systemd[1]: systemd-udevd.service: Deactivated successfully. May 17 00:42:34.658274 systemd[1]: Stopped systemd-udevd.service. May 17 00:42:34.675063 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. May 17 00:42:34.675177 systemd[1]: Closed systemd-udevd-control.socket. May 17 00:42:34.690701 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. May 17 00:42:34.690770 systemd[1]: Closed systemd-udevd-kernel.socket. May 17 00:42:34.708537 systemd[1]: dracut-pre-udev.service: Deactivated successfully. May 17 00:42:34.708658 systemd[1]: Stopped dracut-pre-udev.service. May 17 00:42:34.723606 systemd[1]: dracut-cmdline.service: Deactivated successfully. May 17 00:42:34.723686 systemd[1]: Stopped dracut-cmdline.service. May 17 00:42:34.738702 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. May 17 00:42:34.738797 systemd[1]: Stopped dracut-cmdline-ask.service. May 17 00:42:34.755784 systemd[1]: Starting initrd-udevadm-cleanup-db.service... May 17 00:42:34.780419 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. May 17 00:42:34.780572 systemd[1]: Stopped systemd-vconsole-setup.service. May 17 00:42:34.781351 systemd[1]: network-cleanup.service: Deactivated successfully. May 17 00:42:34.781500 systemd[1]: Stopped network-cleanup.service. May 17 00:42:34.803914 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. May 17 00:42:34.804046 systemd[1]: Finished initrd-udevadm-cleanup-db.service. May 17 00:42:34.821746 systemd[1]: Reached target initrd-switch-root.target. May 17 00:42:34.840730 systemd[1]: Starting initrd-switch-root.service... May 17 00:42:34.865140 systemd[1]: Switching root. May 17 00:42:34.910724 systemd-journald[189]: Journal stopped May 17 00:42:40.138841 kernel: SELinux: Class mctp_socket not defined in policy. May 17 00:42:40.138979 kernel: SELinux: Class anon_inode not defined in policy. May 17 00:42:40.139027 kernel: SELinux: the above unknown classes and permissions will be allowed May 17 00:42:40.139065 kernel: SELinux: policy capability network_peer_controls=1 May 17 00:42:40.139103 kernel: SELinux: policy capability open_perms=1 May 17 00:42:40.139140 kernel: SELinux: policy capability extended_socket_class=1 May 17 00:42:40.139178 kernel: SELinux: policy capability always_check_network=0 May 17 00:42:40.139221 kernel: SELinux: policy capability cgroup_seclabel=1 May 17 00:42:40.139349 kernel: SELinux: policy capability nnp_nosuid_transition=1 May 17 00:42:40.139394 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 May 17 00:42:40.139431 kernel: SELinux: policy capability ioctl_skip_cloexec=0 May 17 00:42:40.139467 kernel: kauditd_printk_skb: 48 callbacks suppressed May 17 00:42:40.139505 kernel: audit: type=1403 audit(1747442555.330:85): auid=4294967295 ses=4294967295 lsm=selinux res=1 May 17 00:42:40.139545 systemd[1]: Successfully loaded SELinux policy in 117.379ms. May 17 00:42:40.139600 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.238ms. May 17 00:42:40.139633 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) May 17 00:42:40.139666 systemd[1]: Detected virtualization kvm. May 17 00:42:40.139697 systemd[1]: Detected architecture x86-64. May 17 00:42:40.139735 systemd[1]: Detected first boot. May 17 00:42:40.139774 systemd[1]: Initializing machine ID from VM UUID. May 17 00:42:40.139805 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). May 17 00:42:40.139845 kernel: audit: type=1400 audit(1747442555.687:86): avc: denied { associate } for pid=913 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" May 17 00:42:40.139885 kernel: audit: type=1300 audit(1747442555.687:86): arch=c000003e syscall=188 success=yes exit=0 a0=c0001896b2 a1=c00002cb40 a2=c00002aa40 a3=32 items=0 ppid=896 pid=913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:42:40.139919 kernel: audit: type=1327 audit(1747442555.687:86): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 May 17 00:42:40.139950 kernel: audit: type=1400 audit(1747442555.698:87): avc: denied { associate } for pid=913 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 May 17 00:42:40.139982 kernel: audit: type=1300 audit(1747442555.698:87): arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c000189789 a2=1ed a3=0 items=2 ppid=896 pid=913 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:42:40.140017 kernel: audit: type=1307 audit(1747442555.698:87): cwd="/" May 17 00:42:40.140049 kernel: audit: type=1302 audit(1747442555.698:87): item=0 name=(null) inode=2 dev=00:2a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:40.140079 kernel: audit: type=1302 audit(1747442555.698:87): item=1 name=(null) inode=3 dev=00:2a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:40.140110 kernel: audit: type=1327 audit(1747442555.698:87): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 May 17 00:42:40.140143 systemd[1]: Populated /etc with preset unit settings. May 17 00:42:40.140177 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:42:40.140222 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:42:40.140272 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:42:40.140303 systemd[1]: Queued start job for default target multi-user.target. May 17 00:42:40.140335 systemd[1]: Unnecessary job was removed for dev-sda6.device. May 17 00:42:40.140366 systemd[1]: Created slice system-addon\x2dconfig.slice. May 17 00:42:40.140399 systemd[1]: Created slice system-addon\x2drun.slice. May 17 00:42:40.140439 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. May 17 00:42:40.140476 systemd[1]: Created slice system-getty.slice. May 17 00:42:40.140508 systemd[1]: Created slice system-modprobe.slice. May 17 00:42:40.140539 systemd[1]: Created slice system-serial\x2dgetty.slice. May 17 00:42:40.140573 systemd[1]: Created slice system-system\x2dcloudinit.slice. May 17 00:42:40.140603 systemd[1]: Created slice system-systemd\x2dfsck.slice. May 17 00:42:40.140633 systemd[1]: Created slice user.slice. May 17 00:42:40.140666 systemd[1]: Started systemd-ask-password-console.path. May 17 00:42:40.140697 systemd[1]: Started systemd-ask-password-wall.path. May 17 00:42:40.140733 systemd[1]: Set up automount boot.automount. May 17 00:42:40.140764 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. May 17 00:42:40.140796 systemd[1]: Reached target integritysetup.target. May 17 00:42:40.140830 systemd[1]: Reached target remote-cryptsetup.target. May 17 00:42:40.140862 systemd[1]: Reached target remote-fs.target. May 17 00:42:40.140902 systemd[1]: Reached target slices.target. May 17 00:42:40.140934 systemd[1]: Reached target swap.target. May 17 00:42:40.140965 systemd[1]: Reached target torcx.target. May 17 00:42:40.140998 systemd[1]: Reached target veritysetup.target. May 17 00:42:40.141033 systemd[1]: Listening on systemd-coredump.socket. May 17 00:42:40.141066 systemd[1]: Listening on systemd-initctl.socket. May 17 00:42:40.141097 systemd[1]: Listening on systemd-journald-audit.socket. May 17 00:42:40.141130 systemd[1]: Listening on systemd-journald-dev-log.socket. May 17 00:42:40.141171 systemd[1]: Listening on systemd-journald.socket. May 17 00:42:40.141209 systemd[1]: Listening on systemd-networkd.socket. May 17 00:42:40.141241 systemd[1]: Listening on systemd-udevd-control.socket. May 17 00:42:40.141293 systemd[1]: Listening on systemd-udevd-kernel.socket. May 17 00:42:40.141332 systemd[1]: Listening on systemd-userdbd.socket. May 17 00:42:40.141367 systemd[1]: Mounting dev-hugepages.mount... May 17 00:42:40.141411 systemd[1]: Mounting dev-mqueue.mount... May 17 00:42:40.141443 systemd[1]: Mounting media.mount... May 17 00:42:40.141485 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:40.141524 systemd[1]: Mounting sys-kernel-debug.mount... May 17 00:42:40.141558 systemd[1]: Mounting sys-kernel-tracing.mount... May 17 00:42:40.141596 systemd[1]: Mounting tmp.mount... May 17 00:42:40.141635 systemd[1]: Starting flatcar-tmpfiles.service... May 17 00:42:40.141668 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:42:40.141706 systemd[1]: Starting kmod-static-nodes.service... May 17 00:42:40.141743 systemd[1]: Starting modprobe@configfs.service... May 17 00:42:40.141780 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:42:40.141812 systemd[1]: Starting modprobe@drm.service... May 17 00:42:40.141858 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:42:40.141898 systemd[1]: Starting modprobe@fuse.service... May 17 00:42:40.141929 systemd[1]: Starting modprobe@loop.service... May 17 00:42:40.141967 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). May 17 00:42:40.141999 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. May 17 00:42:40.142032 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) May 17 00:42:40.142065 kernel: fuse: init (API version 7.34) May 17 00:42:40.142095 systemd[1]: Starting systemd-journald.service... May 17 00:42:40.142129 systemd[1]: Starting systemd-modules-load.service... May 17 00:42:40.142175 kernel: loop: module loaded May 17 00:42:40.142209 systemd[1]: Starting systemd-network-generator.service... May 17 00:42:40.142242 systemd[1]: Starting systemd-remount-fs.service... May 17 00:42:40.142288 systemd[1]: Starting systemd-udev-trigger.service... May 17 00:42:40.142322 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:40.142370 systemd-journald[1022]: Journal started May 17 00:42:40.142510 systemd-journald[1022]: Runtime Journal (/run/log/journal/1478b8602b6203bc8936d3dde1dce9ce) is 8.0M, max 148.8M, 140.8M free. May 17 00:42:39.643000 audit[1]: AVC avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 May 17 00:42:39.643000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 May 17 00:42:40.134000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 May 17 00:42:40.134000 audit[1022]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=6 a1=7ffcdc332db0 a2=4000 a3=7ffcdc332e4c items=0 ppid=1 pid=1022 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:42:40.134000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" May 17 00:42:40.158322 systemd[1]: Started systemd-journald.service. May 17 00:42:40.165000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.169155 systemd[1]: Mounted dev-hugepages.mount. May 17 00:42:40.176657 systemd[1]: Mounted dev-mqueue.mount. May 17 00:42:40.183657 systemd[1]: Mounted media.mount. May 17 00:42:40.190692 systemd[1]: Mounted sys-kernel-debug.mount. May 17 00:42:40.199679 systemd[1]: Mounted sys-kernel-tracing.mount. May 17 00:42:40.208685 systemd[1]: Mounted tmp.mount. May 17 00:42:40.216261 systemd[1]: Finished flatcar-tmpfiles.service. May 17 00:42:40.223000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.225449 systemd[1]: Finished kmod-static-nodes.service. May 17 00:42:40.233000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.235069 systemd[1]: modprobe@configfs.service: Deactivated successfully. May 17 00:42:40.235476 systemd[1]: Finished modprobe@configfs.service. May 17 00:42:40.244000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.244000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.246402 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:42:40.246764 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:42:40.254000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.254000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.256138 systemd[1]: modprobe@drm.service: Deactivated successfully. May 17 00:42:40.256444 systemd[1]: Finished modprobe@drm.service. May 17 00:42:40.264000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.264000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.266069 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:42:40.266395 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:42:40.274000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.274000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.275993 systemd[1]: modprobe@fuse.service: Deactivated successfully. May 17 00:42:40.276320 systemd[1]: Finished modprobe@fuse.service. May 17 00:42:40.283000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.283000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.284964 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:42:40.285272 systemd[1]: Finished modprobe@loop.service. May 17 00:42:40.292000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.292000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.294054 systemd[1]: Finished systemd-modules-load.service. May 17 00:42:40.301000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.303058 systemd[1]: Finished systemd-network-generator.service. May 17 00:42:40.310000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.312031 systemd[1]: Finished systemd-remount-fs.service. May 17 00:42:40.319000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.320994 systemd[1]: Finished systemd-udev-trigger.service. May 17 00:42:40.328000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.330133 systemd[1]: Reached target network-pre.target. May 17 00:42:40.340090 systemd[1]: Mounting sys-fs-fuse-connections.mount... May 17 00:42:40.351447 systemd[1]: Mounting sys-kernel-config.mount... May 17 00:42:40.358457 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). May 17 00:42:40.362383 systemd[1]: Starting systemd-hwdb-update.service... May 17 00:42:40.371818 systemd[1]: Starting systemd-journal-flush.service... May 17 00:42:40.381368 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:42:40.383856 systemd[1]: Starting systemd-random-seed.service... May 17 00:42:40.391484 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:42:40.393867 systemd[1]: Starting systemd-sysctl.service... May 17 00:42:40.401311 systemd-journald[1022]: Time spent on flushing to /var/log/journal/1478b8602b6203bc8936d3dde1dce9ce is 91.520ms for 1103 entries. May 17 00:42:40.401311 systemd-journald[1022]: System Journal (/var/log/journal/1478b8602b6203bc8936d3dde1dce9ce) is 8.0M, max 584.8M, 576.8M free. May 17 00:42:40.528512 systemd-journald[1022]: Received client request to flush runtime journal. May 17 00:42:40.528628 kernel: kauditd_printk_skb: 24 callbacks suppressed May 17 00:42:40.528679 kernel: audit: type=1130 audit(1747442560.456:110): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.528730 kernel: audit: type=1130 audit(1747442560.502:111): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.456000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.502000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.403958 systemd[1]: Starting systemd-sysusers.service... May 17 00:42:40.418653 systemd[1]: Starting systemd-udev-settle.service... May 17 00:42:40.534391 udevadm[1049]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. May 17 00:42:40.430750 systemd[1]: Mounted sys-fs-fuse-connections.mount. May 17 00:42:40.439598 systemd[1]: Mounted sys-kernel-config.mount. May 17 00:42:40.448952 systemd[1]: Finished systemd-random-seed.service. May 17 00:42:40.461665 systemd[1]: Reached target first-boot-complete.target. May 17 00:42:40.495199 systemd[1]: Finished systemd-sysctl.service. May 17 00:42:40.531665 systemd[1]: Finished systemd-journal-flush.service. May 17 00:42:40.541188 systemd[1]: Finished systemd-sysusers.service. May 17 00:42:40.539000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.568437 kernel: audit: type=1130 audit(1747442560.539:112): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.571000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.575237 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... May 17 00:42:40.600390 kernel: audit: type=1130 audit(1747442560.571:113): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.665580 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. May 17 00:42:40.673000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:40.698306 kernel: audit: type=1130 audit(1747442560.673:114): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.256550 systemd[1]: Finished systemd-hwdb-update.service. May 17 00:42:41.264000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.268231 systemd[1]: Starting systemd-udevd.service... May 17 00:42:41.288375 kernel: audit: type=1130 audit(1747442561.264:115): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.313878 systemd-udevd[1060]: Using default interface naming scheme 'v252'. May 17 00:42:41.365611 systemd[1]: Started systemd-udevd.service. May 17 00:42:41.401519 kernel: audit: type=1130 audit(1747442561.373:116): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.373000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.401214 systemd[1]: Starting systemd-networkd.service... May 17 00:42:41.416839 systemd[1]: Starting systemd-userdbd.service... May 17 00:42:41.487720 systemd[1]: Found device dev-ttyS0.device. May 17 00:42:41.505348 systemd[1]: Started systemd-userdbd.service. May 17 00:42:41.512000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.536282 kernel: audit: type=1130 audit(1747442561.512:117): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.694280 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 May 17 00:42:41.699752 systemd-networkd[1079]: lo: Link UP May 17 00:42:41.700453 systemd-networkd[1079]: lo: Gained carrier May 17 00:42:41.701496 systemd-networkd[1079]: Enumeration completed May 17 00:42:41.701743 systemd[1]: Started systemd-networkd.service. May 17 00:42:41.702498 systemd-networkd[1079]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. May 17 00:42:41.704951 systemd-networkd[1079]: eth0: Link UP May 17 00:42:41.705128 systemd-networkd[1079]: eth0: Gained carrier May 17 00:42:41.709000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.715550 systemd-networkd[1079]: eth0: Overlong DHCP hostname received, shortened from 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9.c.flatcar-212911.internal' to 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9' May 17 00:42:41.715583 systemd-networkd[1079]: eth0: DHCPv4 address 10.128.0.10/32, gateway 10.128.0.1 acquired from 169.254.169.254 May 17 00:42:41.737555 kernel: audit: type=1130 audit(1747442561.709:118): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.817288 kernel: ACPI: button: Power Button [PWRF] May 17 00:42:41.832315 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input3 May 17 00:42:41.834586 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. May 17 00:42:41.770000 audit[1078]: AVC avc: denied { confidentiality } for pid=1078 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 May 17 00:42:41.860272 kernel: audit: type=1400 audit(1747442561.770:119): avc: denied { confidentiality } for pid=1078 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 May 17 00:42:41.866827 kernel: ACPI: button: Sleep Button [SLPF] May 17 00:42:41.770000 audit[1078]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=562fce9d69b0 a1=338ac a2=7fdb67c1bbc5 a3=5 items=110 ppid=1060 pid=1078 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:42:41.770000 audit: CWD cwd="/" May 17 00:42:41.770000 audit: PATH item=0 name=(null) inode=40 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=1 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=2 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.892355 kernel: EDAC MC: Ver: 3.0.0 May 17 00:42:41.770000 audit: PATH item=3 name=(null) inode=12974 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=4 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=5 name=(null) inode=12975 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=6 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=7 name=(null) inode=12976 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=8 name=(null) inode=12976 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=9 name=(null) inode=12977 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=10 name=(null) inode=12976 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=11 name=(null) inode=12978 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=12 name=(null) inode=12976 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=13 name=(null) inode=12979 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=14 name=(null) inode=12976 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=15 name=(null) inode=12980 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=16 name=(null) inode=12976 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=17 name=(null) inode=12981 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=18 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=19 name=(null) inode=12982 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=20 name=(null) inode=12982 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=21 name=(null) inode=12983 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=22 name=(null) inode=12982 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=23 name=(null) inode=12984 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=24 name=(null) inode=12982 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=25 name=(null) inode=12985 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=26 name=(null) inode=12982 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=27 name=(null) inode=12986 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=28 name=(null) inode=12982 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=29 name=(null) inode=12987 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=30 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=31 name=(null) inode=12988 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=32 name=(null) inode=12988 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=33 name=(null) inode=12989 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=34 name=(null) inode=12988 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=35 name=(null) inode=12990 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=36 name=(null) inode=12988 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=37 name=(null) inode=12991 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=38 name=(null) inode=12988 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=39 name=(null) inode=12992 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=40 name=(null) inode=12988 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=41 name=(null) inode=12993 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=42 name=(null) inode=12973 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=43 name=(null) inode=12994 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=44 name=(null) inode=12994 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=45 name=(null) inode=12995 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=46 name=(null) inode=12994 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=47 name=(null) inode=12996 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=48 name=(null) inode=12994 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=49 name=(null) inode=12997 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=50 name=(null) inode=12994 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=51 name=(null) inode=12998 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=52 name=(null) inode=12994 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=53 name=(null) inode=12999 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=54 name=(null) inode=40 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=55 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=56 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=57 name=(null) inode=13001 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=58 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=59 name=(null) inode=13002 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=60 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=61 name=(null) inode=13003 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=62 name=(null) inode=13003 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=63 name=(null) inode=13004 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=64 name=(null) inode=13003 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=65 name=(null) inode=13005 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=66 name=(null) inode=13003 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=67 name=(null) inode=13006 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=68 name=(null) inode=13003 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=69 name=(null) inode=13007 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=70 name=(null) inode=13003 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=71 name=(null) inode=13008 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=72 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=73 name=(null) inode=13009 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=74 name=(null) inode=13009 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=75 name=(null) inode=13010 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=76 name=(null) inode=13009 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=77 name=(null) inode=13011 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=78 name=(null) inode=13009 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=79 name=(null) inode=13012 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=80 name=(null) inode=13009 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=81 name=(null) inode=13013 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=82 name=(null) inode=13009 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=83 name=(null) inode=13014 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=84 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=85 name=(null) inode=13015 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=86 name=(null) inode=13015 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=87 name=(null) inode=13016 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=88 name=(null) inode=13015 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=89 name=(null) inode=13017 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=90 name=(null) inode=13015 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=91 name=(null) inode=13018 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=92 name=(null) inode=13015 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=93 name=(null) inode=13019 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=94 name=(null) inode=13015 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=95 name=(null) inode=13020 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=96 name=(null) inode=13000 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=97 name=(null) inode=13021 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=98 name=(null) inode=13021 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=99 name=(null) inode=13022 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=100 name=(null) inode=13021 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=101 name=(null) inode=13023 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=102 name=(null) inode=13021 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=103 name=(null) inode=13024 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=104 name=(null) inode=13021 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=105 name=(null) inode=13025 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=106 name=(null) inode=13021 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=107 name=(null) inode=13026 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PATH item=109 name=(null) inode=13030 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 May 17 00:42:41.770000 audit: PROCTITLE proctitle="(udev-worker)" May 17 00:42:41.909312 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr May 17 00:42:41.933778 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 May 17 00:42:41.951300 kernel: mousedev: PS/2 mouse device common for all mice May 17 00:42:41.969059 systemd[1]: Finished systemd-udev-settle.service. May 17 00:42:41.976000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:41.979597 systemd[1]: Starting lvm2-activation-early.service... May 17 00:42:42.010146 lvm[1098]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 17 00:42:42.047218 systemd[1]: Finished lvm2-activation-early.service. May 17 00:42:42.055000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.056897 systemd[1]: Reached target cryptsetup.target. May 17 00:42:42.067357 systemd[1]: Starting lvm2-activation.service... May 17 00:42:42.075201 lvm[1100]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. May 17 00:42:42.104065 systemd[1]: Finished lvm2-activation.service. May 17 00:42:42.112000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.113827 systemd[1]: Reached target local-fs-pre.target. May 17 00:42:42.122438 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). May 17 00:42:42.122496 systemd[1]: Reached target local-fs.target. May 17 00:42:42.131436 systemd[1]: Reached target machines.target. May 17 00:42:42.141338 systemd[1]: Starting ldconfig.service... May 17 00:42:42.149339 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:42:42.149477 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:42.152136 systemd[1]: Starting systemd-boot-update.service... May 17 00:42:42.164583 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... May 17 00:42:42.176620 systemd[1]: Starting systemd-machine-id-commit.service... May 17 00:42:42.187777 systemd[1]: Starting systemd-sysext.service... May 17 00:42:42.196321 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1103 (bootctl) May 17 00:42:42.199594 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... May 17 00:42:42.220360 systemd[1]: Unmounting usr-share-oem.mount... May 17 00:42:42.228442 systemd[1]: usr-share-oem.mount: Deactivated successfully. May 17 00:42:42.229032 systemd[1]: Unmounted usr-share-oem.mount. May 17 00:42:42.233000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.234532 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. May 17 00:42:42.264966 kernel: loop0: detected capacity change from 0 to 221472 May 17 00:42:42.384316 systemd-fsck[1115]: fsck.fat 4.2 (2021-01-31) May 17 00:42:42.384316 systemd-fsck[1115]: /dev/sda1: 790 files, 120726/258078 clusters May 17 00:42:42.387953 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. May 17 00:42:42.397000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.401153 systemd[1]: Mounting boot.mount... May 17 00:42:42.446187 systemd[1]: Mounted boot.mount. May 17 00:42:42.478794 systemd[1]: Finished systemd-boot-update.service. May 17 00:42:42.486000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.748282 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher May 17 00:42:42.759782 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. May 17 00:42:42.762010 systemd[1]: Finished systemd-machine-id-commit.service. May 17 00:42:42.769000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.781410 kernel: loop1: detected capacity change from 0 to 221472 May 17 00:42:42.813660 (sd-sysext)[1125]: Using extensions 'kubernetes'. May 17 00:42:42.814482 (sd-sysext)[1125]: Merged extensions into '/usr'. May 17 00:42:42.855432 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:42.858648 ldconfig[1102]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. May 17 00:42:42.859513 systemd[1]: Mounting usr-share-oem.mount... May 17 00:42:42.867162 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:42:42.871054 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:42:42.873444 systemd-networkd[1079]: eth0: Gained IPv6LL May 17 00:42:42.884990 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:42:42.894946 systemd[1]: Starting modprobe@loop.service... May 17 00:42:42.902571 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:42:42.902839 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:42.903053 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:42.908977 systemd[1]: Finished ldconfig.service. May 17 00:42:42.915000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.917095 systemd[1]: Mounted usr-share-oem.mount. May 17 00:42:42.925224 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:42:42.925594 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:42:42.933000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.933000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.935146 systemd[1]: Finished systemd-sysext.service. May 17 00:42:42.942000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.943932 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:42:42.944219 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:42:42.952000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.952000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.953993 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:42:42.954483 systemd[1]: Finished modprobe@loop.service. May 17 00:42:42.962000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.962000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:42.966829 systemd[1]: Starting ensure-sysext.service... May 17 00:42:42.974523 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:42:42.974670 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:42:42.976672 systemd[1]: Starting systemd-tmpfiles-setup.service... May 17 00:42:42.989338 systemd[1]: Reloading. May 17 00:42:42.999241 systemd-tmpfiles[1140]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. May 17 00:42:43.001538 systemd-tmpfiles[1140]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. May 17 00:42:43.005750 systemd-tmpfiles[1140]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. May 17 00:42:43.113109 /usr/lib/systemd/system-generators/torcx-generator[1159]: time="2025-05-17T00:42:43Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 17 00:42:43.113166 /usr/lib/systemd/system-generators/torcx-generator[1159]: time="2025-05-17T00:42:43Z" level=info msg="torcx already run" May 17 00:42:43.329681 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:42:43.329713 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:42:43.354810 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:42:43.452815 systemd[1]: Finished systemd-tmpfiles-setup.service. May 17 00:42:43.461000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:43.468673 systemd[1]: Starting audit-rules.service... May 17 00:42:43.479216 systemd[1]: Starting clean-ca-certificates.service... May 17 00:42:43.490224 systemd[1]: Starting oem-gce-enable-oslogin.service... May 17 00:42:43.501339 systemd[1]: Starting systemd-journal-catalog-update.service... May 17 00:42:43.514427 systemd[1]: Starting systemd-resolved.service... May 17 00:42:43.525801 systemd[1]: Starting systemd-timesyncd.service... May 17 00:42:43.536695 systemd[1]: Starting systemd-update-utmp.service... May 17 00:42:43.545000 audit[1235]: SYSTEM_BOOT pid=1235 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' May 17 00:42:43.546159 systemd[1]: Finished clean-ca-certificates.service. May 17 00:42:43.553000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:43.555559 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. May 17 00:42:43.556005 systemd[1]: Finished oem-gce-enable-oslogin.service. May 17 00:42:43.563000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=oem-gce-enable-oslogin comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:43.563000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=oem-gce-enable-oslogin comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' May 17 00:42:43.573489 augenrules[1243]: No rules May 17 00:42:43.572000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 May 17 00:42:43.572000 audit[1243]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe98fd9680 a2=420 a3=0 items=0 ppid=1211 pid=1243 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) May 17 00:42:43.572000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 May 17 00:42:43.578887 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:43.579661 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:42:43.585555 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:42:43.594869 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:42:43.605060 systemd[1]: Starting modprobe@loop.service... May 17 00:42:43.615110 systemd[1]: Starting oem-gce-enable-oslogin.service... May 17 00:42:43.623525 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:42:43.623879 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:43.624181 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:42:43.624434 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:43.627443 systemd[1]: Finished audit-rules.service. May 17 00:42:43.635661 systemd[1]: Finished systemd-journal-catalog-update.service. May 17 00:42:43.640943 enable-oslogin[1256]: /etc/pam.d/sshd already exists. Not enabling OS Login May 17 00:42:43.646833 systemd[1]: Finished systemd-update-utmp.service. May 17 00:42:43.656473 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:42:43.656826 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:42:43.666206 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:42:43.666559 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:42:43.676377 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:42:43.676720 systemd[1]: Finished modprobe@loop.service. May 17 00:42:43.686224 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. May 17 00:42:43.686677 systemd[1]: Finished oem-gce-enable-oslogin.service. May 17 00:42:43.698021 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:42:43.698284 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:42:43.701389 systemd[1]: Starting systemd-update-done.service... May 17 00:42:43.712167 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:43.712766 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:42:43.716594 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:42:43.725949 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:42:43.735892 systemd[1]: Starting modprobe@loop.service... May 17 00:42:43.746554 systemd[1]: Starting oem-gce-enable-oslogin.service... May 17 00:42:43.755477 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:42:43.755792 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:43.756067 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:42:43.756293 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:43.756971 enable-oslogin[1268]: /etc/pam.d/sshd already exists. Not enabling OS Login May 17 00:42:43.759337 systemd[1]: Finished systemd-update-done.service. May 17 00:42:43.768621 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:42:43.768940 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:42:43.778463 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:42:43.778810 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:42:43.788457 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:42:43.788822 systemd[1]: Finished modprobe@loop.service. May 17 00:42:43.798387 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. May 17 00:42:43.798800 systemd[1]: Finished oem-gce-enable-oslogin.service. May 17 00:42:43.808241 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:42:43.808474 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:42:43.814089 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:43.815719 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. May 17 00:42:43.818418 systemd[1]: Starting modprobe@dm_mod.service... May 17 00:42:43.828132 systemd[1]: Starting modprobe@drm.service... May 17 00:42:43.839219 systemd[1]: Starting modprobe@efi_pstore.service... May 17 00:42:43.850344 systemd[1]: Starting modprobe@loop.service... May 17 00:42:43.860123 systemd[1]: Starting oem-gce-enable-oslogin.service... May 17 00:42:43.868604 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. May 17 00:42:43.868896 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:43.869741 systemd-resolved[1225]: Positive Trust Anchors: May 17 00:42:43.870283 systemd-resolved[1225]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d May 17 00:42:43.870503 systemd-resolved[1225]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test May 17 00:42:43.872030 systemd[1]: Starting systemd-networkd-wait-online.service... May 17 00:42:44.275998 systemd-timesyncd[1231]: Contacted time server 169.254.169.254:123 (169.254.169.254). May 17 00:42:44.276114 systemd-timesyncd[1231]: Initial clock synchronization to Sat 2025-05-17 00:42:44.275882 UTC. May 17 00:42:44.277627 enable-oslogin[1280]: /etc/pam.d/sshd already exists. Not enabling OS Login May 17 00:42:44.282596 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). May 17 00:42:44.282869 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). May 17 00:42:44.285703 systemd[1]: Started systemd-timesyncd.service. May 17 00:42:44.296585 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. May 17 00:42:44.296908 systemd[1]: Finished modprobe@dm_mod.service. May 17 00:42:44.306369 systemd[1]: modprobe@drm.service: Deactivated successfully. May 17 00:42:44.306710 systemd[1]: Finished modprobe@drm.service. May 17 00:42:44.316641 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. May 17 00:42:44.316964 systemd[1]: Finished modprobe@efi_pstore.service. May 17 00:42:44.326397 systemd[1]: modprobe@loop.service: Deactivated successfully. May 17 00:42:44.326724 systemd[1]: Finished modprobe@loop.service. May 17 00:42:44.327612 systemd-resolved[1225]: Defaulting to hostname 'linux'. May 17 00:42:44.336448 systemd[1]: Started systemd-resolved.service. May 17 00:42:44.345244 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. May 17 00:42:44.345677 systemd[1]: Finished oem-gce-enable-oslogin.service. May 17 00:42:44.355433 systemd[1]: Finished systemd-networkd-wait-online.service. May 17 00:42:44.366757 systemd[1]: Reached target network.target. May 17 00:42:44.375585 systemd[1]: Reached target network-online.target. May 17 00:42:44.384528 systemd[1]: Reached target nss-lookup.target. May 17 00:42:44.393524 systemd[1]: Reached target time-set.target. May 17 00:42:44.401549 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). May 17 00:42:44.401622 systemd[1]: Reached target sysinit.target. May 17 00:42:44.410641 systemd[1]: Started motdgen.path. May 17 00:42:44.417598 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. May 17 00:42:44.427825 systemd[1]: Started logrotate.timer. May 17 00:42:44.434713 systemd[1]: Started mdadm.timer. May 17 00:42:44.442554 systemd[1]: Started systemd-tmpfiles-clean.timer. May 17 00:42:44.451569 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). May 17 00:42:44.451654 systemd[1]: Reached target paths.target. May 17 00:42:44.458554 systemd[1]: Reached target timers.target. May 17 00:42:44.466435 systemd[1]: Listening on dbus.socket. May 17 00:42:44.475337 systemd[1]: Starting docker.socket... May 17 00:42:44.485002 systemd[1]: Listening on sshd.socket. May 17 00:42:44.492657 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:44.492778 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. May 17 00:42:44.493924 systemd[1]: Finished ensure-sysext.service. May 17 00:42:44.502869 systemd[1]: Listening on docker.socket. May 17 00:42:44.510685 systemd[1]: Reached target sockets.target. May 17 00:42:44.519525 systemd[1]: Reached target basic.target. May 17 00:42:44.526850 systemd[1]: System is tainted: cgroupsv1 May 17 00:42:44.526962 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. May 17 00:42:44.527007 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. May 17 00:42:44.528905 systemd[1]: Starting containerd.service... May 17 00:42:44.538651 systemd[1]: Starting coreos-metadata-sshkeys@core.service... May 17 00:42:44.550882 systemd[1]: Starting dbus.service... May 17 00:42:44.558983 systemd[1]: Starting enable-oem-cloudinit.service... May 17 00:42:44.569142 systemd[1]: Starting extend-filesystems.service... May 17 00:42:44.575696 jq[1292]: false May 17 00:42:44.576524 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). May 17 00:42:44.579556 systemd[1]: Starting kubelet.service... May 17 00:42:44.590204 systemd[1]: Starting motdgen.service... May 17 00:42:44.598036 systemd[1]: Starting oem-gce.service... May 17 00:42:44.607856 systemd[1]: Starting prepare-helm.service... May 17 00:42:44.617055 systemd[1]: Starting ssh-key-proc-cmdline.service... May 17 00:42:44.627940 systemd[1]: Starting sshd-keygen.service... May 17 00:42:44.639994 systemd[1]: Starting systemd-logind.service... May 17 00:42:44.647577 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). May 17 00:42:44.647745 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionSecurity=!tpm2). May 17 00:42:44.650756 systemd[1]: Starting update-engine.service... May 17 00:42:44.659935 systemd[1]: Starting update-ssh-keys-after-ignition.service... May 17 00:42:44.674895 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. May 17 00:42:44.675437 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. May 17 00:42:44.687145 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. May 17 00:42:44.687724 systemd[1]: Finished ssh-key-proc-cmdline.service. May 17 00:42:44.720432 jq[1313]: true May 17 00:42:44.724568 mkfs.ext4[1328]: mke2fs 1.46.5 (30-Dec-2021) May 17 00:42:44.731943 mkfs.ext4[1328]: Discarding device blocks: 0/262144\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008 \u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008done May 17 00:42:44.731943 mkfs.ext4[1328]: Creating filesystem with 262144 4k blocks and 65536 inodes May 17 00:42:44.731943 mkfs.ext4[1328]: Filesystem UUID: 089c6783-a483-4460-a146-71339674bf1a May 17 00:42:44.731943 mkfs.ext4[1328]: Superblock backups stored on blocks: May 17 00:42:44.731943 mkfs.ext4[1328]: 32768, 98304, 163840, 229376 May 17 00:42:44.731943 mkfs.ext4[1328]: Allocating group tables: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done May 17 00:42:44.731943 mkfs.ext4[1328]: Writing inode tables: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done May 17 00:42:44.742400 mkfs.ext4[1328]: Creating journal (8192 blocks): done May 17 00:42:44.752908 mkfs.ext4[1328]: Writing superblocks and filesystem accounting information: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done May 17 00:42:44.791646 jq[1330]: true May 17 00:42:44.792104 extend-filesystems[1293]: Found loop1 May 17 00:42:44.792104 extend-filesystems[1293]: Found sda May 17 00:42:44.792104 extend-filesystems[1293]: Found sda1 May 17 00:42:44.792104 extend-filesystems[1293]: Found sda2 May 17 00:42:44.792104 extend-filesystems[1293]: Found sda3 May 17 00:42:44.792104 extend-filesystems[1293]: Found usr May 17 00:42:44.839668 extend-filesystems[1293]: Found sda4 May 17 00:42:44.839668 extend-filesystems[1293]: Found sda6 May 17 00:42:44.839668 extend-filesystems[1293]: Found sda7 May 17 00:42:44.839668 extend-filesystems[1293]: Found sda9 May 17 00:42:44.839668 extend-filesystems[1293]: Checking size of /dev/sda9 May 17 00:42:44.877085 systemd[1]: motdgen.service: Deactivated successfully. May 17 00:42:44.891591 umount[1338]: umount: /var/lib/flatcar-oem-gce.img: not mounted. May 17 00:42:44.877555 systemd[1]: Finished motdgen.service. May 17 00:42:44.895424 tar[1323]: linux-amd64/helm May 17 00:42:44.905385 kernel: loop2: detected capacity change from 0 to 2097152 May 17 00:42:44.911915 extend-filesystems[1293]: Resized partition /dev/sda9 May 17 00:42:44.921962 extend-filesystems[1359]: resize2fs 1.46.5 (30-Dec-2021) May 17 00:42:44.939386 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 2538491 blocks May 17 00:42:44.948989 update_engine[1312]: I0517 00:42:44.948915 1312 main.cc:92] Flatcar Update Engine starting May 17 00:42:44.957824 dbus-daemon[1291]: [system] SELinux support is enabled May 17 00:42:44.960063 systemd[1]: Started dbus.service. May 17 00:42:44.970606 dbus-daemon[1291]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1079 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") May 17 00:42:44.972084 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). May 17 00:42:44.972139 systemd[1]: Reached target system-config.target. May 17 00:42:44.980615 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). May 17 00:42:44.980667 systemd[1]: Reached target user-config.target. May 17 00:42:44.996944 update_engine[1312]: I0517 00:42:44.996760 1312 update_check_scheduler.cc:74] Next update check in 7m34s May 17 00:42:45.013506 kernel: EXT4-fs (sda9): resized filesystem to 2538491 May 17 00:42:45.019457 kernel: EXT4-fs (loop2): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. May 17 00:42:45.019538 extend-filesystems[1359]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required May 17 00:42:45.019538 extend-filesystems[1359]: old_desc_blocks = 1, new_desc_blocks = 2 May 17 00:42:45.019538 extend-filesystems[1359]: The filesystem on /dev/sda9 is now 2538491 (4k) blocks long. May 17 00:42:45.017514 systemd[1]: Started update-engine.service. May 17 00:42:45.088779 extend-filesystems[1293]: Resized filesystem in /dev/sda9 May 17 00:42:45.021724 dbus-daemon[1291]: [system] Successfully activated service 'org.freedesktop.systemd1' May 17 00:42:45.026288 systemd[1]: extend-filesystems.service: Deactivated successfully. May 17 00:42:45.026734 systemd[1]: Finished extend-filesystems.service. May 17 00:42:45.041860 systemd[1]: Started locksmithd.service. May 17 00:42:45.060828 systemd[1]: Starting systemd-hostnamed.service... May 17 00:42:45.129384 bash[1369]: Updated "/home/core/.ssh/authorized_keys" May 17 00:42:45.134072 systemd[1]: Finished update-ssh-keys-after-ignition.service. May 17 00:42:45.157692 env[1327]: time="2025-05-17T00:42:45.157561951Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 May 17 00:42:45.202408 coreos-metadata[1290]: May 17 00:42:45.202 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/sshKeys: Attempt #1 May 17 00:42:45.207419 coreos-metadata[1290]: May 17 00:42:45.207 INFO Fetch failed with 404: resource not found May 17 00:42:45.207419 coreos-metadata[1290]: May 17 00:42:45.207 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/ssh-keys: Attempt #1 May 17 00:42:45.207986 coreos-metadata[1290]: May 17 00:42:45.207 INFO Fetch successful May 17 00:42:45.207986 coreos-metadata[1290]: May 17 00:42:45.207 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/block-project-ssh-keys: Attempt #1 May 17 00:42:45.209468 coreos-metadata[1290]: May 17 00:42:45.209 INFO Fetch failed with 404: resource not found May 17 00:42:45.209754 coreos-metadata[1290]: May 17 00:42:45.209 INFO Fetching http://169.254.169.254/computeMetadata/v1/project/attributes/sshKeys: Attempt #1 May 17 00:42:45.211253 coreos-metadata[1290]: May 17 00:42:45.211 INFO Fetch failed with 404: resource not found May 17 00:42:45.211567 coreos-metadata[1290]: May 17 00:42:45.211 INFO Fetching http://169.254.169.254/computeMetadata/v1/project/attributes/ssh-keys: Attempt #1 May 17 00:42:45.212603 coreos-metadata[1290]: May 17 00:42:45.212 INFO Fetch successful May 17 00:42:45.215077 unknown[1290]: wrote ssh authorized keys file for user: core May 17 00:42:45.278698 update-ssh-keys[1383]: Updated "/home/core/.ssh/authorized_keys" May 17 00:42:45.280165 systemd[1]: Finished coreos-metadata-sshkeys@core.service. May 17 00:42:45.343015 systemd-logind[1310]: Watching system buttons on /dev/input/event1 (Power Button) May 17 00:42:45.349470 systemd-logind[1310]: Watching system buttons on /dev/input/event2 (Sleep Button) May 17 00:42:45.349718 systemd-logind[1310]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) May 17 00:42:45.352182 systemd-logind[1310]: New seat seat0. May 17 00:42:45.362831 env[1327]: time="2025-05-17T00:42:45.362613358Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 May 17 00:42:45.362997 env[1327]: time="2025-05-17T00:42:45.362927990Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 May 17 00:42:45.364816 systemd[1]: Started systemd-logind.service. May 17 00:42:45.366113 env[1327]: time="2025-05-17T00:42:45.366035104Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.182-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 May 17 00:42:45.366244 env[1327]: time="2025-05-17T00:42:45.366116235Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 May 17 00:42:45.366847 env[1327]: time="2025-05-17T00:42:45.366782201Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:42:45.366847 env[1327]: time="2025-05-17T00:42:45.366834827Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 May 17 00:42:45.367073 env[1327]: time="2025-05-17T00:42:45.366869220Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" May 17 00:42:45.367073 env[1327]: time="2025-05-17T00:42:45.366891444Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 May 17 00:42:45.367204 env[1327]: time="2025-05-17T00:42:45.367072428Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 May 17 00:42:45.388969 env[1327]: time="2025-05-17T00:42:45.388913434Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 May 17 00:42:45.389467 env[1327]: time="2025-05-17T00:42:45.389405330Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 May 17 00:42:45.389467 env[1327]: time="2025-05-17T00:42:45.389461504Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 May 17 00:42:45.389654 env[1327]: time="2025-05-17T00:42:45.389595641Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" May 17 00:42:45.389654 env[1327]: time="2025-05-17T00:42:45.389622042Z" level=info msg="metadata content store policy set" policy=shared May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.400995272Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401066240Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401098581Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401182857Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401269732Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401302538Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401339410Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401396962Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401429453Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401459730Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401489912Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401519635Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401742548Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 May 17 00:42:45.403387 env[1327]: time="2025-05-17T00:42:45.401888019Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402501516Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402551411Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402583083Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402663933Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402689108Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402713714Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402740620Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402769262Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402798358Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402830813Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402854542Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.402886005Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.403106556Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.403131805Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 May 17 00:42:45.404220 env[1327]: time="2025-05-17T00:42:45.403154546Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 May 17 00:42:45.407741 env[1327]: time="2025-05-17T00:42:45.403175803Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 May 17 00:42:45.407741 env[1327]: time="2025-05-17T00:42:45.403204762Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 May 17 00:42:45.407741 env[1327]: time="2025-05-17T00:42:45.403230696Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 May 17 00:42:45.407741 env[1327]: time="2025-05-17T00:42:45.403265974Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" May 17 00:42:45.407741 env[1327]: time="2025-05-17T00:42:45.403326662Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 May 17 00:42:45.408047 env[1327]: time="2025-05-17T00:42:45.405542536Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" May 17 00:42:45.408047 env[1327]: time="2025-05-17T00:42:45.406068138Z" level=info msg="Connect containerd service" May 17 00:42:45.408047 env[1327]: time="2025-05-17T00:42:45.406148153Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.408169237Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.408923715Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.409009138Z" level=info msg=serving... address=/run/containerd/containerd.sock May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.409706913Z" level=info msg="containerd successfully booted in 0.294727s" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.409888640Z" level=info msg="Start subscribing containerd event" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.409960163Z" level=info msg="Start recovering state" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.410097616Z" level=info msg="Start event monitor" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.410137585Z" level=info msg="Start snapshots syncer" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.410157429Z" level=info msg="Start cni network conf syncer for default" May 17 00:42:45.411967 env[1327]: time="2025-05-17T00:42:45.410174893Z" level=info msg="Start streaming server" May 17 00:42:45.409392 systemd[1]: Started containerd.service. May 17 00:42:45.762169 dbus-daemon[1291]: [system] Successfully activated service 'org.freedesktop.hostname1' May 17 00:42:45.762450 systemd[1]: Started systemd-hostnamed.service. May 17 00:42:45.763820 dbus-daemon[1291]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.6' (uid=0 pid=1372 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") May 17 00:42:45.777973 systemd[1]: Starting polkit.service... May 17 00:42:45.865730 polkitd[1406]: Started polkitd version 121 May 17 00:42:45.902109 polkitd[1406]: Loading rules from directory /etc/polkit-1/rules.d May 17 00:42:45.902224 polkitd[1406]: Loading rules from directory /usr/share/polkit-1/rules.d May 17 00:42:45.913103 polkitd[1406]: Finished loading, compiling and executing 2 rules May 17 00:42:45.914119 dbus-daemon[1291]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' May 17 00:42:45.914515 systemd[1]: Started polkit.service. May 17 00:42:45.915930 polkitd[1406]: Acquired the name org.freedesktop.PolicyKit1 on the system bus May 17 00:42:45.971031 systemd-hostnamed[1372]: Hostname set to (transient) May 17 00:42:45.976375 systemd-resolved[1225]: System hostname changed to 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9'. May 17 00:42:46.892515 tar[1323]: linux-amd64/LICENSE May 17 00:42:46.903525 tar[1323]: linux-amd64/README.md May 17 00:42:46.922843 systemd[1]: Finished prepare-helm.service. May 17 00:42:47.456057 systemd[1]: Started kubelet.service. May 17 00:42:48.192213 locksmithd[1371]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" May 17 00:42:48.780616 kubelet[1425]: E0517 00:42:48.780557 1425 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:42:48.783912 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:42:48.784279 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:42:52.042186 sshd_keygen[1334]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 May 17 00:42:52.092795 systemd[1]: Finished sshd-keygen.service. May 17 00:42:52.104922 systemd[1]: Starting issuegen.service... May 17 00:42:52.118743 systemd[1]: issuegen.service: Deactivated successfully. May 17 00:42:52.119226 systemd[1]: Finished issuegen.service. May 17 00:42:52.130596 systemd[1]: Starting systemd-user-sessions.service... May 17 00:42:52.144049 systemd[1]: Finished systemd-user-sessions.service. May 17 00:42:52.154961 systemd[1]: Started getty@tty1.service. May 17 00:42:52.165281 systemd[1]: Started serial-getty@ttyS0.service. May 17 00:42:52.173976 systemd[1]: Reached target getty.target. May 17 00:42:52.432110 systemd[1]: var-lib-flatcar\x2doem\x2dgce.mount: Deactivated successfully. May 17 00:42:53.360705 systemd[1]: Created slice system-sshd.slice. May 17 00:42:53.371980 systemd[1]: Started sshd@0-10.128.0.10:22-139.178.89.65:34962.service. May 17 00:42:53.712225 sshd[1450]: Accepted publickey for core from 139.178.89.65 port 34962 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:42:53.724814 sshd[1450]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:42:53.762698 systemd[1]: Created slice user-500.slice. May 17 00:42:53.773707 systemd[1]: Starting user-runtime-dir@500.service... May 17 00:42:53.796857 systemd-logind[1310]: New session 1 of user core. May 17 00:42:53.822237 systemd[1]: Finished user-runtime-dir@500.service. May 17 00:42:53.835652 systemd[1]: Starting user@500.service... May 17 00:42:53.871137 (systemd)[1455]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) May 17 00:42:54.043865 systemd[1455]: Queued start job for default target default.target. May 17 00:42:54.045053 systemd[1455]: Reached target paths.target. May 17 00:42:54.045109 systemd[1455]: Reached target sockets.target. May 17 00:42:54.045140 systemd[1455]: Reached target timers.target. May 17 00:42:54.045174 systemd[1455]: Reached target basic.target. May 17 00:42:54.045460 systemd[1]: Started user@500.service. May 17 00:42:54.045642 systemd[1455]: Reached target default.target. May 17 00:42:54.045737 systemd[1455]: Startup finished in 155ms. May 17 00:42:54.054720 systemd[1]: Started session-1.scope. May 17 00:42:54.284340 systemd[1]: Started sshd@1-10.128.0.10:22-139.178.89.65:34966.service. May 17 00:42:54.579578 sshd[1464]: Accepted publickey for core from 139.178.89.65 port 34966 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:42:54.581803 sshd[1464]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:42:54.588208 systemd-logind[1310]: New session 2 of user core. May 17 00:42:54.590268 systemd[1]: Started session-2.scope. May 17 00:42:54.769399 kernel: loop2: detected capacity change from 0 to 2097152 May 17 00:42:54.798734 systemd-nspawn[1469]: Spawning container oem-gce on /var/lib/flatcar-oem-gce.img. May 17 00:42:54.798734 systemd-nspawn[1469]: Press ^] three times within 1s to kill container. May 17 00:42:54.800659 sshd[1464]: pam_unix(sshd:session): session closed for user core May 17 00:42:54.806454 systemd[1]: sshd@1-10.128.0.10:22-139.178.89.65:34966.service: Deactivated successfully. May 17 00:42:54.808179 systemd[1]: session-2.scope: Deactivated successfully. May 17 00:42:54.813974 systemd-logind[1310]: Session 2 logged out. Waiting for processes to exit. May 17 00:42:54.815624 kernel: EXT4-fs (loop2): mounted filesystem without journal. Opts: norecovery. Quota mode: none. May 17 00:42:54.816701 systemd-logind[1310]: Removed session 2. May 17 00:42:54.844712 systemd[1]: Started sshd@2-10.128.0.10:22-139.178.89.65:34972.service. May 17 00:42:54.906110 systemd[1]: Started oem-gce.service. May 17 00:42:54.914087 systemd[1]: Reached target multi-user.target. May 17 00:42:54.925435 systemd[1]: Starting systemd-update-utmp-runlevel.service... May 17 00:42:54.938966 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. May 17 00:42:54.939277 systemd[1]: Finished systemd-update-utmp-runlevel.service. May 17 00:42:54.953911 systemd[1]: Startup finished in 10.502s (kernel) + 19.358s (userspace) = 29.861s. May 17 00:42:54.981332 systemd-nspawn[1469]: + '[' -e /etc/default/instance_configs.cfg.template ']' May 17 00:42:54.981332 systemd-nspawn[1469]: + echo -e '[InstanceSetup]\nset_host_keys = false' May 17 00:42:54.981633 systemd-nspawn[1469]: + /usr/bin/google_instance_setup May 17 00:42:55.158886 sshd[1477]: Accepted publickey for core from 139.178.89.65 port 34972 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:42:55.161871 sshd[1477]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:42:55.171147 systemd[1]: Started session-3.scope. May 17 00:42:55.173206 systemd-logind[1310]: New session 3 of user core. May 17 00:42:55.380450 sshd[1477]: pam_unix(sshd:session): session closed for user core May 17 00:42:55.385604 systemd[1]: sshd@2-10.128.0.10:22-139.178.89.65:34972.service: Deactivated successfully. May 17 00:42:55.387442 systemd-logind[1310]: Session 3 logged out. Waiting for processes to exit. May 17 00:42:55.387600 systemd[1]: session-3.scope: Deactivated successfully. May 17 00:42:55.393442 systemd-logind[1310]: Removed session 3. May 17 00:42:55.765943 instance-setup[1481]: INFO Running google_set_multiqueue. May 17 00:42:55.782917 instance-setup[1481]: INFO Set channels for eth0 to 2. May 17 00:42:55.786840 instance-setup[1481]: INFO Setting /proc/irq/27/smp_affinity_list to 0 for device virtio1. May 17 00:42:55.788615 instance-setup[1481]: INFO /proc/irq/27/smp_affinity_list: real affinity 0 May 17 00:42:55.788851 instance-setup[1481]: INFO Setting /proc/irq/28/smp_affinity_list to 0 for device virtio1. May 17 00:42:55.790827 instance-setup[1481]: INFO /proc/irq/28/smp_affinity_list: real affinity 0 May 17 00:42:55.791300 instance-setup[1481]: INFO Setting /proc/irq/29/smp_affinity_list to 1 for device virtio1. May 17 00:42:55.793035 instance-setup[1481]: INFO /proc/irq/29/smp_affinity_list: real affinity 1 May 17 00:42:55.793833 instance-setup[1481]: INFO Setting /proc/irq/30/smp_affinity_list to 1 for device virtio1. May 17 00:42:55.795571 instance-setup[1481]: INFO /proc/irq/30/smp_affinity_list: real affinity 1 May 17 00:42:55.810107 instance-setup[1481]: INFO Queue 0 XPS=1 for /sys/class/net/eth0/queues/tx-0/xps_cpus May 17 00:42:55.810575 instance-setup[1481]: INFO Queue 1 XPS=2 for /sys/class/net/eth0/queues/tx-1/xps_cpus May 17 00:42:55.858701 systemd-nspawn[1469]: + /usr/bin/google_metadata_script_runner --script-type startup May 17 00:42:56.241859 startup-script[1517]: INFO Starting startup scripts. May 17 00:42:56.257361 startup-script[1517]: INFO No startup scripts found in metadata. May 17 00:42:56.257563 startup-script[1517]: INFO Finished running startup scripts. May 17 00:42:56.296974 systemd-nspawn[1469]: + trap 'stopping=1 ; kill "${daemon_pids[@]}" || :' SIGTERM May 17 00:42:56.296974 systemd-nspawn[1469]: + daemon_pids=() May 17 00:42:56.296974 systemd-nspawn[1469]: + for d in accounts clock_skew network May 17 00:42:56.297376 systemd-nspawn[1469]: + daemon_pids+=($!) May 17 00:42:56.297534 systemd-nspawn[1469]: + for d in accounts clock_skew network May 17 00:42:56.297933 systemd-nspawn[1469]: + daemon_pids+=($!) May 17 00:42:56.298118 systemd-nspawn[1469]: + for d in accounts clock_skew network May 17 00:42:56.298308 systemd-nspawn[1469]: + /usr/bin/google_clock_skew_daemon May 17 00:42:56.298736 systemd-nspawn[1469]: + /usr/bin/google_accounts_daemon May 17 00:42:56.298912 systemd-nspawn[1469]: + daemon_pids+=($!) May 17 00:42:56.299127 systemd-nspawn[1469]: + NOTIFY_SOCKET=/run/systemd/notify May 17 00:42:56.299127 systemd-nspawn[1469]: + /usr/bin/systemd-notify --ready May 17 00:42:56.304806 systemd-nspawn[1469]: + /usr/bin/google_network_daemon May 17 00:42:56.370077 systemd-nspawn[1469]: + wait -n 36 37 38 May 17 00:42:57.071170 google-clock-skew[1521]: INFO Starting Google Clock Skew daemon. May 17 00:42:57.090261 google-clock-skew[1521]: INFO Clock drift token has changed: 0. May 17 00:42:57.099739 google-clock-skew[1521]: WARNING Failed to sync system time with hardware clock. May 17 00:42:57.100131 systemd-nspawn[1469]: hwclock: Cannot access the Hardware Clock via any known method. May 17 00:42:57.100131 systemd-nspawn[1469]: hwclock: Use the --verbose option to see the details of our search for an access method. May 17 00:42:57.121617 google-networking[1522]: INFO Starting Google Networking daemon. May 17 00:42:57.127725 groupadd[1531]: group added to /etc/group: name=google-sudoers, GID=1000 May 17 00:42:57.132853 groupadd[1531]: group added to /etc/gshadow: name=google-sudoers May 17 00:42:57.140455 groupadd[1531]: new group: name=google-sudoers, GID=1000 May 17 00:42:57.158496 google-accounts[1520]: INFO Starting Google Accounts daemon. May 17 00:42:57.188604 google-accounts[1520]: WARNING OS Login not installed. May 17 00:42:57.189863 google-accounts[1520]: INFO Creating a new user account for 0. May 17 00:42:57.197840 systemd-nspawn[1469]: useradd: invalid user name '0': use --badname to ignore May 17 00:42:57.198707 google-accounts[1520]: WARNING Could not create user 0. Command '['useradd', '-m', '-s', '/bin/bash', '-p', '*', '0']' returned non-zero exit status 3.. May 17 00:42:59.008844 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. May 17 00:42:59.009245 systemd[1]: Stopped kubelet.service. May 17 00:42:59.011879 systemd[1]: Starting kubelet.service... May 17 00:42:59.346442 systemd[1]: Started kubelet.service. May 17 00:42:59.411548 kubelet[1550]: E0517 00:42:59.411500 1550 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:42:59.415817 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:42:59.416187 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:43:05.422483 systemd[1]: Started sshd@3-10.128.0.10:22-139.178.89.65:46092.service. May 17 00:43:05.715321 sshd[1556]: Accepted publickey for core from 139.178.89.65 port 46092 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:43:05.717435 sshd[1556]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:43:05.724386 systemd-logind[1310]: New session 4 of user core. May 17 00:43:05.725149 systemd[1]: Started session-4.scope. May 17 00:43:05.930524 sshd[1556]: pam_unix(sshd:session): session closed for user core May 17 00:43:05.935174 systemd[1]: sshd@3-10.128.0.10:22-139.178.89.65:46092.service: Deactivated successfully. May 17 00:43:05.937252 systemd[1]: session-4.scope: Deactivated successfully. May 17 00:43:05.937367 systemd-logind[1310]: Session 4 logged out. Waiting for processes to exit. May 17 00:43:05.939336 systemd-logind[1310]: Removed session 4. May 17 00:43:05.974293 systemd[1]: Started sshd@4-10.128.0.10:22-139.178.89.65:46108.service. May 17 00:43:06.266300 sshd[1563]: Accepted publickey for core from 139.178.89.65 port 46108 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:43:06.268511 sshd[1563]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:43:06.276563 systemd[1]: Started session-5.scope. May 17 00:43:06.277190 systemd-logind[1310]: New session 5 of user core. May 17 00:43:06.473732 sshd[1563]: pam_unix(sshd:session): session closed for user core May 17 00:43:06.478616 systemd[1]: sshd@4-10.128.0.10:22-139.178.89.65:46108.service: Deactivated successfully. May 17 00:43:06.480938 systemd-logind[1310]: Session 5 logged out. Waiting for processes to exit. May 17 00:43:06.481098 systemd[1]: session-5.scope: Deactivated successfully. May 17 00:43:06.483559 systemd-logind[1310]: Removed session 5. May 17 00:43:06.518858 systemd[1]: Started sshd@5-10.128.0.10:22-139.178.89.65:36042.service. May 17 00:43:06.811869 sshd[1570]: Accepted publickey for core from 139.178.89.65 port 36042 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:43:06.813932 sshd[1570]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:43:06.821484 systemd[1]: Started session-6.scope. May 17 00:43:06.821854 systemd-logind[1310]: New session 6 of user core. May 17 00:43:07.027401 sshd[1570]: pam_unix(sshd:session): session closed for user core May 17 00:43:07.032452 systemd[1]: sshd@5-10.128.0.10:22-139.178.89.65:36042.service: Deactivated successfully. May 17 00:43:07.034592 systemd[1]: session-6.scope: Deactivated successfully. May 17 00:43:07.034621 systemd-logind[1310]: Session 6 logged out. Waiting for processes to exit. May 17 00:43:07.037180 systemd-logind[1310]: Removed session 6. May 17 00:43:07.072620 systemd[1]: Started sshd@6-10.128.0.10:22-139.178.89.65:36046.service. May 17 00:43:07.372908 sshd[1577]: Accepted publickey for core from 139.178.89.65 port 36046 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:43:07.375207 sshd[1577]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:43:07.383485 systemd-logind[1310]: New session 7 of user core. May 17 00:43:07.384047 systemd[1]: Started session-7.scope. May 17 00:43:07.577052 sudo[1581]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh May 17 00:43:07.577577 sudo[1581]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) May 17 00:43:07.615529 systemd[1]: Starting docker.service... May 17 00:43:07.674792 env[1591]: time="2025-05-17T00:43:07.673826771Z" level=info msg="Starting up" May 17 00:43:07.676506 env[1591]: time="2025-05-17T00:43:07.676465081Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 17 00:43:07.677018 env[1591]: time="2025-05-17T00:43:07.676982664Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 17 00:43:07.677188 env[1591]: time="2025-05-17T00:43:07.677164794Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 17 00:43:07.677270 env[1591]: time="2025-05-17T00:43:07.677253159Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 17 00:43:07.680167 env[1591]: time="2025-05-17T00:43:07.680135722Z" level=info msg="parsed scheme: \"unix\"" module=grpc May 17 00:43:07.680321 env[1591]: time="2025-05-17T00:43:07.680300240Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc May 17 00:43:07.680462 env[1591]: time="2025-05-17T00:43:07.680438981Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc May 17 00:43:07.680553 env[1591]: time="2025-05-17T00:43:07.680535493Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc May 17 00:43:07.693920 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport1607680913-merged.mount: Deactivated successfully. May 17 00:43:07.890769 env[1591]: time="2025-05-17T00:43:07.890709168Z" level=warning msg="Your kernel does not support cgroup blkio weight" May 17 00:43:07.890769 env[1591]: time="2025-05-17T00:43:07.890753320Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" May 17 00:43:07.891161 env[1591]: time="2025-05-17T00:43:07.891107902Z" level=info msg="Loading containers: start." May 17 00:43:08.086393 kernel: Initializing XFRM netlink socket May 17 00:43:08.137268 env[1591]: time="2025-05-17T00:43:08.137205352Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" May 17 00:43:08.233685 systemd-networkd[1079]: docker0: Link UP May 17 00:43:08.256600 env[1591]: time="2025-05-17T00:43:08.256541503Z" level=info msg="Loading containers: done." May 17 00:43:08.276892 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck2465785078-merged.mount: Deactivated successfully. May 17 00:43:08.284445 env[1591]: time="2025-05-17T00:43:08.284333833Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 May 17 00:43:08.284815 env[1591]: time="2025-05-17T00:43:08.284715594Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 May 17 00:43:08.284915 env[1591]: time="2025-05-17T00:43:08.284885789Z" level=info msg="Daemon has completed initialization" May 17 00:43:08.309576 systemd[1]: Started docker.service. May 17 00:43:08.322924 env[1591]: time="2025-05-17T00:43:08.322659759Z" level=info msg="API listen on /run/docker.sock" May 17 00:43:09.386178 systemd[1]: Started sshd@7-10.128.0.10:22-106.58.166.77:33302.service. May 17 00:43:09.508993 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. May 17 00:43:09.509387 systemd[1]: Stopped kubelet.service. May 17 00:43:09.512027 systemd[1]: Starting kubelet.service... May 17 00:43:09.545085 env[1327]: time="2025-05-17T00:43:09.545013976Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.9\"" May 17 00:43:09.865720 systemd[1]: Started kubelet.service. May 17 00:43:09.928142 kubelet[1722]: E0517 00:43:09.928068 1722 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:43:09.930593 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:43:09.930929 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:43:10.225683 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3050236452.mount: Deactivated successfully. May 17 00:43:11.855000 env[1327]: time="2025-05-17T00:43:11.854917235Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:11.858208 env[1327]: time="2025-05-17T00:43:11.858148987Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:11.860989 env[1327]: time="2025-05-17T00:43:11.860931384Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:11.863698 env[1327]: time="2025-05-17T00:43:11.863610907Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:5b68f0df22013422dc8fb9ddfcff513eb6fc92f9dbf8aae41555c895efef5a20,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:11.864922 env[1327]: time="2025-05-17T00:43:11.864829745Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.31.9\" returns image reference \"sha256:0c19e0eafbdfffa1317cf99a16478265a4cd746ef677de27b0be6a8b515f36b1\"" May 17 00:43:11.866017 env[1327]: time="2025-05-17T00:43:11.865977425Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.9\"" May 17 00:43:13.499529 env[1327]: time="2025-05-17T00:43:13.499453141Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:13.502712 env[1327]: time="2025-05-17T00:43:13.502652099Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:13.508656 env[1327]: time="2025-05-17T00:43:13.508584183Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:13.511260 env[1327]: time="2025-05-17T00:43:13.511197503Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:be9e7987d323b38a12e28436cff6d6ec6fc31ffdd3ea11eaa9d74852e9d31248,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:13.512719 env[1327]: time="2025-05-17T00:43:13.512651392Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.31.9\" returns image reference \"sha256:6aa3d581404ae6ae5dc355cb750aaedec843d2c99263d28fce50277e8e2a6ec2\"" May 17 00:43:13.513544 env[1327]: time="2025-05-17T00:43:13.513493354Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.9\"" May 17 00:43:14.882694 env[1327]: time="2025-05-17T00:43:14.882614654Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:14.885908 env[1327]: time="2025-05-17T00:43:14.885841778Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:14.888756 env[1327]: time="2025-05-17T00:43:14.888688936Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:14.891459 env[1327]: time="2025-05-17T00:43:14.891404678Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:eb358c7346bb17ab2c639c3ff8ab76a147dec7ae609f5c0c2800233e42253ed1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:14.892553 env[1327]: time="2025-05-17T00:43:14.892496843Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.31.9\" returns image reference \"sha256:737ed3eafaf27a28ea9e13b736011bfed5bd349785ac6bc220b34eaf4adc51e3\"" May 17 00:43:14.893550 env[1327]: time="2025-05-17T00:43:14.893511331Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.9\"" May 17 00:43:16.006254 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2966462671.mount: Deactivated successfully. May 17 00:43:16.012902 systemd[1]: systemd-hostnamed.service: Deactivated successfully. May 17 00:43:16.792374 env[1327]: time="2025-05-17T00:43:16.792289391Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:16.799726 env[1327]: time="2025-05-17T00:43:16.799663751Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:16.802049 env[1327]: time="2025-05-17T00:43:16.801950739Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.31.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:16.805765 env[1327]: time="2025-05-17T00:43:16.805712953Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:fdf026cf2434537e499e9c739d189ca8fc57101d929ac5ccd8e24f979a9738c1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:16.806452 env[1327]: time="2025-05-17T00:43:16.806406054Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.31.9\" returns image reference \"sha256:11a47a71ed3ecf643e15a11990daed3b656279449ba9344db0b54652c4723578\"" May 17 00:43:16.807221 env[1327]: time="2025-05-17T00:43:16.807182956Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" May 17 00:43:17.224012 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount282675161.mount: Deactivated successfully. May 17 00:43:18.511686 env[1327]: time="2025-05-17T00:43:18.511596067Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.517274 env[1327]: time="2025-05-17T00:43:18.517193969Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.523694 env[1327]: time="2025-05-17T00:43:18.523637706Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.527474 env[1327]: time="2025-05-17T00:43:18.527415324Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.528579 env[1327]: time="2025-05-17T00:43:18.528501687Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:c69fa2e9cbf5f42dc48af631e956d3f95724c13f91596bc567591790e5e36db6\"" May 17 00:43:18.530152 env[1327]: time="2025-05-17T00:43:18.530099951Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" May 17 00:43:18.940608 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2499565719.mount: Deactivated successfully. May 17 00:43:18.948851 env[1327]: time="2025-05-17T00:43:18.948765620Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.951685 env[1327]: time="2025-05-17T00:43:18.951627091Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.954868 env[1327]: time="2025-05-17T00:43:18.954777013Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.10,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.957605 env[1327]: time="2025-05-17T00:43:18.957520019Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:18.959236 env[1327]: time="2025-05-17T00:43:18.959160741Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:873ed75102791e5b0b8a7fcd41606c92fcec98d56d05ead4ac5131650004c136\"" May 17 00:43:18.960078 env[1327]: time="2025-05-17T00:43:18.960021497Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\"" May 17 00:43:19.327231 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount61020302.mount: Deactivated successfully. May 17 00:43:20.008867 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. May 17 00:43:20.009271 systemd[1]: Stopped kubelet.service. May 17 00:43:20.012089 systemd[1]: Starting kubelet.service... May 17 00:43:20.403860 systemd[1]: Started kubelet.service. May 17 00:43:20.504788 kubelet[1739]: E0517 00:43:20.504563 1739 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" May 17 00:43:20.508106 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE May 17 00:43:20.508459 systemd[1]: kubelet.service: Failed with result 'exit-code'. May 17 00:43:22.253169 env[1327]: time="2025-05-17T00:43:22.253087382Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:22.259542 env[1327]: time="2025-05-17T00:43:22.259476602Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:22.263990 env[1327]: time="2025-05-17T00:43:22.263925545Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.15-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:22.268769 env[1327]: time="2025-05-17T00:43:22.268702376Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:a6dc63e6e8cfa0307d7851762fa6b629afb18f28d8aa3fab5a6e91b4af60026a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:22.270203 env[1327]: time="2025-05-17T00:43:22.270111206Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.15-0\" returns image reference \"sha256:2e96e5913fc06e3d26915af3d0f2ca5048cc4b6327e661e80da792cbf8d8d9d4\"" May 17 00:43:26.430713 systemd[1]: Stopped kubelet.service. May 17 00:43:26.434872 systemd[1]: Starting kubelet.service... May 17 00:43:26.491786 systemd[1]: Reloading. May 17 00:43:26.665765 /usr/lib/systemd/system-generators/torcx-generator[1799]: time="2025-05-17T00:43:26Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 17 00:43:26.669503 /usr/lib/systemd/system-generators/torcx-generator[1799]: time="2025-05-17T00:43:26Z" level=info msg="torcx already run" May 17 00:43:26.826969 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:43:26.827001 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:43:26.852528 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:43:26.995164 systemd[1]: Started kubelet.service. May 17 00:43:26.999536 systemd[1]: Stopping kubelet.service... May 17 00:43:27.000739 systemd[1]: kubelet.service: Deactivated successfully. May 17 00:43:27.001425 systemd[1]: Stopped kubelet.service. May 17 00:43:27.005323 systemd[1]: Starting kubelet.service... May 17 00:43:27.354165 systemd[1]: Started kubelet.service. May 17 00:43:27.429732 kubelet[1859]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:43:27.430337 kubelet[1859]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 17 00:43:27.430490 kubelet[1859]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:43:27.431101 kubelet[1859]: I0517 00:43:27.430786 1859 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 17 00:43:28.044528 kubelet[1859]: I0517 00:43:28.044449 1859 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" May 17 00:43:28.044528 kubelet[1859]: I0517 00:43:28.044500 1859 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 17 00:43:28.044938 kubelet[1859]: I0517 00:43:28.044895 1859 server.go:934] "Client rotation is on, will bootstrap in background" May 17 00:43:28.107576 kubelet[1859]: E0517 00:43:28.107523 1859 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://10.128.0.10:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:28.111202 kubelet[1859]: I0517 00:43:28.111142 1859 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 17 00:43:28.124840 kubelet[1859]: E0517 00:43:28.124777 1859 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 17 00:43:28.125249 kubelet[1859]: I0517 00:43:28.125082 1859 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 17 00:43:28.133905 kubelet[1859]: I0517 00:43:28.133842 1859 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 17 00:43:28.134456 kubelet[1859]: I0517 00:43:28.134427 1859 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" May 17 00:43:28.134742 kubelet[1859]: I0517 00:43:28.134673 1859 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 17 00:43:28.135022 kubelet[1859]: I0517 00:43:28.134728 1859 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} May 17 00:43:28.135212 kubelet[1859]: I0517 00:43:28.135031 1859 topology_manager.go:138] "Creating topology manager with none policy" May 17 00:43:28.135212 kubelet[1859]: I0517 00:43:28.135052 1859 container_manager_linux.go:300] "Creating device plugin manager" May 17 00:43:28.135389 kubelet[1859]: I0517 00:43:28.135222 1859 state_mem.go:36] "Initialized new in-memory state store" May 17 00:43:28.143361 kubelet[1859]: I0517 00:43:28.143287 1859 kubelet.go:408] "Attempting to sync node with API server" May 17 00:43:28.143723 kubelet[1859]: I0517 00:43:28.143637 1859 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" May 17 00:43:28.143723 kubelet[1859]: I0517 00:43:28.143710 1859 kubelet.go:314] "Adding apiserver pod source" May 17 00:43:28.143902 kubelet[1859]: I0517 00:43:28.143748 1859 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 17 00:43:28.162608 kubelet[1859]: W0517 00:43:28.162506 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.128.0.10:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9&limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:28.163078 kubelet[1859]: E0517 00:43:28.163028 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.128.0.10:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9&limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:28.163439 kubelet[1859]: I0517 00:43:28.163413 1859 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 17 00:43:28.164578 kubelet[1859]: I0517 00:43:28.164547 1859 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 17 00:43:28.164750 kubelet[1859]: W0517 00:43:28.164694 1859 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. May 17 00:43:28.168432 kubelet[1859]: I0517 00:43:28.168394 1859 server.go:1274] "Started kubelet" May 17 00:43:28.187557 kubelet[1859]: I0517 00:43:28.187496 1859 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 17 00:43:28.202637 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). May 17 00:43:28.206088 kubelet[1859]: I0517 00:43:28.205999 1859 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 17 00:43:28.206331 kubelet[1859]: E0517 00:43:28.194506 1859 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.128.0.10:6443/api/v1/namespaces/default/events\": dial tcp 10.128.0.10:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9.184029cb354162f1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,UID:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,},FirstTimestamp:2025-05-17 00:43:28.168329969 +0000 UTC m=+0.796267012,LastTimestamp:2025-05-17 00:43:28.168329969 +0000 UTC m=+0.796267012,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,}" May 17 00:43:28.206595 kubelet[1859]: I0517 00:43:28.206566 1859 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 17 00:43:28.207482 kubelet[1859]: W0517 00:43:28.187638 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.128.0.10:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:28.207482 kubelet[1859]: E0517 00:43:28.206864 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.128.0.10:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:28.207839 kubelet[1859]: I0517 00:43:28.207808 1859 server.go:449] "Adding debug handlers to kubelet server" May 17 00:43:28.208015 kubelet[1859]: I0517 00:43:28.207995 1859 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 17 00:43:28.214577 kubelet[1859]: I0517 00:43:28.214315 1859 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 17 00:43:28.221073 kubelet[1859]: E0517 00:43:28.219450 1859 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 17 00:43:28.221073 kubelet[1859]: E0517 00:43:28.219545 1859 kubelet_node_status.go:453] "Error getting the current node from lister" err="node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" not found" May 17 00:43:28.221073 kubelet[1859]: I0517 00:43:28.219597 1859 volume_manager.go:289] "Starting Kubelet Volume Manager" May 17 00:43:28.221073 kubelet[1859]: I0517 00:43:28.219958 1859 desired_state_of_world_populator.go:147] "Desired state populator starts to run" May 17 00:43:28.221073 kubelet[1859]: I0517 00:43:28.220043 1859 reconciler.go:26] "Reconciler: start to sync state" May 17 00:43:28.221073 kubelet[1859]: W0517 00:43:28.220655 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.128.0.10:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:28.221073 kubelet[1859]: E0517 00:43:28.220741 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.128.0.10:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:28.221073 kubelet[1859]: I0517 00:43:28.221029 1859 factory.go:221] Registration of the systemd container factory successfully May 17 00:43:28.221975 kubelet[1859]: I0517 00:43:28.221143 1859 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 17 00:43:28.223213 kubelet[1859]: E0517 00:43:28.222717 1859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.10:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9?timeout=10s\": dial tcp 10.128.0.10:6443: connect: connection refused" interval="200ms" May 17 00:43:28.224800 kubelet[1859]: I0517 00:43:28.224606 1859 factory.go:221] Registration of the containerd container factory successfully May 17 00:43:28.266205 kubelet[1859]: I0517 00:43:28.266126 1859 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 17 00:43:28.269942 kubelet[1859]: I0517 00:43:28.269898 1859 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 17 00:43:28.269942 kubelet[1859]: I0517 00:43:28.269945 1859 status_manager.go:217] "Starting to sync pod status with apiserver" May 17 00:43:28.270173 kubelet[1859]: I0517 00:43:28.269975 1859 kubelet.go:2321] "Starting kubelet main sync loop" May 17 00:43:28.270173 kubelet[1859]: E0517 00:43:28.270039 1859 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 17 00:43:28.273098 kubelet[1859]: W0517 00:43:28.273050 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.128.0.10:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:28.273276 kubelet[1859]: E0517 00:43:28.273119 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.128.0.10:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:28.282030 kubelet[1859]: I0517 00:43:28.281915 1859 cpu_manager.go:214] "Starting CPU manager" policy="none" May 17 00:43:28.282030 kubelet[1859]: I0517 00:43:28.281944 1859 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 17 00:43:28.282030 kubelet[1859]: I0517 00:43:28.281974 1859 state_mem.go:36] "Initialized new in-memory state store" May 17 00:43:28.285166 kubelet[1859]: I0517 00:43:28.285112 1859 policy_none.go:49] "None policy: Start" May 17 00:43:28.286583 kubelet[1859]: I0517 00:43:28.286556 1859 memory_manager.go:170] "Starting memorymanager" policy="None" May 17 00:43:28.286918 kubelet[1859]: I0517 00:43:28.286867 1859 state_mem.go:35] "Initializing new in-memory state store" May 17 00:43:28.296442 kubelet[1859]: I0517 00:43:28.294918 1859 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 17 00:43:28.296961 kubelet[1859]: I0517 00:43:28.296932 1859 eviction_manager.go:189] "Eviction manager: starting control loop" May 17 00:43:28.298572 kubelet[1859]: I0517 00:43:28.298477 1859 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 17 00:43:28.299251 kubelet[1859]: I0517 00:43:28.299189 1859 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 17 00:43:28.305609 kubelet[1859]: E0517 00:43:28.305563 1859 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" not found" May 17 00:43:28.405217 kubelet[1859]: I0517 00:43:28.405167 1859 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.406199 kubelet[1859]: E0517 00:43:28.406159 1859 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.128.0.10:6443/api/v1/nodes\": dial tcp 10.128.0.10:6443: connect: connection refused" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.420938 kubelet[1859]: I0517 00:43:28.420871 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-kubeconfig\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421234 kubelet[1859]: I0517 00:43:28.420949 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421234 kubelet[1859]: I0517 00:43:28.420993 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2f54a6b75ee979f46c55e66226f79feb-kubeconfig\") pod \"kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"2f54a6b75ee979f46c55e66226f79feb\") " pod="kube-system/kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421234 kubelet[1859]: I0517 00:43:28.421026 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-flexvolume-dir\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421234 kubelet[1859]: I0517 00:43:28.421058 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-k8s-certs\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421757 kubelet[1859]: I0517 00:43:28.421134 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/da1027880b2f3ae58454bfe774080950-ca-certs\") pod \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"da1027880b2f3ae58454bfe774080950\") " pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421757 kubelet[1859]: I0517 00:43:28.421169 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/da1027880b2f3ae58454bfe774080950-k8s-certs\") pod \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"da1027880b2f3ae58454bfe774080950\") " pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421757 kubelet[1859]: I0517 00:43:28.421198 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/da1027880b2f3ae58454bfe774080950-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"da1027880b2f3ae58454bfe774080950\") " pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.421757 kubelet[1859]: I0517 00:43:28.421227 1859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-ca-certs\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.424408 kubelet[1859]: E0517 00:43:28.424296 1859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.10:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9?timeout=10s\": dial tcp 10.128.0.10:6443: connect: connection refused" interval="400ms" May 17 00:43:28.613439 kubelet[1859]: I0517 00:43:28.612327 1859 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.613439 kubelet[1859]: E0517 00:43:28.612794 1859 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.128.0.10:6443/api/v1/nodes\": dial tcp 10.128.0.10:6443: connect: connection refused" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:28.687759 env[1327]: time="2025-05-17T00:43:28.687687057Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,Uid:da1027880b2f3ae58454bfe774080950,Namespace:kube-system,Attempt:0,}" May 17 00:43:28.707789 env[1327]: time="2025-05-17T00:43:28.707732684Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,Uid:0ed4d1288c326e677089947e941e90b9,Namespace:kube-system,Attempt:0,}" May 17 00:43:28.715776 env[1327]: time="2025-05-17T00:43:28.715672466Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,Uid:2f54a6b75ee979f46c55e66226f79feb,Namespace:kube-system,Attempt:0,}" May 17 00:43:28.825882 kubelet[1859]: E0517 00:43:28.825782 1859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.10:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9?timeout=10s\": dial tcp 10.128.0.10:6443: connect: connection refused" interval="800ms" May 17 00:43:29.018744 kubelet[1859]: I0517 00:43:29.018704 1859 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:29.019211 kubelet[1859]: E0517 00:43:29.019148 1859 kubelet_node_status.go:95] "Unable to register node with API server" err="Post \"https://10.128.0.10:6443/api/v1/nodes\": dial tcp 10.128.0.10:6443: connect: connection refused" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:29.098771 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3050778746.mount: Deactivated successfully. May 17 00:43:29.111173 env[1327]: time="2025-05-17T00:43:29.111105117Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.112765 env[1327]: time="2025-05-17T00:43:29.112712252Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.117642 env[1327]: time="2025-05-17T00:43:29.117574150Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.120959 env[1327]: time="2025-05-17T00:43:29.120884441Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.122380 env[1327]: time="2025-05-17T00:43:29.122302068Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.123959 env[1327]: time="2025-05-17T00:43:29.123890732Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.126837 env[1327]: time="2025-05-17T00:43:29.126780483Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.129036 env[1327]: time="2025-05-17T00:43:29.128987023Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.131250 env[1327]: time="2025-05-17T00:43:29.131186662Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.132921 env[1327]: time="2025-05-17T00:43:29.132861533Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.139386 env[1327]: time="2025-05-17T00:43:29.139305992Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.141261 env[1327]: time="2025-05-17T00:43:29.141175626Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:29.174919 kubelet[1859]: W0517 00:43:29.174832 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.128.0.10:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:29.175096 kubelet[1859]: E0517 00:43:29.174942 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://10.128.0.10:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:29.187090 env[1327]: time="2025-05-17T00:43:29.179214356Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:43:29.187090 env[1327]: time="2025-05-17T00:43:29.179306229Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:43:29.187090 env[1327]: time="2025-05-17T00:43:29.179333412Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:43:29.204043 env[1327]: time="2025-05-17T00:43:29.201762777Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/337ba12893eede6ffb81ec78d9cd1048e6c29aff00da4996611524f06971d615 pid=1899 runtime=io.containerd.runc.v2 May 17 00:43:29.218813 env[1327]: time="2025-05-17T00:43:29.218713412Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:43:29.219118 env[1327]: time="2025-05-17T00:43:29.219074861Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:43:29.219308 env[1327]: time="2025-05-17T00:43:29.219265771Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:43:29.224375 env[1327]: time="2025-05-17T00:43:29.224192881Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:43:29.224618 env[1327]: time="2025-05-17T00:43:29.224374443Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:43:29.224618 env[1327]: time="2025-05-17T00:43:29.224427011Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:43:29.224772 env[1327]: time="2025-05-17T00:43:29.224700934Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/28bf1cb81cee1081700e43a46ec58975529061e4d2a52fe9bf057067722d574c pid=1922 runtime=io.containerd.runc.v2 May 17 00:43:29.228142 env[1327]: time="2025-05-17T00:43:29.227832426Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/39ee6e8fd62bfc9dc3fa00d185d66e30dc99380f6e82347c768afea58181f487 pid=1935 runtime=io.containerd.runc.v2 May 17 00:43:29.392145 env[1327]: time="2025-05-17T00:43:29.391993934Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,Uid:0ed4d1288c326e677089947e941e90b9,Namespace:kube-system,Attempt:0,} returns sandbox id \"28bf1cb81cee1081700e43a46ec58975529061e4d2a52fe9bf057067722d574c\"" May 17 00:43:29.395379 kubelet[1859]: E0517 00:43:29.395268 1859 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" hostnameMaxLen=63 truncatedHostname="kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb" May 17 00:43:29.397726 env[1327]: time="2025-05-17T00:43:29.397669858Z" level=info msg="CreateContainer within sandbox \"28bf1cb81cee1081700e43a46ec58975529061e4d2a52fe9bf057067722d574c\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" May 17 00:43:29.404255 env[1327]: time="2025-05-17T00:43:29.403621127Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,Uid:2f54a6b75ee979f46c55e66226f79feb,Namespace:kube-system,Attempt:0,} returns sandbox id \"337ba12893eede6ffb81ec78d9cd1048e6c29aff00da4996611524f06971d615\"" May 17 00:43:29.405294 kubelet[1859]: E0517 00:43:29.405250 1859 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" hostnameMaxLen=63 truncatedHostname="kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d" May 17 00:43:29.406886 env[1327]: time="2025-05-17T00:43:29.406828552Z" level=info msg="CreateContainer within sandbox \"337ba12893eede6ffb81ec78d9cd1048e6c29aff00da4996611524f06971d615\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" May 17 00:43:29.422851 env[1327]: time="2025-05-17T00:43:29.422778289Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9,Uid:da1027880b2f3ae58454bfe774080950,Namespace:kube-system,Attempt:0,} returns sandbox id \"39ee6e8fd62bfc9dc3fa00d185d66e30dc99380f6e82347c768afea58181f487\"" May 17 00:43:29.425604 kubelet[1859]: E0517 00:43:29.425542 1859 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" hostnameMaxLen=63 truncatedHostname="kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d" May 17 00:43:29.427410 env[1327]: time="2025-05-17T00:43:29.427315539Z" level=info msg="CreateContainer within sandbox \"39ee6e8fd62bfc9dc3fa00d185d66e30dc99380f6e82347c768afea58181f487\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" May 17 00:43:29.429897 env[1327]: time="2025-05-17T00:43:29.429846862Z" level=info msg="CreateContainer within sandbox \"28bf1cb81cee1081700e43a46ec58975529061e4d2a52fe9bf057067722d574c\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"058d8b7889fffeffaf96b7d17433f5d3bcddf0864ea245b5c7999e31c807bba1\"" May 17 00:43:29.431017 env[1327]: time="2025-05-17T00:43:29.430980030Z" level=info msg="StartContainer for \"058d8b7889fffeffaf96b7d17433f5d3bcddf0864ea245b5c7999e31c807bba1\"" May 17 00:43:29.441400 env[1327]: time="2025-05-17T00:43:29.441312606Z" level=info msg="CreateContainer within sandbox \"337ba12893eede6ffb81ec78d9cd1048e6c29aff00da4996611524f06971d615\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"8386f9903d0a5e299c109a7b97243aa97b71fbe819875d89669520b39361d23e\"" May 17 00:43:29.443122 env[1327]: time="2025-05-17T00:43:29.443062453Z" level=info msg="StartContainer for \"8386f9903d0a5e299c109a7b97243aa97b71fbe819875d89669520b39361d23e\"" May 17 00:43:29.464326 env[1327]: time="2025-05-17T00:43:29.464252247Z" level=info msg="CreateContainer within sandbox \"39ee6e8fd62bfc9dc3fa00d185d66e30dc99380f6e82347c768afea58181f487\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"2f670dcb3649b282a831f8113fabdd4d3e0fc77141512aa8e2bf8e766bcf8222\"" May 17 00:43:29.468023 env[1327]: time="2025-05-17T00:43:29.467961070Z" level=info msg="StartContainer for \"2f670dcb3649b282a831f8113fabdd4d3e0fc77141512aa8e2bf8e766bcf8222\"" May 17 00:43:29.526173 kubelet[1859]: W0517 00:43:29.526066 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.128.0.10:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:29.526173 kubelet[1859]: E0517 00:43:29.526173 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://10.128.0.10:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:29.529019 kubelet[1859]: W0517 00:43:29.528928 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.128.0.10:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9&limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:29.529233 kubelet[1859]: E0517 00:43:29.529031 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://10.128.0.10:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9&limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:29.612704 env[1327]: time="2025-05-17T00:43:29.612621988Z" level=info msg="StartContainer for \"058d8b7889fffeffaf96b7d17433f5d3bcddf0864ea245b5c7999e31c807bba1\" returns successfully" May 17 00:43:29.627108 kubelet[1859]: E0517 00:43:29.627025 1859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.10:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9?timeout=10s\": dial tcp 10.128.0.10:6443: connect: connection refused" interval="1.6s" May 17 00:43:29.653231 env[1327]: time="2025-05-17T00:43:29.653088297Z" level=info msg="StartContainer for \"8386f9903d0a5e299c109a7b97243aa97b71fbe819875d89669520b39361d23e\" returns successfully" May 17 00:43:29.703081 env[1327]: time="2025-05-17T00:43:29.703014957Z" level=info msg="StartContainer for \"2f670dcb3649b282a831f8113fabdd4d3e0fc77141512aa8e2bf8e766bcf8222\" returns successfully" May 17 00:43:29.742729 kubelet[1859]: W0517 00:43:29.742507 1859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.128.0.10:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.128.0.10:6443: connect: connection refused May 17 00:43:29.742729 kubelet[1859]: E0517 00:43:29.742659 1859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://10.128.0.10:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 10.128.0.10:6443: connect: connection refused" logger="UnhandledError" May 17 00:43:29.827035 kubelet[1859]: I0517 00:43:29.826997 1859 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:30.107197 update_engine[1312]: I0517 00:43:30.106412 1312 update_attempter.cc:509] Updating boot flags... May 17 00:43:33.001675 kubelet[1859]: E0517 00:43:33.001620 1859 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" not found" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:33.126559 kubelet[1859]: I0517 00:43:33.126513 1859 kubelet_node_status.go:75] "Successfully registered node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:33.175984 kubelet[1859]: I0517 00:43:33.175942 1859 apiserver.go:52] "Watching apiserver" May 17 00:43:33.220744 kubelet[1859]: I0517 00:43:33.220689 1859 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" May 17 00:43:35.365116 systemd[1]: Reloading. May 17 00:43:35.584941 /usr/lib/systemd/system-generators/torcx-generator[2166]: time="2025-05-17T00:43:35Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.7 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.7 /var/lib/torcx/store]" May 17 00:43:35.584999 /usr/lib/systemd/system-generators/torcx-generator[2166]: time="2025-05-17T00:43:35Z" level=info msg="torcx already run" May 17 00:43:35.731894 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. May 17 00:43:35.731926 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. May 17 00:43:35.761220 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. May 17 00:43:35.932906 systemd[1]: Stopping kubelet.service... May 17 00:43:35.957433 systemd[1]: kubelet.service: Deactivated successfully. May 17 00:43:35.958311 systemd[1]: Stopped kubelet.service. May 17 00:43:35.964755 systemd[1]: Starting kubelet.service... May 17 00:43:36.277191 systemd[1]: Started kubelet.service. May 17 00:43:36.425060 kubelet[2224]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:43:36.425060 kubelet[2224]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. May 17 00:43:36.425716 kubelet[2224]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. May 17 00:43:36.425716 kubelet[2224]: I0517 00:43:36.425202 2224 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" May 17 00:43:36.430911 sudo[2235]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin May 17 00:43:36.432166 sudo[2235]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) May 17 00:43:36.442138 kubelet[2224]: I0517 00:43:36.442038 2224 server.go:491] "Kubelet version" kubeletVersion="v1.31.8" May 17 00:43:36.442138 kubelet[2224]: I0517 00:43:36.442083 2224 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" May 17 00:43:36.443379 kubelet[2224]: I0517 00:43:36.442661 2224 server.go:934] "Client rotation is on, will bootstrap in background" May 17 00:43:36.445247 kubelet[2224]: I0517 00:43:36.445205 2224 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". May 17 00:43:36.453605 kubelet[2224]: I0517 00:43:36.453554 2224 dynamic_cafile_content.go:160] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" May 17 00:43:36.463395 kubelet[2224]: E0517 00:43:36.462668 2224 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" May 17 00:43:36.463395 kubelet[2224]: I0517 00:43:36.462709 2224 server.go:1408] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." May 17 00:43:36.467394 kubelet[2224]: I0517 00:43:36.467337 2224 server.go:749] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" May 17 00:43:36.468150 kubelet[2224]: I0517 00:43:36.468121 2224 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" May 17 00:43:36.468415 kubelet[2224]: I0517 00:43:36.468376 2224 container_manager_linux.go:264] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] May 17 00:43:36.468711 kubelet[2224]: I0517 00:43:36.468421 2224 container_manager_linux.go:269] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":1} May 17 00:43:36.468906 kubelet[2224]: I0517 00:43:36.468732 2224 topology_manager.go:138] "Creating topology manager with none policy" May 17 00:43:36.468906 kubelet[2224]: I0517 00:43:36.468752 2224 container_manager_linux.go:300] "Creating device plugin manager" May 17 00:43:36.468906 kubelet[2224]: I0517 00:43:36.468797 2224 state_mem.go:36] "Initialized new in-memory state store" May 17 00:43:36.469089 kubelet[2224]: I0517 00:43:36.468944 2224 kubelet.go:408] "Attempting to sync node with API server" May 17 00:43:36.469089 kubelet[2224]: I0517 00:43:36.468966 2224 kubelet.go:303] "Adding static pod path" path="/etc/kubernetes/manifests" May 17 00:43:36.469089 kubelet[2224]: I0517 00:43:36.469007 2224 kubelet.go:314] "Adding apiserver pod source" May 17 00:43:36.469089 kubelet[2224]: I0517 00:43:36.469025 2224 apiserver.go:42] "Waiting for node sync before watching apiserver pods" May 17 00:43:36.510310 kubelet[2224]: I0517 00:43:36.510269 2224 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" May 17 00:43:36.511129 kubelet[2224]: I0517 00:43:36.511097 2224 kubelet.go:837] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" May 17 00:43:36.511816 kubelet[2224]: I0517 00:43:36.511789 2224 server.go:1274] "Started kubelet" May 17 00:43:36.514510 kubelet[2224]: I0517 00:43:36.514476 2224 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" May 17 00:43:36.529584 kubelet[2224]: E0517 00:43:36.529443 2224 kubelet.go:1478] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" May 17 00:43:36.535174 kubelet[2224]: I0517 00:43:36.535114 2224 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 May 17 00:43:36.537205 kubelet[2224]: I0517 00:43:36.537168 2224 server.go:449] "Adding debug handlers to kubelet server" May 17 00:43:36.538706 kubelet[2224]: I0517 00:43:36.538649 2224 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 May 17 00:43:36.538989 kubelet[2224]: I0517 00:43:36.538963 2224 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" May 17 00:43:36.539387 kubelet[2224]: I0517 00:43:36.539337 2224 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" May 17 00:43:36.541409 kubelet[2224]: I0517 00:43:36.541379 2224 volume_manager.go:289] "Starting Kubelet Volume Manager" May 17 00:43:36.545362 kubelet[2224]: I0517 00:43:36.545313 2224 desired_state_of_world_populator.go:147] "Desired state populator starts to run" May 17 00:43:36.546112 kubelet[2224]: I0517 00:43:36.545657 2224 reconciler.go:26] "Reconciler: start to sync state" May 17 00:43:36.555017 kubelet[2224]: I0517 00:43:36.554978 2224 factory.go:221] Registration of the systemd container factory successfully May 17 00:43:36.555206 kubelet[2224]: I0517 00:43:36.555158 2224 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory May 17 00:43:36.559893 kubelet[2224]: I0517 00:43:36.559857 2224 factory.go:221] Registration of the containerd container factory successfully May 17 00:43:36.600051 kubelet[2224]: I0517 00:43:36.599979 2224 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" May 17 00:43:36.604715 kubelet[2224]: I0517 00:43:36.604668 2224 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" May 17 00:43:36.604715 kubelet[2224]: I0517 00:43:36.604717 2224 status_manager.go:217] "Starting to sync pod status with apiserver" May 17 00:43:36.604973 kubelet[2224]: I0517 00:43:36.604751 2224 kubelet.go:2321] "Starting kubelet main sync loop" May 17 00:43:36.604973 kubelet[2224]: E0517 00:43:36.604854 2224 kubelet.go:2345] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" May 17 00:43:36.705806 kubelet[2224]: E0517 00:43:36.705759 2224 kubelet.go:2345] "Skipping pod synchronization" err="container runtime status check may not have completed yet" May 17 00:43:36.708568 kubelet[2224]: I0517 00:43:36.708534 2224 cpu_manager.go:214] "Starting CPU manager" policy="none" May 17 00:43:36.708568 kubelet[2224]: I0517 00:43:36.708565 2224 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" May 17 00:43:36.708816 kubelet[2224]: I0517 00:43:36.708595 2224 state_mem.go:36] "Initialized new in-memory state store" May 17 00:43:36.708894 kubelet[2224]: I0517 00:43:36.708878 2224 state_mem.go:88] "Updated default CPUSet" cpuSet="" May 17 00:43:36.708958 kubelet[2224]: I0517 00:43:36.708897 2224 state_mem.go:96] "Updated CPUSet assignments" assignments={} May 17 00:43:36.708958 kubelet[2224]: I0517 00:43:36.708929 2224 policy_none.go:49] "None policy: Start" May 17 00:43:36.709834 kubelet[2224]: I0517 00:43:36.709804 2224 memory_manager.go:170] "Starting memorymanager" policy="None" May 17 00:43:36.709978 kubelet[2224]: I0517 00:43:36.709847 2224 state_mem.go:35] "Initializing new in-memory state store" May 17 00:43:36.710105 kubelet[2224]: I0517 00:43:36.710080 2224 state_mem.go:75] "Updated machine memory state" May 17 00:43:36.718916 kubelet[2224]: I0517 00:43:36.718873 2224 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" May 17 00:43:36.719181 kubelet[2224]: I0517 00:43:36.719152 2224 eviction_manager.go:189] "Eviction manager: starting control loop" May 17 00:43:36.719300 kubelet[2224]: I0517 00:43:36.719188 2224 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" May 17 00:43:36.724836 kubelet[2224]: I0517 00:43:36.724797 2224 plugin_manager.go:118] "Starting Kubelet Plugin Manager" May 17 00:43:36.844563 kubelet[2224]: I0517 00:43:36.844403 2224 kubelet_node_status.go:72] "Attempting to register node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.858881 kubelet[2224]: I0517 00:43:36.858821 2224 kubelet_node_status.go:111] "Node was previously registered" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.859088 kubelet[2224]: I0517 00:43:36.858960 2224 kubelet_node_status.go:75] "Successfully registered node" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.927823 kubelet[2224]: W0517 00:43:36.927770 2224 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] May 17 00:43:36.928433 kubelet[2224]: W0517 00:43:36.928407 2224 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] May 17 00:43:36.929192 kubelet[2224]: W0517 00:43:36.929164 2224 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] May 17 00:43:36.954003 kubelet[2224]: I0517 00:43:36.953938 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954228 kubelet[2224]: I0517 00:43:36.954073 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/2f54a6b75ee979f46c55e66226f79feb-kubeconfig\") pod \"kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"2f54a6b75ee979f46c55e66226f79feb\") " pod="kube-system/kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954228 kubelet[2224]: I0517 00:43:36.954142 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/da1027880b2f3ae58454bfe774080950-ca-certs\") pod \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"da1027880b2f3ae58454bfe774080950\") " pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954228 kubelet[2224]: I0517 00:43:36.954208 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/da1027880b2f3ae58454bfe774080950-k8s-certs\") pod \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"da1027880b2f3ae58454bfe774080950\") " pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954479 kubelet[2224]: I0517 00:43:36.954242 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/da1027880b2f3ae58454bfe774080950-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"da1027880b2f3ae58454bfe774080950\") " pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954479 kubelet[2224]: I0517 00:43:36.954381 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-k8s-certs\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954479 kubelet[2224]: I0517 00:43:36.954421 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-kubeconfig\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954700 kubelet[2224]: I0517 00:43:36.954497 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-ca-certs\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:36.954700 kubelet[2224]: I0517 00:43:36.954569 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/0ed4d1288c326e677089947e941e90b9-flexvolume-dir\") pod \"kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" (UID: \"0ed4d1288c326e677089947e941e90b9\") " pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:37.395484 sudo[2235]: pam_unix(sudo:session): session closed for user root May 17 00:43:37.486511 kubelet[2224]: I0517 00:43:37.486465 2224 apiserver.go:52] "Watching apiserver" May 17 00:43:37.546219 kubelet[2224]: I0517 00:43:37.546171 2224 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" May 17 00:43:37.673641 kubelet[2224]: W0517 00:43:37.673494 2224 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters] May 17 00:43:37.673837 kubelet[2224]: E0517 00:43:37.673696 2224 kubelet.go:1915] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" already exists" pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" May 17 00:43:37.773571 kubelet[2224]: I0517 00:43:37.773449 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" podStartSLOduration=1.773408265 podStartE2EDuration="1.773408265s" podCreationTimestamp="2025-05-17 00:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:43:37.735867259 +0000 UTC m=+1.428427925" watchObservedRunningTime="2025-05-17 00:43:37.773408265 +0000 UTC m=+1.465968935" May 17 00:43:37.773856 kubelet[2224]: I0517 00:43:37.773695 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" podStartSLOduration=1.773654278 podStartE2EDuration="1.773654278s" podCreationTimestamp="2025-05-17 00:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:43:37.772379209 +0000 UTC m=+1.464939878" watchObservedRunningTime="2025-05-17 00:43:37.773654278 +0000 UTC m=+1.466214948" May 17 00:43:37.833160 kubelet[2224]: I0517 00:43:37.832655 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" podStartSLOduration=1.832627542 podStartE2EDuration="1.832627542s" podCreationTimestamp="2025-05-17 00:43:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:43:37.808661049 +0000 UTC m=+1.501221716" watchObservedRunningTime="2025-05-17 00:43:37.832627542 +0000 UTC m=+1.525188213" May 17 00:43:40.004318 kubelet[2224]: I0517 00:43:40.004270 2224 kuberuntime_manager.go:1635] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" May 17 00:43:40.005202 env[1327]: time="2025-05-17T00:43:40.004934129Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." May 17 00:43:40.005853 kubelet[2224]: I0517 00:43:40.005383 2224 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" May 17 00:43:40.297565 sudo[1581]: pam_unix(sudo:session): session closed for user root May 17 00:43:40.342764 sshd[1577]: pam_unix(sshd:session): session closed for user core May 17 00:43:40.348663 systemd[1]: sshd@6-10.128.0.10:22-139.178.89.65:36046.service: Deactivated successfully. May 17 00:43:40.349966 systemd[1]: session-7.scope: Deactivated successfully. May 17 00:43:40.350460 systemd-logind[1310]: Session 7 logged out. Waiting for processes to exit. May 17 00:43:40.353215 systemd-logind[1310]: Removed session 7. May 17 00:43:40.752015 kubelet[2224]: W0517 00:43:40.751873 2224 reflector.go:561] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9' and this object May 17 00:43:40.752015 kubelet[2224]: E0517 00:43:40.751936 2224 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"cilium-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"cilium-config\" is forbidden: User \"system:node:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9' and this object" logger="UnhandledError" May 17 00:43:40.753037 kubelet[2224]: W0517 00:43:40.752996 2224 reflector.go:561] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9' and this object May 17 00:43:40.753224 kubelet[2224]: E0517 00:43:40.753053 2224 reflector.go:158] "Unhandled Error" err="object-\"kube-system\"/\"hubble-server-certs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"hubble-server-certs\" is forbidden: User \"system:node:ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" cannot list resource \"secrets\" in API group \"\" in the namespace \"kube-system\": no relationship found between node 'ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9' and this object" logger="UnhandledError" May 17 00:43:40.780609 kubelet[2224]: I0517 00:43:40.780546 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-hostproc\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.780844 kubelet[2224]: I0517 00:43:40.780623 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-xtables-lock\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.780844 kubelet[2224]: I0517 00:43:40.780659 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aec630c4-dcf2-429b-a0f2-de98af83b20d-clustermesh-secrets\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.780844 kubelet[2224]: I0517 00:43:40.780693 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-config-path\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.780844 kubelet[2224]: I0517 00:43:40.780723 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-kernel\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.780844 kubelet[2224]: I0517 00:43:40.780754 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3342ed97-2fe3-4b1a-8fdf-38442658169a-kube-proxy\") pod \"kube-proxy-8zjlb\" (UID: \"3342ed97-2fe3-4b1a-8fdf-38442658169a\") " pod="kube-system/kube-proxy-8zjlb" May 17 00:43:40.781172 kubelet[2224]: I0517 00:43:40.780782 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3342ed97-2fe3-4b1a-8fdf-38442658169a-lib-modules\") pod \"kube-proxy-8zjlb\" (UID: \"3342ed97-2fe3-4b1a-8fdf-38442658169a\") " pod="kube-system/kube-proxy-8zjlb" May 17 00:43:40.781172 kubelet[2224]: I0517 00:43:40.780811 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-bpf-maps\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781172 kubelet[2224]: I0517 00:43:40.780836 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cni-path\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781172 kubelet[2224]: I0517 00:43:40.780888 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-hubble-tls\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781172 kubelet[2224]: I0517 00:43:40.780918 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-lib-modules\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781172 kubelet[2224]: I0517 00:43:40.780951 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-etc-cni-netd\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781599 kubelet[2224]: I0517 00:43:40.780993 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwgn4\" (UniqueName: \"kubernetes.io/projected/3342ed97-2fe3-4b1a-8fdf-38442658169a-kube-api-access-qwgn4\") pod \"kube-proxy-8zjlb\" (UID: \"3342ed97-2fe3-4b1a-8fdf-38442658169a\") " pod="kube-system/kube-proxy-8zjlb" May 17 00:43:40.781599 kubelet[2224]: I0517 00:43:40.781024 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-run\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781599 kubelet[2224]: I0517 00:43:40.781053 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-cgroup\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781599 kubelet[2224]: I0517 00:43:40.781096 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-net\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781599 kubelet[2224]: I0517 00:43:40.781139 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h47fm\" (UniqueName: \"kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-kube-api-access-h47fm\") pod \"cilium-px5wt\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " pod="kube-system/cilium-px5wt" May 17 00:43:40.781909 kubelet[2224]: I0517 00:43:40.781174 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3342ed97-2fe3-4b1a-8fdf-38442658169a-xtables-lock\") pod \"kube-proxy-8zjlb\" (UID: \"3342ed97-2fe3-4b1a-8fdf-38442658169a\") " pod="kube-system/kube-proxy-8zjlb" May 17 00:43:40.808843 kubelet[2224]: E0517 00:43:40.808781 2224 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-h47fm lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-px5wt" podUID="aec630c4-dcf2-429b-a0f2-de98af83b20d" May 17 00:43:40.882966 kubelet[2224]: I0517 00:43:40.882916 2224 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" May 17 00:43:41.018269 env[1327]: time="2025-05-17T00:43:41.018183257Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8zjlb,Uid:3342ed97-2fe3-4b1a-8fdf-38442658169a,Namespace:kube-system,Attempt:0,}" May 17 00:43:41.058931 env[1327]: time="2025-05-17T00:43:41.058622936Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:43:41.058931 env[1327]: time="2025-05-17T00:43:41.058698271Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:43:41.058931 env[1327]: time="2025-05-17T00:43:41.058721287Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:43:41.062278 env[1327]: time="2025-05-17T00:43:41.062125974Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/42b1c126be3d86f2c1a4f99222054aadab15f0cc0880ffa85614f4d441f36bbf pid=2306 runtime=io.containerd.runc.v2 May 17 00:43:41.171587 env[1327]: time="2025-05-17T00:43:41.171520352Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-8zjlb,Uid:3342ed97-2fe3-4b1a-8fdf-38442658169a,Namespace:kube-system,Attempt:0,} returns sandbox id \"42b1c126be3d86f2c1a4f99222054aadab15f0cc0880ffa85614f4d441f36bbf\"" May 17 00:43:41.176214 env[1327]: time="2025-05-17T00:43:41.176157037Z" level=info msg="CreateContainer within sandbox \"42b1c126be3d86f2c1a4f99222054aadab15f0cc0880ffa85614f4d441f36bbf\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" May 17 00:43:41.203512 env[1327]: time="2025-05-17T00:43:41.203413687Z" level=info msg="CreateContainer within sandbox \"42b1c126be3d86f2c1a4f99222054aadab15f0cc0880ffa85614f4d441f36bbf\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"994658bbda53f4c61f0e0b15405a3ade89d10126602af140bb823e4f1b28ecba\"" May 17 00:43:41.204559 env[1327]: time="2025-05-17T00:43:41.204511571Z" level=info msg="StartContainer for \"994658bbda53f4c61f0e0b15405a3ade89d10126602af140bb823e4f1b28ecba\"" May 17 00:43:41.285170 kubelet[2224]: I0517 00:43:41.284913 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/546317f2-a52c-4374-b0d6-e64927e14f57-cilium-config-path\") pod \"cilium-operator-5d85765b45-2l94x\" (UID: \"546317f2-a52c-4374-b0d6-e64927e14f57\") " pod="kube-system/cilium-operator-5d85765b45-2l94x" May 17 00:43:41.286031 kubelet[2224]: I0517 00:43:41.285194 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94kvj\" (UniqueName: \"kubernetes.io/projected/546317f2-a52c-4374-b0d6-e64927e14f57-kube-api-access-94kvj\") pod \"cilium-operator-5d85765b45-2l94x\" (UID: \"546317f2-a52c-4374-b0d6-e64927e14f57\") " pod="kube-system/cilium-operator-5d85765b45-2l94x" May 17 00:43:41.360553 env[1327]: time="2025-05-17T00:43:41.360481876Z" level=info msg="StartContainer for \"994658bbda53f4c61f0e0b15405a3ade89d10126602af140bb823e4f1b28ecba\" returns successfully" May 17 00:43:41.697640 kubelet[2224]: I0517 00:43:41.697462 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-8zjlb" podStartSLOduration=1.697437702 podStartE2EDuration="1.697437702s" podCreationTimestamp="2025-05-17 00:43:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:43:41.697399368 +0000 UTC m=+5.389960035" watchObservedRunningTime="2025-05-17 00:43:41.697437702 +0000 UTC m=+5.389998348" May 17 00:43:41.793826 kubelet[2224]: I0517 00:43:41.793770 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-etc-cni-netd\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794026 kubelet[2224]: I0517 00:43:41.793856 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-bpf-maps\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794026 kubelet[2224]: I0517 00:43:41.793883 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cni-path\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794026 kubelet[2224]: I0517 00:43:41.793912 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-net\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794026 kubelet[2224]: I0517 00:43:41.793940 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-lib-modules\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794026 kubelet[2224]: I0517 00:43:41.793968 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-run\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794026 kubelet[2224]: I0517 00:43:41.793996 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-kernel\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794434 kubelet[2224]: I0517 00:43:41.794043 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aec630c4-dcf2-429b-a0f2-de98af83b20d-clustermesh-secrets\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794434 kubelet[2224]: I0517 00:43:41.794073 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-cgroup\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794434 kubelet[2224]: I0517 00:43:41.794110 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-hostproc\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794434 kubelet[2224]: I0517 00:43:41.794140 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-xtables-lock\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794434 kubelet[2224]: I0517 00:43:41.794177 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h47fm\" (UniqueName: \"kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-kube-api-access-h47fm\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:41.794823 kubelet[2224]: I0517 00:43:41.794803 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.794895 kubelet[2224]: I0517 00:43:41.794857 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.794895 kubelet[2224]: I0517 00:43:41.794887 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cni-path" (OuterVolumeSpecName: "cni-path") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795022 kubelet[2224]: I0517 00:43:41.794914 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795022 kubelet[2224]: I0517 00:43:41.794944 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795022 kubelet[2224]: I0517 00:43:41.794968 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795022 kubelet[2224]: I0517 00:43:41.794997 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795503 kubelet[2224]: I0517 00:43:41.795462 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795627 kubelet[2224]: I0517 00:43:41.795521 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.795627 kubelet[2224]: I0517 00:43:41.795557 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-hostproc" (OuterVolumeSpecName: "hostproc") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:43:41.800879 kubelet[2224]: I0517 00:43:41.800820 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aec630c4-dcf2-429b-a0f2-de98af83b20d-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 17 00:43:41.803574 kubelet[2224]: I0517 00:43:41.803528 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-kube-api-access-h47fm" (OuterVolumeSpecName: "kube-api-access-h47fm") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "kube-api-access-h47fm". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:43:41.883029 kubelet[2224]: E0517 00:43:41.882969 2224 projected.go:263] Couldn't get secret kube-system/hubble-server-certs: failed to sync secret cache: timed out waiting for the condition May 17 00:43:41.883029 kubelet[2224]: E0517 00:43:41.883027 2224 projected.go:194] Error preparing data for projected volume hubble-tls for pod kube-system/cilium-px5wt: failed to sync secret cache: timed out waiting for the condition May 17 00:43:41.883389 kubelet[2224]: E0517 00:43:41.883145 2224 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-hubble-tls podName:aec630c4-dcf2-429b-a0f2-de98af83b20d nodeName:}" failed. No retries permitted until 2025-05-17 00:43:42.383110974 +0000 UTC m=+6.075671630 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "hubble-tls" (UniqueName: "kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-hubble-tls") pod "cilium-px5wt" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d") : failed to sync secret cache: timed out waiting for the condition May 17 00:43:41.883389 kubelet[2224]: E0517 00:43:41.883337 2224 configmap.go:193] Couldn't get configMap kube-system/cilium-config: failed to sync configmap cache: timed out waiting for the condition May 17 00:43:41.883614 kubelet[2224]: E0517 00:43:41.883423 2224 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-config-path podName:aec630c4-dcf2-429b-a0f2-de98af83b20d nodeName:}" failed. No retries permitted until 2025-05-17 00:43:42.38340475 +0000 UTC m=+6.075965408 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-config-path" (UniqueName: "kubernetes.io/configmap/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-config-path") pod "cilium-px5wt" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d") : failed to sync configmap cache: timed out waiting for the condition May 17 00:43:41.897685 kubelet[2224]: I0517 00:43:41.897619 2224 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-etc-cni-netd\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897685 kubelet[2224]: I0517 00:43:41.897687 2224 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-bpf-maps\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897710 2224 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cni-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897727 2224 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-net\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897746 2224 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-lib-modules\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897765 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-run\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897782 2224 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-host-proc-sys-kernel\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897803 2224 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/aec630c4-dcf2-429b-a0f2-de98af83b20d-clustermesh-secrets\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.897944 kubelet[2224]: I0517 00:43:41.897822 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-cgroup\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.898473 kubelet[2224]: I0517 00:43:41.897837 2224 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-hostproc\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.898473 kubelet[2224]: I0517 00:43:41.897854 2224 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/aec630c4-dcf2-429b-a0f2-de98af83b20d-xtables-lock\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.898473 kubelet[2224]: I0517 00:43:41.897871 2224 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h47fm\" (UniqueName: \"kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-kube-api-access-h47fm\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:41.904778 systemd[1]: var-lib-kubelet-pods-aec630c4\x2ddcf2\x2d429b\x2da0f2\x2dde98af83b20d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dh47fm.mount: Deactivated successfully. May 17 00:43:41.905075 systemd[1]: var-lib-kubelet-pods-aec630c4\x2ddcf2\x2d429b\x2da0f2\x2dde98af83b20d-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 17 00:43:42.433563 env[1327]: time="2025-05-17T00:43:42.433477557Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-2l94x,Uid:546317f2-a52c-4374-b0d6-e64927e14f57,Namespace:kube-system,Attempt:0,}" May 17 00:43:42.469225 env[1327]: time="2025-05-17T00:43:42.468892073Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:43:42.469225 env[1327]: time="2025-05-17T00:43:42.468952299Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:43:42.469225 env[1327]: time="2025-05-17T00:43:42.468974412Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:43:42.469798 env[1327]: time="2025-05-17T00:43:42.469662564Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1 pid=2516 runtime=io.containerd.runc.v2 May 17 00:43:42.509075 kubelet[2224]: I0517 00:43:42.501135 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-config-path\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:42.509075 kubelet[2224]: I0517 00:43:42.501193 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-hubble-tls\") pod \"aec630c4-dcf2-429b-a0f2-de98af83b20d\" (UID: \"aec630c4-dcf2-429b-a0f2-de98af83b20d\") " May 17 00:43:42.515080 kubelet[2224]: I0517 00:43:42.515005 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 17 00:43:42.522952 kubelet[2224]: I0517 00:43:42.522886 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "aec630c4-dcf2-429b-a0f2-de98af83b20d" (UID: "aec630c4-dcf2-429b-a0f2-de98af83b20d"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:43:42.583016 env[1327]: time="2025-05-17T00:43:42.582944108Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5d85765b45-2l94x,Uid:546317f2-a52c-4374-b0d6-e64927e14f57,Namespace:kube-system,Attempt:0,} returns sandbox id \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\"" May 17 00:43:42.585751 env[1327]: time="2025-05-17T00:43:42.585703675Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" May 17 00:43:42.602299 kubelet[2224]: I0517 00:43:42.602227 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/aec630c4-dcf2-429b-a0f2-de98af83b20d-cilium-config-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:42.602299 kubelet[2224]: I0517 00:43:42.602305 2224 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/aec630c4-dcf2-429b-a0f2-de98af83b20d-hubble-tls\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:43:42.803072 kubelet[2224]: I0517 00:43:42.802985 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-run\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803318 kubelet[2224]: I0517 00:43:42.803102 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-net\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803318 kubelet[2224]: I0517 00:43:42.803177 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-xtables-lock\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803318 kubelet[2224]: I0517 00:43:42.803210 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-cgroup\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803318 kubelet[2224]: I0517 00:43:42.803281 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cni-path\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803318 kubelet[2224]: I0517 00:43:42.803305 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/38a4631e-531e-44c6-933e-1b9cfdebcdec-clustermesh-secrets\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803682 kubelet[2224]: I0517 00:43:42.803382 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-bpf-maps\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803682 kubelet[2224]: I0517 00:43:42.803455 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-hostproc\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803682 kubelet[2224]: I0517 00:43:42.803487 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-lib-modules\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803682 kubelet[2224]: I0517 00:43:42.803564 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-config-path\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803682 kubelet[2224]: I0517 00:43:42.803601 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-kernel\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.803682 kubelet[2224]: I0517 00:43:42.803674 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-hubble-tls\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.804079 kubelet[2224]: I0517 00:43:42.803762 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88sf4\" (UniqueName: \"kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-kube-api-access-88sf4\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.804079 kubelet[2224]: I0517 00:43:42.803837 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-etc-cni-netd\") pod \"cilium-wdsmk\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " pod="kube-system/cilium-wdsmk" May 17 00:43:42.897331 systemd[1]: run-containerd-runc-k8s.io-7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1-runc.RizsDs.mount: Deactivated successfully. May 17 00:43:42.897623 systemd[1]: var-lib-kubelet-pods-aec630c4\x2ddcf2\x2d429b\x2da0f2\x2dde98af83b20d-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 17 00:43:43.057733 env[1327]: time="2025-05-17T00:43:43.056290069Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wdsmk,Uid:38a4631e-531e-44c6-933e-1b9cfdebcdec,Namespace:kube-system,Attempt:0,}" May 17 00:43:43.100223 env[1327]: time="2025-05-17T00:43:43.100067002Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:43:43.100505 env[1327]: time="2025-05-17T00:43:43.100245755Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:43:43.100505 env[1327]: time="2025-05-17T00:43:43.100307698Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:43:43.100704 env[1327]: time="2025-05-17T00:43:43.100634669Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a pid=2564 runtime=io.containerd.runc.v2 May 17 00:43:43.174669 env[1327]: time="2025-05-17T00:43:43.174567372Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-wdsmk,Uid:38a4631e-531e-44c6-933e-1b9cfdebcdec,Namespace:kube-system,Attempt:0,} returns sandbox id \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\"" May 17 00:43:44.608574 kubelet[2224]: I0517 00:43:44.608523 2224 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aec630c4-dcf2-429b-a0f2-de98af83b20d" path="/var/lib/kubelet/pods/aec630c4-dcf2-429b-a0f2-de98af83b20d/volumes" May 17 00:43:44.934580 env[1327]: time="2025-05-17T00:43:44.934424731Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:44.937657 env[1327]: time="2025-05-17T00:43:44.937611429Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:44.940333 env[1327]: time="2025-05-17T00:43:44.940290359Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:44.941261 env[1327]: time="2025-05-17T00:43:44.941177974Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" May 17 00:43:44.944583 env[1327]: time="2025-05-17T00:43:44.944528055Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" May 17 00:43:44.947365 env[1327]: time="2025-05-17T00:43:44.947275242Z" level=info msg="CreateContainer within sandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" May 17 00:43:44.975541 env[1327]: time="2025-05-17T00:43:44.975468687Z" level=info msg="CreateContainer within sandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\"" May 17 00:43:44.977598 env[1327]: time="2025-05-17T00:43:44.977552328Z" level=info msg="StartContainer for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\"" May 17 00:43:45.073151 env[1327]: time="2025-05-17T00:43:45.073086285Z" level=info msg="StartContainer for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" returns successfully" May 17 00:43:46.207105 kubelet[2224]: I0517 00:43:46.207008 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-5d85765b45-2l94x" podStartSLOduration=2.848802076 podStartE2EDuration="5.206978463s" podCreationTimestamp="2025-05-17 00:43:41 +0000 UTC" firstStartedPulling="2025-05-17 00:43:42.584952211 +0000 UTC m=+6.277512852" lastFinishedPulling="2025-05-17 00:43:44.943128582 +0000 UTC m=+8.635689239" observedRunningTime="2025-05-17 00:43:45.833869358 +0000 UTC m=+9.526430026" watchObservedRunningTime="2025-05-17 00:43:46.206978463 +0000 UTC m=+9.899539120" May 17 00:43:51.777514 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3232343137.mount: Deactivated successfully. May 17 00:43:55.392998 env[1327]: time="2025-05-17T00:43:55.392857612Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:55.397804 env[1327]: time="2025-05-17T00:43:55.397722627Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:55.400514 env[1327]: time="2025-05-17T00:43:55.400453654Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" May 17 00:43:55.402367 env[1327]: time="2025-05-17T00:43:55.402255600Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" May 17 00:43:55.418113 env[1327]: time="2025-05-17T00:43:55.417933466Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:43:55.439976 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3788057321.mount: Deactivated successfully. May 17 00:43:55.456541 env[1327]: time="2025-05-17T00:43:55.456490696Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\"" May 17 00:43:55.458630 env[1327]: time="2025-05-17T00:43:55.457690591Z" level=info msg="StartContainer for \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\"" May 17 00:43:55.555629 env[1327]: time="2025-05-17T00:43:55.555555443Z" level=info msg="StartContainer for \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\" returns successfully" May 17 00:43:56.431436 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046-rootfs.mount: Deactivated successfully. May 17 00:43:56.898670 env[1327]: time="2025-05-17T00:43:56.898554271Z" level=error msg="collecting metrics for d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046" error="cgroups: cgroup deleted: unknown" May 17 00:43:57.837505 env[1327]: time="2025-05-17T00:43:57.837416283Z" level=info msg="shim disconnected" id=d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046 May 17 00:43:57.837505 env[1327]: time="2025-05-17T00:43:57.837505860Z" level=warning msg="cleaning up after shim disconnected" id=d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046 namespace=k8s.io May 17 00:43:57.837973 env[1327]: time="2025-05-17T00:43:57.837545284Z" level=info msg="cleaning up dead shim" May 17 00:43:57.851250 env[1327]: time="2025-05-17T00:43:57.851185425Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:43:57Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2687 runtime=io.containerd.runc.v2\n" May 17 00:43:58.781792 env[1327]: time="2025-05-17T00:43:58.781703862Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 17 00:43:58.819307 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4082806011.mount: Deactivated successfully. May 17 00:43:58.823959 env[1327]: time="2025-05-17T00:43:58.820361132Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\"" May 17 00:43:58.833583 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4244212150.mount: Deactivated successfully. May 17 00:43:58.837619 env[1327]: time="2025-05-17T00:43:58.836986595Z" level=info msg="StartContainer for \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\"" May 17 00:43:58.934125 env[1327]: time="2025-05-17T00:43:58.933018532Z" level=info msg="StartContainer for \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\" returns successfully" May 17 00:43:58.948624 systemd[1]: systemd-sysctl.service: Deactivated successfully. May 17 00:43:58.949163 systemd[1]: Stopped systemd-sysctl.service. May 17 00:43:58.950265 systemd[1]: Stopping systemd-sysctl.service... May 17 00:43:58.955564 systemd[1]: Starting systemd-sysctl.service... May 17 00:43:58.983566 systemd[1]: Finished systemd-sysctl.service. May 17 00:43:59.003310 env[1327]: time="2025-05-17T00:43:59.003239459Z" level=info msg="shim disconnected" id=9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91 May 17 00:43:59.003310 env[1327]: time="2025-05-17T00:43:59.003313772Z" level=warning msg="cleaning up after shim disconnected" id=9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91 namespace=k8s.io May 17 00:43:59.003753 env[1327]: time="2025-05-17T00:43:59.003331061Z" level=info msg="cleaning up dead shim" May 17 00:43:59.017630 env[1327]: time="2025-05-17T00:43:59.017554725Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:43:59Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2756 runtime=io.containerd.runc.v2\n" May 17 00:43:59.789379 env[1327]: time="2025-05-17T00:43:59.785905411Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 17 00:43:59.800524 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91-rootfs.mount: Deactivated successfully. May 17 00:43:59.827362 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4071708494.mount: Deactivated successfully. May 17 00:43:59.831508 env[1327]: time="2025-05-17T00:43:59.831439878Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\"" May 17 00:43:59.834019 env[1327]: time="2025-05-17T00:43:59.832500587Z" level=info msg="StartContainer for \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\"" May 17 00:43:59.944606 env[1327]: time="2025-05-17T00:43:59.944489245Z" level=info msg="StartContainer for \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\" returns successfully" May 17 00:43:59.992113 env[1327]: time="2025-05-17T00:43:59.991999477Z" level=info msg="shim disconnected" id=c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1 May 17 00:43:59.992113 env[1327]: time="2025-05-17T00:43:59.992106547Z" level=warning msg="cleaning up after shim disconnected" id=c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1 namespace=k8s.io May 17 00:43:59.992508 env[1327]: time="2025-05-17T00:43:59.992125492Z" level=info msg="cleaning up dead shim" May 17 00:44:00.007382 env[1327]: time="2025-05-17T00:44:00.007278543Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:43:59Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2811 runtime=io.containerd.runc.v2\n" May 17 00:44:00.788070 env[1327]: time="2025-05-17T00:44:00.787600914Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 17 00:44:00.802327 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1-rootfs.mount: Deactivated successfully. May 17 00:44:00.833934 env[1327]: time="2025-05-17T00:44:00.833875013Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\"" May 17 00:44:00.836673 env[1327]: time="2025-05-17T00:44:00.836601274Z" level=info msg="StartContainer for \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\"" May 17 00:44:00.930807 env[1327]: time="2025-05-17T00:44:00.930741350Z" level=info msg="StartContainer for \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\" returns successfully" May 17 00:44:00.968525 env[1327]: time="2025-05-17T00:44:00.968413943Z" level=info msg="shim disconnected" id=cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2 May 17 00:44:00.968525 env[1327]: time="2025-05-17T00:44:00.968529519Z" level=warning msg="cleaning up after shim disconnected" id=cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2 namespace=k8s.io May 17 00:44:00.969092 env[1327]: time="2025-05-17T00:44:00.968548631Z" level=info msg="cleaning up dead shim" May 17 00:44:00.982539 env[1327]: time="2025-05-17T00:44:00.982396749Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:44:00Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2868 runtime=io.containerd.runc.v2\n" May 17 00:44:01.803416 env[1327]: time="2025-05-17T00:44:01.793703386Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 17 00:44:01.801801 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2-rootfs.mount: Deactivated successfully. May 17 00:44:01.847430 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3332150565.mount: Deactivated successfully. May 17 00:44:01.868951 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2945488134.mount: Deactivated successfully. May 17 00:44:01.870450 env[1327]: time="2025-05-17T00:44:01.870372364Z" level=info msg="CreateContainer within sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\"" May 17 00:44:01.873959 env[1327]: time="2025-05-17T00:44:01.873881212Z" level=info msg="StartContainer for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\"" May 17 00:44:01.980538 env[1327]: time="2025-05-17T00:44:01.980462236Z" level=info msg="StartContainer for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" returns successfully" May 17 00:44:02.188555 kubelet[2224]: I0517 00:44:02.187190 2224 kubelet_node_status.go:488] "Fast updating node status as it just became ready" May 17 00:44:02.448960 kubelet[2224]: I0517 00:44:02.448803 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bwxh\" (UniqueName: \"kubernetes.io/projected/19683d25-b34e-4661-98ab-e5edf456a9bf-kube-api-access-9bwxh\") pod \"coredns-7c65d6cfc9-z4kps\" (UID: \"19683d25-b34e-4661-98ab-e5edf456a9bf\") " pod="kube-system/coredns-7c65d6cfc9-z4kps" May 17 00:44:02.449454 kubelet[2224]: I0517 00:44:02.449414 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qjmj\" (UniqueName: \"kubernetes.io/projected/ad7e1eca-d168-4377-852c-25e536fe3a24-kube-api-access-7qjmj\") pod \"coredns-7c65d6cfc9-qtnnf\" (UID: \"ad7e1eca-d168-4377-852c-25e536fe3a24\") " pod="kube-system/coredns-7c65d6cfc9-qtnnf" May 17 00:44:02.449720 kubelet[2224]: I0517 00:44:02.449681 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad7e1eca-d168-4377-852c-25e536fe3a24-config-volume\") pod \"coredns-7c65d6cfc9-qtnnf\" (UID: \"ad7e1eca-d168-4377-852c-25e536fe3a24\") " pod="kube-system/coredns-7c65d6cfc9-qtnnf" May 17 00:44:02.449952 kubelet[2224]: I0517 00:44:02.449898 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/19683d25-b34e-4661-98ab-e5edf456a9bf-config-volume\") pod \"coredns-7c65d6cfc9-z4kps\" (UID: \"19683d25-b34e-4661-98ab-e5edf456a9bf\") " pod="kube-system/coredns-7c65d6cfc9-z4kps" May 17 00:44:02.641832 env[1327]: time="2025-05-17T00:44:02.641295947Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-z4kps,Uid:19683d25-b34e-4661-98ab-e5edf456a9bf,Namespace:kube-system,Attempt:0,}" May 17 00:44:02.643798 env[1327]: time="2025-05-17T00:44:02.643126560Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-qtnnf,Uid:ad7e1eca-d168-4377-852c-25e536fe3a24,Namespace:kube-system,Attempt:0,}" May 17 00:44:02.871672 kubelet[2224]: I0517 00:44:02.871572 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-wdsmk" podStartSLOduration=8.644990517 podStartE2EDuration="20.871541614s" podCreationTimestamp="2025-05-17 00:43:42 +0000 UTC" firstStartedPulling="2025-05-17 00:43:43.177844495 +0000 UTC m=+6.870405157" lastFinishedPulling="2025-05-17 00:43:55.404395597 +0000 UTC m=+19.096956254" observedRunningTime="2025-05-17 00:44:02.868017929 +0000 UTC m=+26.560578596" watchObservedRunningTime="2025-05-17 00:44:02.871541614 +0000 UTC m=+26.564102282" May 17 00:44:04.514243 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready May 17 00:44:04.514659 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready May 17 00:44:04.500714 systemd-networkd[1079]: cilium_host: Link UP May 17 00:44:04.501060 systemd-networkd[1079]: cilium_net: Link UP May 17 00:44:04.507121 systemd-networkd[1079]: cilium_net: Gained carrier May 17 00:44:04.517980 systemd-networkd[1079]: cilium_host: Gained carrier May 17 00:44:04.519672 systemd-networkd[1079]: cilium_net: Gained IPv6LL May 17 00:44:04.667568 systemd-networkd[1079]: cilium_host: Gained IPv6LL May 17 00:44:04.684924 systemd-networkd[1079]: cilium_vxlan: Link UP May 17 00:44:04.684937 systemd-networkd[1079]: cilium_vxlan: Gained carrier May 17 00:44:04.976455 kernel: NET: Registered PF_ALG protocol family May 17 00:44:05.958719 systemd-networkd[1079]: lxc_health: Link UP May 17 00:44:05.996423 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready May 17 00:44:06.000822 systemd-networkd[1079]: lxc_health: Gained carrier May 17 00:44:06.028205 systemd-networkd[1079]: cilium_vxlan: Gained IPv6LL May 17 00:44:06.227479 systemd-networkd[1079]: lxc40a7e43896ce: Link UP May 17 00:44:06.247379 kernel: eth0: renamed from tmp3c05f May 17 00:44:06.260942 systemd-networkd[1079]: lxc82a1e7e33e11: Link UP May 17 00:44:06.270572 kernel: eth0: renamed from tmp0c50e May 17 00:44:06.303961 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc40a7e43896ce: link becomes ready May 17 00:44:06.303499 systemd-networkd[1079]: lxc40a7e43896ce: Gained carrier May 17 00:44:06.316399 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc82a1e7e33e11: link becomes ready May 17 00:44:06.331301 systemd-networkd[1079]: lxc82a1e7e33e11: Gained carrier May 17 00:44:07.499606 systemd-networkd[1079]: lxc82a1e7e33e11: Gained IPv6LL May 17 00:44:07.627624 systemd-networkd[1079]: lxc_health: Gained IPv6LL May 17 00:44:08.140275 systemd-networkd[1079]: lxc40a7e43896ce: Gained IPv6LL May 17 00:44:11.869468 env[1327]: time="2025-05-17T00:44:11.869253968Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:44:11.870144 env[1327]: time="2025-05-17T00:44:11.869447340Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:44:11.870144 env[1327]: time="2025-05-17T00:44:11.869765442Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:44:11.870331 env[1327]: time="2025-05-17T00:44:11.870179298Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3c05f9de35f027d6b4eb51b381c0d6e5fb019cd4a37046add27c4889689d8b59 pid=3423 runtime=io.containerd.runc.v2 May 17 00:44:11.901663 env[1327]: time="2025-05-17T00:44:11.901528494Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:44:11.901914 env[1327]: time="2025-05-17T00:44:11.901700162Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:44:11.901914 env[1327]: time="2025-05-17T00:44:11.901777118Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:44:11.902158 env[1327]: time="2025-05-17T00:44:11.902054604Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0c50e4ab6e46910699c7b9501a5390e501be4e41b98947e570acc94f70e9a01d pid=3435 runtime=io.containerd.runc.v2 May 17 00:44:11.985083 systemd[1]: run-containerd-runc-k8s.io-3c05f9de35f027d6b4eb51b381c0d6e5fb019cd4a37046add27c4889689d8b59-runc.o2j3WD.mount: Deactivated successfully. May 17 00:44:12.122130 env[1327]: time="2025-05-17T00:44:12.121941640Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-qtnnf,Uid:ad7e1eca-d168-4377-852c-25e536fe3a24,Namespace:kube-system,Attempt:0,} returns sandbox id \"3c05f9de35f027d6b4eb51b381c0d6e5fb019cd4a37046add27c4889689d8b59\"" May 17 00:44:12.129238 env[1327]: time="2025-05-17T00:44:12.129177349Z" level=info msg="CreateContainer within sandbox \"3c05f9de35f027d6b4eb51b381c0d6e5fb019cd4a37046add27c4889689d8b59\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 17 00:44:12.172368 env[1327]: time="2025-05-17T00:44:12.172280993Z" level=info msg="CreateContainer within sandbox \"3c05f9de35f027d6b4eb51b381c0d6e5fb019cd4a37046add27c4889689d8b59\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"ecfe29496b7436df409c6978d35d6637849092055a9d793ea59df710460dad4d\"" May 17 00:44:12.176289 env[1327]: time="2025-05-17T00:44:12.176131186Z" level=info msg="StartContainer for \"ecfe29496b7436df409c6978d35d6637849092055a9d793ea59df710460dad4d\"" May 17 00:44:12.200430 env[1327]: time="2025-05-17T00:44:12.198811864Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7c65d6cfc9-z4kps,Uid:19683d25-b34e-4661-98ab-e5edf456a9bf,Namespace:kube-system,Attempt:0,} returns sandbox id \"0c50e4ab6e46910699c7b9501a5390e501be4e41b98947e570acc94f70e9a01d\"" May 17 00:44:12.209407 env[1327]: time="2025-05-17T00:44:12.209112674Z" level=info msg="CreateContainer within sandbox \"0c50e4ab6e46910699c7b9501a5390e501be4e41b98947e570acc94f70e9a01d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" May 17 00:44:12.268324 env[1327]: time="2025-05-17T00:44:12.268260438Z" level=info msg="CreateContainer within sandbox \"0c50e4ab6e46910699c7b9501a5390e501be4e41b98947e570acc94f70e9a01d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"8fbaf458c5ce0ba704c16269048b371abd9778036a5af21c8264cbfa52d3c80e\"" May 17 00:44:12.272592 env[1327]: time="2025-05-17T00:44:12.272539390Z" level=info msg="StartContainer for \"8fbaf458c5ce0ba704c16269048b371abd9778036a5af21c8264cbfa52d3c80e\"" May 17 00:44:12.303159 env[1327]: time="2025-05-17T00:44:12.303061879Z" level=info msg="StartContainer for \"ecfe29496b7436df409c6978d35d6637849092055a9d793ea59df710460dad4d\" returns successfully" May 17 00:44:12.411075 env[1327]: time="2025-05-17T00:44:12.408130332Z" level=info msg="StartContainer for \"8fbaf458c5ce0ba704c16269048b371abd9778036a5af21c8264cbfa52d3c80e\" returns successfully" May 17 00:44:12.886013 systemd[1]: run-containerd-runc-k8s.io-0c50e4ab6e46910699c7b9501a5390e501be4e41b98947e570acc94f70e9a01d-runc.q6BER3.mount: Deactivated successfully. May 17 00:44:12.896771 kubelet[2224]: I0517 00:44:12.896690 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-z4kps" podStartSLOduration=31.896666661 podStartE2EDuration="31.896666661s" podCreationTimestamp="2025-05-17 00:43:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:44:12.895832113 +0000 UTC m=+36.588392783" watchObservedRunningTime="2025-05-17 00:44:12.896666661 +0000 UTC m=+36.589227329" May 17 00:44:12.952969 kubelet[2224]: I0517 00:44:12.952880 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7c65d6cfc9-qtnnf" podStartSLOduration=31.952833728 podStartE2EDuration="31.952833728s" podCreationTimestamp="2025-05-17 00:43:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:44:12.928609946 +0000 UTC m=+36.621170615" watchObservedRunningTime="2025-05-17 00:44:12.952833728 +0000 UTC m=+36.645394395" May 17 00:44:35.274848 systemd[1]: Started sshd@8-10.128.0.10:22-139.178.89.65:34480.service. May 17 00:44:35.575105 sshd[3580]: Accepted publickey for core from 139.178.89.65 port 34480 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:35.577900 sshd[3580]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:35.586544 systemd-logind[1310]: New session 8 of user core. May 17 00:44:35.586762 systemd[1]: Started session-8.scope. May 17 00:44:35.893077 sshd[3580]: pam_unix(sshd:session): session closed for user core May 17 00:44:35.898731 systemd[1]: sshd@8-10.128.0.10:22-139.178.89.65:34480.service: Deactivated successfully. May 17 00:44:35.900874 systemd[1]: session-8.scope: Deactivated successfully. May 17 00:44:35.901628 systemd-logind[1310]: Session 8 logged out. Waiting for processes to exit. May 17 00:44:35.904293 systemd-logind[1310]: Removed session 8. May 17 00:44:40.938448 systemd[1]: Started sshd@9-10.128.0.10:22-139.178.89.65:41358.service. May 17 00:44:41.229297 sshd[3595]: Accepted publickey for core from 139.178.89.65 port 41358 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:41.231792 sshd[3595]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:41.240651 systemd[1]: Started session-9.scope. May 17 00:44:41.242532 systemd-logind[1310]: New session 9 of user core. May 17 00:44:41.527144 sshd[3595]: pam_unix(sshd:session): session closed for user core May 17 00:44:41.532508 systemd[1]: sshd@9-10.128.0.10:22-139.178.89.65:41358.service: Deactivated successfully. May 17 00:44:41.534298 systemd[1]: session-9.scope: Deactivated successfully. May 17 00:44:41.537524 systemd-logind[1310]: Session 9 logged out. Waiting for processes to exit. May 17 00:44:41.539078 systemd-logind[1310]: Removed session 9. May 17 00:44:46.572709 systemd[1]: Started sshd@10-10.128.0.10:22-139.178.89.65:51146.service. May 17 00:44:46.865771 sshd[3610]: Accepted publickey for core from 139.178.89.65 port 51146 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:46.868195 sshd[3610]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:46.876607 systemd[1]: Started session-10.scope. May 17 00:44:46.878577 systemd-logind[1310]: New session 10 of user core. May 17 00:44:47.172544 sshd[3610]: pam_unix(sshd:session): session closed for user core May 17 00:44:47.178433 systemd[1]: sshd@10-10.128.0.10:22-139.178.89.65:51146.service: Deactivated successfully. May 17 00:44:47.180636 systemd[1]: session-10.scope: Deactivated successfully. May 17 00:44:47.180724 systemd-logind[1310]: Session 10 logged out. Waiting for processes to exit. May 17 00:44:47.183594 systemd-logind[1310]: Removed session 10. May 17 00:44:52.218749 systemd[1]: Started sshd@11-10.128.0.10:22-139.178.89.65:51156.service. May 17 00:44:52.514406 sshd[3623]: Accepted publickey for core from 139.178.89.65 port 51156 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:52.516658 sshd[3623]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:52.525337 systemd[1]: Started session-11.scope. May 17 00:44:52.526971 systemd-logind[1310]: New session 11 of user core. May 17 00:44:52.805946 sshd[3623]: pam_unix(sshd:session): session closed for user core May 17 00:44:52.811872 systemd[1]: sshd@11-10.128.0.10:22-139.178.89.65:51156.service: Deactivated successfully. May 17 00:44:52.814225 systemd[1]: session-11.scope: Deactivated successfully. May 17 00:44:52.814489 systemd-logind[1310]: Session 11 logged out. Waiting for processes to exit. May 17 00:44:52.817058 systemd-logind[1310]: Removed session 11. May 17 00:44:52.850284 systemd[1]: Started sshd@12-10.128.0.10:22-139.178.89.65:51160.service. May 17 00:44:53.145501 sshd[3637]: Accepted publickey for core from 139.178.89.65 port 51160 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:53.147465 sshd[3637]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:53.156333 systemd[1]: Started session-12.scope. May 17 00:44:53.156797 systemd-logind[1310]: New session 12 of user core. May 17 00:44:53.491698 sshd[3637]: pam_unix(sshd:session): session closed for user core May 17 00:44:53.497929 systemd[1]: sshd@12-10.128.0.10:22-139.178.89.65:51160.service: Deactivated successfully. May 17 00:44:53.500199 systemd[1]: session-12.scope: Deactivated successfully. May 17 00:44:53.500806 systemd-logind[1310]: Session 12 logged out. Waiting for processes to exit. May 17 00:44:53.504696 systemd-logind[1310]: Removed session 12. May 17 00:44:53.535947 systemd[1]: Started sshd@13-10.128.0.10:22-139.178.89.65:51170.service. May 17 00:44:53.832528 sshd[3647]: Accepted publickey for core from 139.178.89.65 port 51170 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:53.834890 sshd[3647]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:53.842436 systemd-logind[1310]: New session 13 of user core. May 17 00:44:53.843091 systemd[1]: Started session-13.scope. May 17 00:44:54.137433 sshd[3647]: pam_unix(sshd:session): session closed for user core May 17 00:44:54.142874 systemd[1]: sshd@13-10.128.0.10:22-139.178.89.65:51170.service: Deactivated successfully. May 17 00:44:54.145128 systemd-logind[1310]: Session 13 logged out. Waiting for processes to exit. May 17 00:44:54.145244 systemd[1]: session-13.scope: Deactivated successfully. May 17 00:44:54.149007 systemd-logind[1310]: Removed session 13. May 17 00:44:59.186705 systemd[1]: Started sshd@14-10.128.0.10:22-139.178.89.65:41442.service. May 17 00:44:59.493895 sshd[3660]: Accepted publickey for core from 139.178.89.65 port 41442 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:44:59.496672 sshd[3660]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:44:59.505215 systemd[1]: Started session-14.scope. May 17 00:44:59.505763 systemd-logind[1310]: New session 14 of user core. May 17 00:44:59.790475 sshd[3660]: pam_unix(sshd:session): session closed for user core May 17 00:44:59.796379 systemd[1]: sshd@14-10.128.0.10:22-139.178.89.65:41442.service: Deactivated successfully. May 17 00:44:59.797948 systemd[1]: session-14.scope: Deactivated successfully. May 17 00:44:59.798410 systemd-logind[1310]: Session 14 logged out. Waiting for processes to exit. May 17 00:44:59.800305 systemd-logind[1310]: Removed session 14. May 17 00:45:04.837656 systemd[1]: Started sshd@15-10.128.0.10:22-139.178.89.65:41456.service. May 17 00:45:05.135262 sshd[3673]: Accepted publickey for core from 139.178.89.65 port 41456 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:05.137601 sshd[3673]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:05.146947 systemd[1]: Started session-15.scope. May 17 00:45:05.148509 systemd-logind[1310]: New session 15 of user core. May 17 00:45:05.445218 sshd[3673]: pam_unix(sshd:session): session closed for user core May 17 00:45:05.451657 systemd[1]: sshd@15-10.128.0.10:22-139.178.89.65:41456.service: Deactivated successfully. May 17 00:45:05.453800 systemd[1]: session-15.scope: Deactivated successfully. May 17 00:45:05.454492 systemd-logind[1310]: Session 15 logged out. Waiting for processes to exit. May 17 00:45:05.458458 systemd-logind[1310]: Removed session 15. May 17 00:45:09.399471 systemd[1]: sshd@7-10.128.0.10:22-106.58.166.77:33302.service: Deactivated successfully. May 17 00:45:10.491072 systemd[1]: Started sshd@16-10.128.0.10:22-139.178.89.65:57018.service. May 17 00:45:10.788994 sshd[3688]: Accepted publickey for core from 139.178.89.65 port 57018 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:10.791305 sshd[3688]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:10.798710 systemd-logind[1310]: New session 16 of user core. May 17 00:45:10.799452 systemd[1]: Started session-16.scope. May 17 00:45:11.089676 sshd[3688]: pam_unix(sshd:session): session closed for user core May 17 00:45:11.095233 systemd[1]: sshd@16-10.128.0.10:22-139.178.89.65:57018.service: Deactivated successfully. May 17 00:45:11.097784 systemd[1]: session-16.scope: Deactivated successfully. May 17 00:45:11.098431 systemd-logind[1310]: Session 16 logged out. Waiting for processes to exit. May 17 00:45:11.101148 systemd-logind[1310]: Removed session 16. May 17 00:45:11.134077 systemd[1]: Started sshd@17-10.128.0.10:22-139.178.89.65:57032.service. May 17 00:45:11.426481 sshd[3701]: Accepted publickey for core from 139.178.89.65 port 57032 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:11.428684 sshd[3701]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:11.436850 systemd[1]: Started session-17.scope. May 17 00:45:11.437753 systemd-logind[1310]: New session 17 of user core. May 17 00:45:11.799429 sshd[3701]: pam_unix(sshd:session): session closed for user core May 17 00:45:11.804618 systemd[1]: sshd@17-10.128.0.10:22-139.178.89.65:57032.service: Deactivated successfully. May 17 00:45:11.807707 systemd[1]: session-17.scope: Deactivated successfully. May 17 00:45:11.808881 systemd-logind[1310]: Session 17 logged out. Waiting for processes to exit. May 17 00:45:11.811339 systemd-logind[1310]: Removed session 17. May 17 00:45:11.844759 systemd[1]: Started sshd@18-10.128.0.10:22-139.178.89.65:57048.service. May 17 00:45:12.139075 sshd[3713]: Accepted publickey for core from 139.178.89.65 port 57048 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:12.141307 sshd[3713]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:12.149508 systemd[1]: Started session-18.scope. May 17 00:45:12.149962 systemd-logind[1310]: New session 18 of user core. May 17 00:45:14.062309 sshd[3713]: pam_unix(sshd:session): session closed for user core May 17 00:45:14.069034 systemd-logind[1310]: Session 18 logged out. Waiting for processes to exit. May 17 00:45:14.071045 systemd[1]: sshd@18-10.128.0.10:22-139.178.89.65:57048.service: Deactivated successfully. May 17 00:45:14.072489 systemd[1]: session-18.scope: Deactivated successfully. May 17 00:45:14.075558 systemd-logind[1310]: Removed session 18. May 17 00:45:14.105656 systemd[1]: Started sshd@19-10.128.0.10:22-139.178.89.65:57060.service. May 17 00:45:14.402329 sshd[3731]: Accepted publickey for core from 139.178.89.65 port 57060 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:14.404266 sshd[3731]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:14.412463 systemd[1]: Started session-19.scope. May 17 00:45:14.415281 systemd-logind[1310]: New session 19 of user core. May 17 00:45:14.845228 sshd[3731]: pam_unix(sshd:session): session closed for user core May 17 00:45:14.850608 systemd[1]: sshd@19-10.128.0.10:22-139.178.89.65:57060.service: Deactivated successfully. May 17 00:45:14.852281 systemd[1]: session-19.scope: Deactivated successfully. May 17 00:45:14.856205 systemd-logind[1310]: Session 19 logged out. Waiting for processes to exit. May 17 00:45:14.858614 systemd-logind[1310]: Removed session 19. May 17 00:45:14.890388 systemd[1]: Started sshd@20-10.128.0.10:22-139.178.89.65:57066.service. May 17 00:45:15.185373 sshd[3741]: Accepted publickey for core from 139.178.89.65 port 57066 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:15.188336 sshd[3741]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:15.196251 systemd[1]: Started session-20.scope. May 17 00:45:15.197914 systemd-logind[1310]: New session 20 of user core. May 17 00:45:15.491014 sshd[3741]: pam_unix(sshd:session): session closed for user core May 17 00:45:15.496743 systemd[1]: sshd@20-10.128.0.10:22-139.178.89.65:57066.service: Deactivated successfully. May 17 00:45:15.498754 systemd[1]: session-20.scope: Deactivated successfully. May 17 00:45:15.498831 systemd-logind[1310]: Session 20 logged out. Waiting for processes to exit. May 17 00:45:15.501064 systemd-logind[1310]: Removed session 20. May 17 00:45:20.536648 systemd[1]: Started sshd@21-10.128.0.10:22-139.178.89.65:55618.service. May 17 00:45:20.835434 sshd[3754]: Accepted publickey for core from 139.178.89.65 port 55618 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:20.837806 sshd[3754]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:20.846607 systemd[1]: Started session-21.scope. May 17 00:45:20.848455 systemd-logind[1310]: New session 21 of user core. May 17 00:45:21.127294 sshd[3754]: pam_unix(sshd:session): session closed for user core May 17 00:45:21.140005 systemd[1]: sshd@21-10.128.0.10:22-139.178.89.65:55618.service: Deactivated successfully. May 17 00:45:21.142362 systemd[1]: session-21.scope: Deactivated successfully. May 17 00:45:21.142986 systemd-logind[1310]: Session 21 logged out. Waiting for processes to exit. May 17 00:45:21.146945 systemd-logind[1310]: Removed session 21. May 17 00:45:26.171855 systemd[1]: Started sshd@22-10.128.0.10:22-139.178.89.65:55620.service. May 17 00:45:26.467747 sshd[3770]: Accepted publickey for core from 139.178.89.65 port 55620 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:26.470672 sshd[3770]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:26.478726 systemd[1]: Started session-22.scope. May 17 00:45:26.480404 systemd-logind[1310]: New session 22 of user core. May 17 00:45:26.758523 sshd[3770]: pam_unix(sshd:session): session closed for user core May 17 00:45:26.763631 systemd[1]: sshd@22-10.128.0.10:22-139.178.89.65:55620.service: Deactivated successfully. May 17 00:45:26.765719 systemd[1]: session-22.scope: Deactivated successfully. May 17 00:45:26.765940 systemd-logind[1310]: Session 22 logged out. Waiting for processes to exit. May 17 00:45:26.771834 systemd-logind[1310]: Removed session 22. May 17 00:45:31.803801 systemd[1]: Started sshd@23-10.128.0.10:22-139.178.89.65:53732.service. May 17 00:45:32.100419 sshd[3783]: Accepted publickey for core from 139.178.89.65 port 53732 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:32.102823 sshd[3783]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:32.111087 systemd[1]: Started session-23.scope. May 17 00:45:32.112710 systemd-logind[1310]: New session 23 of user core. May 17 00:45:32.404469 sshd[3783]: pam_unix(sshd:session): session closed for user core May 17 00:45:32.409930 systemd[1]: sshd@23-10.128.0.10:22-139.178.89.65:53732.service: Deactivated successfully. May 17 00:45:32.411881 systemd-logind[1310]: Session 23 logged out. Waiting for processes to exit. May 17 00:45:32.412041 systemd[1]: session-23.scope: Deactivated successfully. May 17 00:45:32.415670 systemd-logind[1310]: Removed session 23. May 17 00:45:32.448902 systemd[1]: Started sshd@24-10.128.0.10:22-139.178.89.65:53742.service. May 17 00:45:32.741008 sshd[3796]: Accepted publickey for core from 139.178.89.65 port 53742 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:32.743677 sshd[3796]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:32.751484 systemd-logind[1310]: New session 24 of user core. May 17 00:45:32.752010 systemd[1]: Started session-24.scope. May 17 00:45:34.906952 env[1327]: time="2025-05-17T00:45:34.906887025Z" level=info msg="StopContainer for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" with timeout 30 (s)" May 17 00:45:34.910366 env[1327]: time="2025-05-17T00:45:34.910290542Z" level=info msg="Stop container \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" with signal terminated" May 17 00:45:34.951615 env[1327]: time="2025-05-17T00:45:34.951525431Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" May 17 00:45:34.960729 env[1327]: time="2025-05-17T00:45:34.960647747Z" level=info msg="StopContainer for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" with timeout 2 (s)" May 17 00:45:34.961686 env[1327]: time="2025-05-17T00:45:34.961645223Z" level=info msg="Stop container \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" with signal terminated" May 17 00:45:34.980465 systemd-networkd[1079]: lxc_health: Link DOWN May 17 00:45:34.980484 systemd-networkd[1079]: lxc_health: Lost carrier May 17 00:45:35.017091 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686-rootfs.mount: Deactivated successfully. May 17 00:45:35.050581 env[1327]: time="2025-05-17T00:45:35.050514917Z" level=info msg="shim disconnected" id=48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686 May 17 00:45:35.051042 env[1327]: time="2025-05-17T00:45:35.050992579Z" level=warning msg="cleaning up after shim disconnected" id=48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686 namespace=k8s.io May 17 00:45:35.051261 env[1327]: time="2025-05-17T00:45:35.051209890Z" level=info msg="cleaning up dead shim" May 17 00:45:35.059560 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067-rootfs.mount: Deactivated successfully. May 17 00:45:35.072874 env[1327]: time="2025-05-17T00:45:35.072725866Z" level=info msg="shim disconnected" id=c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067 May 17 00:45:35.072874 env[1327]: time="2025-05-17T00:45:35.072830754Z" level=warning msg="cleaning up after shim disconnected" id=c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067 namespace=k8s.io May 17 00:45:35.072874 env[1327]: time="2025-05-17T00:45:35.072849044Z" level=info msg="cleaning up dead shim" May 17 00:45:35.082501 env[1327]: time="2025-05-17T00:45:35.082431113Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:35Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3864 runtime=io.containerd.runc.v2\n" May 17 00:45:35.085665 env[1327]: time="2025-05-17T00:45:35.085507906Z" level=info msg="StopContainer for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" returns successfully" May 17 00:45:35.087139 env[1327]: time="2025-05-17T00:45:35.087070235Z" level=info msg="StopPodSandbox for \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\"" May 17 00:45:35.091579 env[1327]: time="2025-05-17T00:45:35.087183316Z" level=info msg="Container to stop \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:35.091743 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1-shm.mount: Deactivated successfully. May 17 00:45:35.100407 env[1327]: time="2025-05-17T00:45:35.100310337Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:35Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3873 runtime=io.containerd.runc.v2\n" May 17 00:45:35.103613 env[1327]: time="2025-05-17T00:45:35.103550063Z" level=info msg="StopContainer for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" returns successfully" May 17 00:45:35.104377 env[1327]: time="2025-05-17T00:45:35.104312950Z" level=info msg="StopPodSandbox for \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\"" May 17 00:45:35.104520 env[1327]: time="2025-05-17T00:45:35.104436996Z" level=info msg="Container to stop \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:35.104520 env[1327]: time="2025-05-17T00:45:35.104470287Z" level=info msg="Container to stop \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:35.104520 env[1327]: time="2025-05-17T00:45:35.104490538Z" level=info msg="Container to stop \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:35.104520 env[1327]: time="2025-05-17T00:45:35.104509509Z" level=info msg="Container to stop \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:35.104866 env[1327]: time="2025-05-17T00:45:35.104535449Z" level=info msg="Container to stop \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:35.109171 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a-shm.mount: Deactivated successfully. May 17 00:45:35.170579 env[1327]: time="2025-05-17T00:45:35.168127163Z" level=info msg="shim disconnected" id=7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1 May 17 00:45:35.170579 env[1327]: time="2025-05-17T00:45:35.168201508Z" level=warning msg="cleaning up after shim disconnected" id=7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1 namespace=k8s.io May 17 00:45:35.170579 env[1327]: time="2025-05-17T00:45:35.168224497Z" level=info msg="cleaning up dead shim" May 17 00:45:35.190268 env[1327]: time="2025-05-17T00:45:35.190192762Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:35Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3931 runtime=io.containerd.runc.v2\n" May 17 00:45:35.190851 env[1327]: time="2025-05-17T00:45:35.190770744Z" level=info msg="TearDown network for sandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" successfully" May 17 00:45:35.191001 env[1327]: time="2025-05-17T00:45:35.190819338Z" level=info msg="StopPodSandbox for \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" returns successfully" May 17 00:45:35.193374 env[1327]: time="2025-05-17T00:45:35.192193872Z" level=info msg="shim disconnected" id=2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a May 17 00:45:35.193374 env[1327]: time="2025-05-17T00:45:35.192265609Z" level=warning msg="cleaning up after shim disconnected" id=2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a namespace=k8s.io May 17 00:45:35.193374 env[1327]: time="2025-05-17T00:45:35.192284772Z" level=info msg="cleaning up dead shim" May 17 00:45:35.221325 env[1327]: time="2025-05-17T00:45:35.221252266Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:35Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3946 runtime=io.containerd.runc.v2\n" May 17 00:45:35.221846 env[1327]: time="2025-05-17T00:45:35.221790278Z" level=info msg="TearDown network for sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" successfully" May 17 00:45:35.221846 env[1327]: time="2025-05-17T00:45:35.221846053Z" level=info msg="StopPodSandbox for \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" returns successfully" May 17 00:45:35.283038 kubelet[2224]: I0517 00:45:35.282952 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/546317f2-a52c-4374-b0d6-e64927e14f57-cilium-config-path\") pod \"546317f2-a52c-4374-b0d6-e64927e14f57\" (UID: \"546317f2-a52c-4374-b0d6-e64927e14f57\") " May 17 00:45:35.283038 kubelet[2224]: I0517 00:45:35.283034 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94kvj\" (UniqueName: \"kubernetes.io/projected/546317f2-a52c-4374-b0d6-e64927e14f57-kube-api-access-94kvj\") pod \"546317f2-a52c-4374-b0d6-e64927e14f57\" (UID: \"546317f2-a52c-4374-b0d6-e64927e14f57\") " May 17 00:45:35.287551 kubelet[2224]: I0517 00:45:35.287485 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/546317f2-a52c-4374-b0d6-e64927e14f57-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "546317f2-a52c-4374-b0d6-e64927e14f57" (UID: "546317f2-a52c-4374-b0d6-e64927e14f57"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 17 00:45:35.289297 kubelet[2224]: I0517 00:45:35.289240 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/546317f2-a52c-4374-b0d6-e64927e14f57-kube-api-access-94kvj" (OuterVolumeSpecName: "kube-api-access-94kvj") pod "546317f2-a52c-4374-b0d6-e64927e14f57" (UID: "546317f2-a52c-4374-b0d6-e64927e14f57"). InnerVolumeSpecName "kube-api-access-94kvj". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:45:35.383382 kubelet[2224]: I0517 00:45:35.383297 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-etc-cni-netd\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.383611 kubelet[2224]: I0517 00:45:35.383397 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cni-path\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.383611 kubelet[2224]: I0517 00:45:35.383428 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-xtables-lock\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.383611 kubelet[2224]: I0517 00:45:35.383455 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-cgroup\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.383611 kubelet[2224]: I0517 00:45:35.383486 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-hostproc\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.383611 kubelet[2224]: I0517 00:45:35.383510 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-kernel\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.383611 kubelet[2224]: I0517 00:45:35.383552 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/38a4631e-531e-44c6-933e-1b9cfdebcdec-clustermesh-secrets\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384131 kubelet[2224]: I0517 00:45:35.383591 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-hubble-tls\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384131 kubelet[2224]: I0517 00:45:35.383624 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88sf4\" (UniqueName: \"kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-kube-api-access-88sf4\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384131 kubelet[2224]: I0517 00:45:35.383656 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-net\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384131 kubelet[2224]: I0517 00:45:35.383683 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-lib-modules\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384131 kubelet[2224]: I0517 00:45:35.383711 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-bpf-maps\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384131 kubelet[2224]: I0517 00:45:35.383746 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-config-path\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384573 kubelet[2224]: I0517 00:45:35.383778 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-run\") pod \"38a4631e-531e-44c6-933e-1b9cfdebcdec\" (UID: \"38a4631e-531e-44c6-933e-1b9cfdebcdec\") " May 17 00:45:35.384573 kubelet[2224]: I0517 00:45:35.383844 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/546317f2-a52c-4374-b0d6-e64927e14f57-cilium-config-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.384573 kubelet[2224]: I0517 00:45:35.383886 2224 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94kvj\" (UniqueName: \"kubernetes.io/projected/546317f2-a52c-4374-b0d6-e64927e14f57-kube-api-access-94kvj\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.384573 kubelet[2224]: I0517 00:45:35.383951 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.384573 kubelet[2224]: I0517 00:45:35.383998 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.384951 kubelet[2224]: I0517 00:45:35.384028 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cni-path" (OuterVolumeSpecName: "cni-path") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.384951 kubelet[2224]: I0517 00:45:35.384052 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.384951 kubelet[2224]: I0517 00:45:35.384080 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.384951 kubelet[2224]: I0517 00:45:35.384104 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-hostproc" (OuterVolumeSpecName: "hostproc") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.384951 kubelet[2224]: I0517 00:45:35.384129 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.385454 kubelet[2224]: I0517 00:45:35.385402 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.385683 kubelet[2224]: I0517 00:45:35.385643 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.385790 kubelet[2224]: I0517 00:45:35.385698 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:35.390088 kubelet[2224]: I0517 00:45:35.390033 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 17 00:45:35.393011 kubelet[2224]: I0517 00:45:35.392963 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:45:35.394582 kubelet[2224]: I0517 00:45:35.394537 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38a4631e-531e-44c6-933e-1b9cfdebcdec-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 17 00:45:35.397070 kubelet[2224]: I0517 00:45:35.397027 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-kube-api-access-88sf4" (OuterVolumeSpecName: "kube-api-access-88sf4") pod "38a4631e-531e-44c6-933e-1b9cfdebcdec" (UID: "38a4631e-531e-44c6-933e-1b9cfdebcdec"). InnerVolumeSpecName "kube-api-access-88sf4". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:45:35.484807 kubelet[2224]: I0517 00:45:35.484631 2224 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cni-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.484807 kubelet[2224]: I0517 00:45:35.484698 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-cgroup\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.484807 kubelet[2224]: I0517 00:45:35.484720 2224 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-hostproc\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.485518 kubelet[2224]: I0517 00:45:35.485482 2224 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-kernel\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.485704 kubelet[2224]: I0517 00:45:35.485682 2224 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-xtables-lock\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.485854 kubelet[2224]: I0517 00:45:35.485834 2224 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/38a4631e-531e-44c6-933e-1b9cfdebcdec-clustermesh-secrets\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.485973 kubelet[2224]: I0517 00:45:35.485953 2224 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-hubble-tls\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.486101 kubelet[2224]: I0517 00:45:35.486078 2224 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88sf4\" (UniqueName: \"kubernetes.io/projected/38a4631e-531e-44c6-933e-1b9cfdebcdec-kube-api-access-88sf4\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.486285 kubelet[2224]: I0517 00:45:35.486253 2224 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-host-proc-sys-net\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.486502 kubelet[2224]: I0517 00:45:35.486461 2224 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-lib-modules\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.486694 kubelet[2224]: I0517 00:45:35.486667 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-config-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.486828 kubelet[2224]: I0517 00:45:35.486807 2224 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-bpf-maps\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.487000 kubelet[2224]: I0517 00:45:35.486975 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-cilium-run\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.487133 kubelet[2224]: I0517 00:45:35.487111 2224 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/38a4631e-531e-44c6-933e-1b9cfdebcdec-etc-cni-netd\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:35.919261 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a-rootfs.mount: Deactivated successfully. May 17 00:45:35.920051 systemd[1]: var-lib-kubelet-pods-38a4631e\x2d531e\x2d44c6\x2d933e\x2d1b9cfdebcdec-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d88sf4.mount: Deactivated successfully. May 17 00:45:35.920311 systemd[1]: var-lib-kubelet-pods-38a4631e\x2d531e\x2d44c6\x2d933e\x2d1b9cfdebcdec-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 17 00:45:35.920557 systemd[1]: var-lib-kubelet-pods-38a4631e\x2d531e\x2d44c6\x2d933e\x2d1b9cfdebcdec-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 17 00:45:35.920812 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1-rootfs.mount: Deactivated successfully. May 17 00:45:35.921010 systemd[1]: var-lib-kubelet-pods-546317f2\x2da52c\x2d4374\x2db0d6\x2de64927e14f57-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d94kvj.mount: Deactivated successfully. May 17 00:45:36.071798 kubelet[2224]: I0517 00:45:36.071712 2224 scope.go:117] "RemoveContainer" containerID="48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686" May 17 00:45:36.079334 env[1327]: time="2025-05-17T00:45:36.078848997Z" level=info msg="RemoveContainer for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\"" May 17 00:45:36.087316 env[1327]: time="2025-05-17T00:45:36.087255533Z" level=info msg="RemoveContainer for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" returns successfully" May 17 00:45:36.097339 kubelet[2224]: I0517 00:45:36.096096 2224 scope.go:117] "RemoveContainer" containerID="48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686" May 17 00:45:36.098675 env[1327]: time="2025-05-17T00:45:36.098529243Z" level=error msg="ContainerStatus for \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\": not found" May 17 00:45:36.099031 kubelet[2224]: E0517 00:45:36.098994 2224 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\": not found" containerID="48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686" May 17 00:45:36.099417 kubelet[2224]: I0517 00:45:36.099233 2224 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686"} err="failed to get container status \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\": rpc error: code = NotFound desc = an error occurred when try to find container \"48684558524ce5541625e3b4862a4c8d37ad9c86fa5a4c4ded23f4434dccb686\": not found" May 17 00:45:36.099596 kubelet[2224]: I0517 00:45:36.099573 2224 scope.go:117] "RemoveContainer" containerID="c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067" May 17 00:45:36.106545 env[1327]: time="2025-05-17T00:45:36.106485348Z" level=info msg="RemoveContainer for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\"" May 17 00:45:36.119242 env[1327]: time="2025-05-17T00:45:36.119170204Z" level=info msg="RemoveContainer for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" returns successfully" May 17 00:45:36.119687 kubelet[2224]: I0517 00:45:36.119652 2224 scope.go:117] "RemoveContainer" containerID="cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2" May 17 00:45:36.122337 env[1327]: time="2025-05-17T00:45:36.122285805Z" level=info msg="RemoveContainer for \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\"" May 17 00:45:36.128046 env[1327]: time="2025-05-17T00:45:36.127964686Z" level=info msg="RemoveContainer for \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\" returns successfully" May 17 00:45:36.128588 kubelet[2224]: I0517 00:45:36.128530 2224 scope.go:117] "RemoveContainer" containerID="c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1" May 17 00:45:36.131402 env[1327]: time="2025-05-17T00:45:36.130753085Z" level=info msg="RemoveContainer for \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\"" May 17 00:45:36.136991 env[1327]: time="2025-05-17T00:45:36.136914790Z" level=info msg="RemoveContainer for \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\" returns successfully" May 17 00:45:36.137301 kubelet[2224]: I0517 00:45:36.137265 2224 scope.go:117] "RemoveContainer" containerID="9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91" May 17 00:45:36.140415 env[1327]: time="2025-05-17T00:45:36.140302933Z" level=info msg="RemoveContainer for \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\"" May 17 00:45:36.146558 env[1327]: time="2025-05-17T00:45:36.146467458Z" level=info msg="RemoveContainer for \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\" returns successfully" May 17 00:45:36.147070 kubelet[2224]: I0517 00:45:36.147000 2224 scope.go:117] "RemoveContainer" containerID="d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046" May 17 00:45:36.149285 env[1327]: time="2025-05-17T00:45:36.149222793Z" level=info msg="RemoveContainer for \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\"" May 17 00:45:36.163122 env[1327]: time="2025-05-17T00:45:36.163030799Z" level=info msg="RemoveContainer for \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\" returns successfully" May 17 00:45:36.163597 kubelet[2224]: I0517 00:45:36.163541 2224 scope.go:117] "RemoveContainer" containerID="c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067" May 17 00:45:36.164157 env[1327]: time="2025-05-17T00:45:36.164035278Z" level=error msg="ContainerStatus for \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\": not found" May 17 00:45:36.164467 kubelet[2224]: E0517 00:45:36.164430 2224 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\": not found" containerID="c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067" May 17 00:45:36.164603 kubelet[2224]: I0517 00:45:36.164489 2224 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067"} err="failed to get container status \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\": rpc error: code = NotFound desc = an error occurred when try to find container \"c63dd18b01c26f5923f82fe28300d1681a82770131ee715c127f56ab1a6d5067\": not found" May 17 00:45:36.164603 kubelet[2224]: I0517 00:45:36.164530 2224 scope.go:117] "RemoveContainer" containerID="cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2" May 17 00:45:36.165128 env[1327]: time="2025-05-17T00:45:36.165019374Z" level=error msg="ContainerStatus for \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\": not found" May 17 00:45:36.165421 kubelet[2224]: E0517 00:45:36.165386 2224 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\": not found" containerID="cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2" May 17 00:45:36.165530 kubelet[2224]: I0517 00:45:36.165433 2224 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2"} err="failed to get container status \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\": rpc error: code = NotFound desc = an error occurred when try to find container \"cc4f0ff8995d7063b1a333d279d060f41c08222bc72eb2ac0c75c036267b00d2\": not found" May 17 00:45:36.165530 kubelet[2224]: I0517 00:45:36.165463 2224 scope.go:117] "RemoveContainer" containerID="c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1" May 17 00:45:36.166144 env[1327]: time="2025-05-17T00:45:36.166007145Z" level=error msg="ContainerStatus for \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\": not found" May 17 00:45:36.166318 kubelet[2224]: E0517 00:45:36.166280 2224 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\": not found" containerID="c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1" May 17 00:45:36.166447 kubelet[2224]: I0517 00:45:36.166330 2224 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1"} err="failed to get container status \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\": rpc error: code = NotFound desc = an error occurred when try to find container \"c5885dcd60b36d59b9a220693d5c08a323933f35f64612a9cd276bc2a0c94ad1\": not found" May 17 00:45:36.166447 kubelet[2224]: I0517 00:45:36.166395 2224 scope.go:117] "RemoveContainer" containerID="9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91" May 17 00:45:36.166916 env[1327]: time="2025-05-17T00:45:36.166822919Z" level=error msg="ContainerStatus for \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\": not found" May 17 00:45:36.167068 kubelet[2224]: E0517 00:45:36.167021 2224 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\": not found" containerID="9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91" May 17 00:45:36.167172 kubelet[2224]: I0517 00:45:36.167080 2224 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91"} err="failed to get container status \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\": rpc error: code = NotFound desc = an error occurred when try to find container \"9084ebd2f817e0d38273427c530d829988e6a0a5fcdc590021c42994f58c7d91\": not found" May 17 00:45:36.167172 kubelet[2224]: I0517 00:45:36.167108 2224 scope.go:117] "RemoveContainer" containerID="d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046" May 17 00:45:36.167702 env[1327]: time="2025-05-17T00:45:36.167583660Z" level=error msg="ContainerStatus for \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\": not found" May 17 00:45:36.167991 kubelet[2224]: E0517 00:45:36.167954 2224 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\": not found" containerID="d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046" May 17 00:45:36.168111 kubelet[2224]: I0517 00:45:36.168000 2224 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046"} err="failed to get container status \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\": rpc error: code = NotFound desc = an error occurred when try to find container \"d90f44abd4297eaae968d15fd03c734748493d2ca17cfc6905ff92f9166e0046\": not found" May 17 00:45:36.533693 env[1327]: time="2025-05-17T00:45:36.533608523Z" level=info msg="StopPodSandbox for \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\"" May 17 00:45:36.533948 env[1327]: time="2025-05-17T00:45:36.533778937Z" level=info msg="TearDown network for sandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" successfully" May 17 00:45:36.533948 env[1327]: time="2025-05-17T00:45:36.533834247Z" level=info msg="StopPodSandbox for \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" returns successfully" May 17 00:45:36.534506 env[1327]: time="2025-05-17T00:45:36.534443595Z" level=info msg="RemovePodSandbox for \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\"" May 17 00:45:36.534636 env[1327]: time="2025-05-17T00:45:36.534497975Z" level=info msg="Forcibly stopping sandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\"" May 17 00:45:36.534636 env[1327]: time="2025-05-17T00:45:36.534621587Z" level=info msg="TearDown network for sandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" successfully" May 17 00:45:36.540726 env[1327]: time="2025-05-17T00:45:36.540635670Z" level=info msg="RemovePodSandbox \"7680af32fcedeef7e04d44f2025abddf99f372ff57ee9beb354acca12b2e2cd1\" returns successfully" May 17 00:45:36.541736 env[1327]: time="2025-05-17T00:45:36.541668799Z" level=info msg="StopPodSandbox for \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\"" May 17 00:45:36.541910 env[1327]: time="2025-05-17T00:45:36.541812615Z" level=info msg="TearDown network for sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" successfully" May 17 00:45:36.541910 env[1327]: time="2025-05-17T00:45:36.541870095Z" level=info msg="StopPodSandbox for \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" returns successfully" May 17 00:45:36.542653 env[1327]: time="2025-05-17T00:45:36.542594744Z" level=info msg="RemovePodSandbox for \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\"" May 17 00:45:36.542799 env[1327]: time="2025-05-17T00:45:36.542657520Z" level=info msg="Forcibly stopping sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\"" May 17 00:45:36.542904 env[1327]: time="2025-05-17T00:45:36.542854677Z" level=info msg="TearDown network for sandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" successfully" May 17 00:45:36.548769 env[1327]: time="2025-05-17T00:45:36.548704112Z" level=info msg="RemovePodSandbox \"2a0ca7ff4fa71586403d6dcad4ce7b31468d32c1fc4706400d18fe619b87836a\" returns successfully" May 17 00:45:36.611095 kubelet[2224]: I0517 00:45:36.609395 2224 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" path="/var/lib/kubelet/pods/38a4631e-531e-44c6-933e-1b9cfdebcdec/volumes" May 17 00:45:36.611095 kubelet[2224]: I0517 00:45:36.610605 2224 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="546317f2-a52c-4374-b0d6-e64927e14f57" path="/var/lib/kubelet/pods/546317f2-a52c-4374-b0d6-e64927e14f57/volumes" May 17 00:45:36.773155 kubelet[2224]: E0517 00:45:36.773078 2224 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 17 00:45:36.866043 sshd[3796]: pam_unix(sshd:session): session closed for user core May 17 00:45:36.872544 systemd[1]: sshd@24-10.128.0.10:22-139.178.89.65:53742.service: Deactivated successfully. May 17 00:45:36.874758 systemd[1]: session-24.scope: Deactivated successfully. May 17 00:45:36.876152 systemd-logind[1310]: Session 24 logged out. Waiting for processes to exit. May 17 00:45:36.879909 systemd-logind[1310]: Removed session 24. May 17 00:45:36.911411 systemd[1]: Started sshd@25-10.128.0.10:22-139.178.89.65:55306.service. May 17 00:45:37.212892 sshd[3967]: Accepted publickey for core from 139.178.89.65 port 55306 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:37.214644 sshd[3967]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:37.223138 systemd[1]: Started session-25.scope. May 17 00:45:37.224049 systemd-logind[1310]: New session 25 of user core. May 17 00:45:38.304872 kubelet[2224]: E0517 00:45:38.304810 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="546317f2-a52c-4374-b0d6-e64927e14f57" containerName="cilium-operator" May 17 00:45:38.304872 kubelet[2224]: E0517 00:45:38.304871 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" containerName="mount-cgroup" May 17 00:45:38.304872 kubelet[2224]: E0517 00:45:38.304886 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" containerName="apply-sysctl-overwrites" May 17 00:45:38.305701 kubelet[2224]: E0517 00:45:38.304901 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" containerName="clean-cilium-state" May 17 00:45:38.305701 kubelet[2224]: E0517 00:45:38.304915 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" containerName="cilium-agent" May 17 00:45:38.305701 kubelet[2224]: E0517 00:45:38.304930 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" containerName="mount-bpf-fs" May 17 00:45:38.305701 kubelet[2224]: I0517 00:45:38.304978 2224 memory_manager.go:354] "RemoveStaleState removing state" podUID="546317f2-a52c-4374-b0d6-e64927e14f57" containerName="cilium-operator" May 17 00:45:38.305701 kubelet[2224]: I0517 00:45:38.304992 2224 memory_manager.go:354] "RemoveStaleState removing state" podUID="38a4631e-531e-44c6-933e-1b9cfdebcdec" containerName="cilium-agent" May 17 00:45:38.321672 sshd[3967]: pam_unix(sshd:session): session closed for user core May 17 00:45:38.327602 systemd[1]: sshd@25-10.128.0.10:22-139.178.89.65:55306.service: Deactivated successfully. May 17 00:45:38.330029 systemd[1]: session-25.scope: Deactivated successfully. May 17 00:45:38.330093 systemd-logind[1310]: Session 25 logged out. Waiting for processes to exit. May 17 00:45:38.332328 systemd-logind[1310]: Removed session 25. May 17 00:45:38.366064 systemd[1]: Started sshd@26-10.128.0.10:22-139.178.89.65:55318.service. May 17 00:45:38.418961 kubelet[2224]: I0517 00:45:38.418886 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/17662258-b964-432b-8d14-4f718fd298b2-cilium-config-path\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419199 kubelet[2224]: I0517 00:45:38.418973 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-hubble-tls\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419199 kubelet[2224]: I0517 00:45:38.419010 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-bpf-maps\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419199 kubelet[2224]: I0517 00:45:38.419036 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-lib-modules\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419199 kubelet[2224]: I0517 00:45:38.419073 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-xtables-lock\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419199 kubelet[2224]: I0517 00:45:38.419098 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-cilium-ipsec-secrets\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419199 kubelet[2224]: I0517 00:45:38.419128 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-kernel\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419669 kubelet[2224]: I0517 00:45:38.419154 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-run\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419669 kubelet[2224]: I0517 00:45:38.419179 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-hostproc\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419669 kubelet[2224]: I0517 00:45:38.419204 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-etc-cni-netd\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419669 kubelet[2224]: I0517 00:45:38.419237 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-cgroup\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419669 kubelet[2224]: I0517 00:45:38.419269 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cni-path\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419669 kubelet[2224]: I0517 00:45:38.419298 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-clustermesh-secrets\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419988 kubelet[2224]: I0517 00:45:38.419329 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-net\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.419988 kubelet[2224]: I0517 00:45:38.419402 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzjrv\" (UniqueName: \"kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-kube-api-access-kzjrv\") pod \"cilium-9xhsv\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " pod="kube-system/cilium-9xhsv" May 17 00:45:38.618509 env[1327]: time="2025-05-17T00:45:38.618319805Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-9xhsv,Uid:17662258-b964-432b-8d14-4f718fd298b2,Namespace:kube-system,Attempt:0,}" May 17 00:45:38.653397 env[1327]: time="2025-05-17T00:45:38.653255703Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:45:38.653397 env[1327]: time="2025-05-17T00:45:38.653312995Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:45:38.653814 env[1327]: time="2025-05-17T00:45:38.653335685Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:45:38.653814 env[1327]: time="2025-05-17T00:45:38.653629412Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2 pid=3993 runtime=io.containerd.runc.v2 May 17 00:45:38.699527 sshd[3978]: Accepted publickey for core from 139.178.89.65 port 55318 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:38.700337 sshd[3978]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:38.713676 systemd[1]: Started session-26.scope. May 17 00:45:38.714705 systemd-logind[1310]: New session 26 of user core. May 17 00:45:38.756275 env[1327]: time="2025-05-17T00:45:38.756214301Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-9xhsv,Uid:17662258-b964-432b-8d14-4f718fd298b2,Namespace:kube-system,Attempt:0,} returns sandbox id \"2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2\"" May 17 00:45:38.760362 env[1327]: time="2025-05-17T00:45:38.760294047Z" level=info msg="CreateContainer within sandbox \"2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:45:38.779026 env[1327]: time="2025-05-17T00:45:38.778943212Z" level=info msg="CreateContainer within sandbox \"2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416\"" May 17 00:45:38.780707 env[1327]: time="2025-05-17T00:45:38.780467984Z" level=info msg="StartContainer for \"0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416\"" May 17 00:45:38.873468 env[1327]: time="2025-05-17T00:45:38.873058525Z" level=info msg="StartContainer for \"0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416\" returns successfully" May 17 00:45:38.892429 kubelet[2224]: I0517 00:45:38.891857 2224 setters.go:600] "Node became not ready" node="ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-05-17T00:45:38Z","lastTransitionTime":"2025-05-17T00:45:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} May 17 00:45:38.994313 env[1327]: time="2025-05-17T00:45:38.994241501Z" level=info msg="shim disconnected" id=0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416 May 17 00:45:38.994313 env[1327]: time="2025-05-17T00:45:38.994316494Z" level=warning msg="cleaning up after shim disconnected" id=0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416 namespace=k8s.io May 17 00:45:38.994754 env[1327]: time="2025-05-17T00:45:38.994333725Z" level=info msg="cleaning up dead shim" May 17 00:45:39.012389 env[1327]: time="2025-05-17T00:45:39.012291559Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:39Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4090 runtime=io.containerd.runc.v2\n" May 17 00:45:39.094702 env[1327]: time="2025-05-17T00:45:39.094638336Z" level=info msg="StopPodSandbox for \"2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2\"" May 17 00:45:39.094941 env[1327]: time="2025-05-17T00:45:39.094739640Z" level=info msg="Container to stop \"0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" May 17 00:45:39.100721 sshd[3978]: pam_unix(sshd:session): session closed for user core May 17 00:45:39.110052 systemd[1]: sshd@26-10.128.0.10:22-139.178.89.65:55318.service: Deactivated successfully. May 17 00:45:39.111554 systemd[1]: session-26.scope: Deactivated successfully. May 17 00:45:39.115464 systemd-logind[1310]: Session 26 logged out. Waiting for processes to exit. May 17 00:45:39.124436 systemd-logind[1310]: Removed session 26. May 17 00:45:39.145408 systemd[1]: Started sshd@27-10.128.0.10:22-139.178.89.65:55322.service. May 17 00:45:39.207442 env[1327]: time="2025-05-17T00:45:39.207294178Z" level=info msg="shim disconnected" id=2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2 May 17 00:45:39.208513 env[1327]: time="2025-05-17T00:45:39.208456665Z" level=warning msg="cleaning up after shim disconnected" id=2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2 namespace=k8s.io May 17 00:45:39.208734 env[1327]: time="2025-05-17T00:45:39.208706855Z" level=info msg="cleaning up dead shim" May 17 00:45:39.224687 env[1327]: time="2025-05-17T00:45:39.224557770Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:39Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4128 runtime=io.containerd.runc.v2\n" May 17 00:45:39.225182 env[1327]: time="2025-05-17T00:45:39.225127665Z" level=info msg="TearDown network for sandbox \"2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2\" successfully" May 17 00:45:39.225182 env[1327]: time="2025-05-17T00:45:39.225171002Z" level=info msg="StopPodSandbox for \"2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2\" returns successfully" May 17 00:45:39.331303 kubelet[2224]: I0517 00:45:39.331221 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-hubble-tls\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.331303 kubelet[2224]: I0517 00:45:39.331292 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-xtables-lock\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332157 kubelet[2224]: I0517 00:45:39.331327 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-hostproc\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332157 kubelet[2224]: I0517 00:45:39.331387 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/17662258-b964-432b-8d14-4f718fd298b2-cilium-config-path\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332157 kubelet[2224]: I0517 00:45:39.331425 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-run\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332157 kubelet[2224]: I0517 00:45:39.331464 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-clustermesh-secrets\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332157 kubelet[2224]: I0517 00:45:39.331500 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-cilium-ipsec-secrets\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332157 kubelet[2224]: I0517 00:45:39.331538 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-etc-cni-netd\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332641 kubelet[2224]: I0517 00:45:39.331623 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-kernel\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332641 kubelet[2224]: I0517 00:45:39.331659 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-lib-modules\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332641 kubelet[2224]: I0517 00:45:39.331690 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-bpf-maps\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332641 kubelet[2224]: I0517 00:45:39.331745 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-net\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332641 kubelet[2224]: I0517 00:45:39.331785 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzjrv\" (UniqueName: \"kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-kube-api-access-kzjrv\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.332641 kubelet[2224]: I0517 00:45:39.331813 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-cgroup\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.333339 kubelet[2224]: I0517 00:45:39.331839 2224 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cni-path\") pod \"17662258-b964-432b-8d14-4f718fd298b2\" (UID: \"17662258-b964-432b-8d14-4f718fd298b2\") " May 17 00:45:39.333339 kubelet[2224]: I0517 00:45:39.331936 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cni-path" (OuterVolumeSpecName: "cni-path") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.333853 kubelet[2224]: I0517 00:45:39.333806 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.334031 kubelet[2224]: I0517 00:45:39.333875 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.334031 kubelet[2224]: I0517 00:45:39.333908 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.334031 kubelet[2224]: I0517 00:45:39.333955 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.334495 kubelet[2224]: I0517 00:45:39.333806 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.334743 kubelet[2224]: I0517 00:45:39.334687 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.334960 kubelet[2224]: I0517 00:45:39.334923 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-hostproc" (OuterVolumeSpecName: "hostproc") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.338569 kubelet[2224]: I0517 00:45:39.338527 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.338715 kubelet[2224]: I0517 00:45:39.338672 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:45:39.340511 kubelet[2224]: I0517 00:45:39.340473 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17662258-b964-432b-8d14-4f718fd298b2-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" May 17 00:45:39.341869 kubelet[2224]: I0517 00:45:39.341823 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" May 17 00:45:39.343972 kubelet[2224]: I0517 00:45:39.343920 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-kube-api-access-kzjrv" (OuterVolumeSpecName: "kube-api-access-kzjrv") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "kube-api-access-kzjrv". PluginName "kubernetes.io/projected", VolumeGidValue "" May 17 00:45:39.351665 kubelet[2224]: I0517 00:45:39.349446 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 17 00:45:39.352168 kubelet[2224]: I0517 00:45:39.352111 2224 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "17662258-b964-432b-8d14-4f718fd298b2" (UID: "17662258-b964-432b-8d14-4f718fd298b2"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432187 2224 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-kernel\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432256 2224 reconciler_common.go:293] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-etc-cni-netd\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432279 2224 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-lib-modules\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432302 2224 reconciler_common.go:293] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-bpf-maps\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432323 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-cgroup\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432374 2224 reconciler_common.go:293] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cni-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.432609 kubelet[2224]: I0517 00:45:39.432398 2224 reconciler_common.go:293] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-host-proc-sys-net\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432421 2224 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzjrv\" (UniqueName: \"kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-kube-api-access-kzjrv\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432442 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/17662258-b964-432b-8d14-4f718fd298b2-cilium-config-path\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432459 2224 reconciler_common.go:293] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/17662258-b964-432b-8d14-4f718fd298b2-hubble-tls\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432475 2224 reconciler_common.go:293] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-xtables-lock\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432491 2224 reconciler_common.go:293] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-hostproc\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432511 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-cilium-ipsec-secrets\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433177 kubelet[2224]: I0517 00:45:39.432528 2224 reconciler_common.go:293] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/17662258-b964-432b-8d14-4f718fd298b2-cilium-run\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.433616 kubelet[2224]: I0517 00:45:39.432548 2224 reconciler_common.go:293] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/17662258-b964-432b-8d14-4f718fd298b2-clustermesh-secrets\") on node \"ci-3510-3-7-nightly-20250516-2100-0e0eb87c6ab29d9711f9\" DevicePath \"\"" May 17 00:45:39.467006 sshd[4113]: Accepted publickey for core from 139.178.89.65 port 55322 ssh2: RSA SHA256:jyE3lnafiBGDGJK6dHnApyF/jgfCnjVgkPORJQqM9Ps May 17 00:45:39.469333 sshd[4113]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) May 17 00:45:39.477742 systemd[1]: Started session-27.scope. May 17 00:45:39.478492 systemd-logind[1310]: New session 27 of user core. May 17 00:45:39.534139 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2627a1503e2f8d0332a62a5d80c89974fe919b1f1c807264de2759a6d50aa3d2-shm.mount: Deactivated successfully. May 17 00:45:39.534593 systemd[1]: var-lib-kubelet-pods-17662258\x2db964\x2d432b\x2d8d14\x2d4f718fd298b2-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkzjrv.mount: Deactivated successfully. May 17 00:45:39.535019 systemd[1]: var-lib-kubelet-pods-17662258\x2db964\x2d432b\x2d8d14\x2d4f718fd298b2-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. May 17 00:45:39.535249 systemd[1]: var-lib-kubelet-pods-17662258\x2db964\x2d432b\x2d8d14\x2d4f718fd298b2-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. May 17 00:45:39.535497 systemd[1]: var-lib-kubelet-pods-17662258\x2db964\x2d432b\x2d8d14\x2d4f718fd298b2-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. May 17 00:45:40.098079 kubelet[2224]: I0517 00:45:40.098023 2224 scope.go:117] "RemoveContainer" containerID="0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416" May 17 00:45:40.102862 env[1327]: time="2025-05-17T00:45:40.102457487Z" level=info msg="RemoveContainer for \"0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416\"" May 17 00:45:40.113649 env[1327]: time="2025-05-17T00:45:40.113584990Z" level=info msg="RemoveContainer for \"0a9952654bfa356d7d8bcc85ede432b63008c81682d8b8ba4c70f08d9becb416\" returns successfully" May 17 00:45:40.162325 kubelet[2224]: E0517 00:45:40.162269 2224 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="17662258-b964-432b-8d14-4f718fd298b2" containerName="mount-cgroup" May 17 00:45:40.162575 kubelet[2224]: I0517 00:45:40.162367 2224 memory_manager.go:354] "RemoveStaleState removing state" podUID="17662258-b964-432b-8d14-4f718fd298b2" containerName="mount-cgroup" May 17 00:45:40.340910 kubelet[2224]: I0517 00:45:40.340807 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-lib-modules\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341636 kubelet[2224]: I0517 00:45:40.340927 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-cilium-run\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341636 kubelet[2224]: I0517 00:45:40.340962 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-cilium-cgroup\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341636 kubelet[2224]: I0517 00:45:40.340996 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-hubble-tls\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341636 kubelet[2224]: I0517 00:45:40.341030 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-hostproc\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341636 kubelet[2224]: I0517 00:45:40.341065 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-host-proc-sys-kernel\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341636 kubelet[2224]: I0517 00:45:40.341101 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgfmz\" (UniqueName: \"kubernetes.io/projected/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-kube-api-access-bgfmz\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341993 kubelet[2224]: I0517 00:45:40.341143 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-clustermesh-secrets\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341993 kubelet[2224]: I0517 00:45:40.341173 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-cni-path\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341993 kubelet[2224]: I0517 00:45:40.341209 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-xtables-lock\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341993 kubelet[2224]: I0517 00:45:40.341264 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-etc-cni-netd\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341993 kubelet[2224]: I0517 00:45:40.341298 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-cilium-config-path\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.341993 kubelet[2224]: I0517 00:45:40.341376 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-cilium-ipsec-secrets\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.342212 kubelet[2224]: I0517 00:45:40.341415 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-bpf-maps\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.342212 kubelet[2224]: I0517 00:45:40.341446 2224 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d3c26f8b-63e2-4570-b8d2-bdae7185b0cc-host-proc-sys-net\") pod \"cilium-vb4r2\" (UID: \"d3c26f8b-63e2-4570-b8d2-bdae7185b0cc\") " pod="kube-system/cilium-vb4r2" May 17 00:45:40.609539 kubelet[2224]: I0517 00:45:40.609171 2224 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17662258-b964-432b-8d14-4f718fd298b2" path="/var/lib/kubelet/pods/17662258-b964-432b-8d14-4f718fd298b2/volumes" May 17 00:45:40.776863 env[1327]: time="2025-05-17T00:45:40.776781117Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vb4r2,Uid:d3c26f8b-63e2-4570-b8d2-bdae7185b0cc,Namespace:kube-system,Attempt:0,}" May 17 00:45:40.821988 env[1327]: time="2025-05-17T00:45:40.821848622Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 May 17 00:45:40.822224 env[1327]: time="2025-05-17T00:45:40.822021918Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 May 17 00:45:40.822224 env[1327]: time="2025-05-17T00:45:40.822086339Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 May 17 00:45:40.822914 env[1327]: time="2025-05-17T00:45:40.822811331Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91 pid=4162 runtime=io.containerd.runc.v2 May 17 00:45:40.871099 systemd[1]: run-containerd-runc-k8s.io-ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91-runc.dQR6eI.mount: Deactivated successfully. May 17 00:45:40.918974 env[1327]: time="2025-05-17T00:45:40.918909667Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vb4r2,Uid:d3c26f8b-63e2-4570-b8d2-bdae7185b0cc,Namespace:kube-system,Attempt:0,} returns sandbox id \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\"" May 17 00:45:40.923813 env[1327]: time="2025-05-17T00:45:40.923703053Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" May 17 00:45:40.948812 env[1327]: time="2025-05-17T00:45:40.948723097Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"e00945c8c058e79ed42e2e03cb34f93faade77f386de6a934bc1b5d66d2ea431\"" May 17 00:45:40.950489 env[1327]: time="2025-05-17T00:45:40.949648324Z" level=info msg="StartContainer for \"e00945c8c058e79ed42e2e03cb34f93faade77f386de6a934bc1b5d66d2ea431\"" May 17 00:45:41.035180 env[1327]: time="2025-05-17T00:45:41.035102430Z" level=info msg="StartContainer for \"e00945c8c058e79ed42e2e03cb34f93faade77f386de6a934bc1b5d66d2ea431\" returns successfully" May 17 00:45:41.093719 env[1327]: time="2025-05-17T00:45:41.093639560Z" level=info msg="shim disconnected" id=e00945c8c058e79ed42e2e03cb34f93faade77f386de6a934bc1b5d66d2ea431 May 17 00:45:41.093719 env[1327]: time="2025-05-17T00:45:41.093712082Z" level=warning msg="cleaning up after shim disconnected" id=e00945c8c058e79ed42e2e03cb34f93faade77f386de6a934bc1b5d66d2ea431 namespace=k8s.io May 17 00:45:41.093719 env[1327]: time="2025-05-17T00:45:41.093731767Z" level=info msg="cleaning up dead shim" May 17 00:45:41.121735 env[1327]: time="2025-05-17T00:45:41.121651756Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:41Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4245 runtime=io.containerd.runc.v2\n" May 17 00:45:41.774856 kubelet[2224]: E0517 00:45:41.774790 2224 kubelet.go:2902] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" May 17 00:45:41.796544 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2555050003.mount: Deactivated successfully. May 17 00:45:42.114257 env[1327]: time="2025-05-17T00:45:42.114100490Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" May 17 00:45:42.139742 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1953093164.mount: Deactivated successfully. May 17 00:45:42.157143 env[1327]: time="2025-05-17T00:45:42.157080115Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"07b975e22fda9e46a7f0426e722dc5438215f1ed039e60f0ea3f7355be1c22e1\"" May 17 00:45:42.162411 env[1327]: time="2025-05-17T00:45:42.159970839Z" level=info msg="StartContainer for \"07b975e22fda9e46a7f0426e722dc5438215f1ed039e60f0ea3f7355be1c22e1\"" May 17 00:45:42.280486 env[1327]: time="2025-05-17T00:45:42.280415483Z" level=info msg="StartContainer for \"07b975e22fda9e46a7f0426e722dc5438215f1ed039e60f0ea3f7355be1c22e1\" returns successfully" May 17 00:45:42.317569 env[1327]: time="2025-05-17T00:45:42.316475201Z" level=info msg="shim disconnected" id=07b975e22fda9e46a7f0426e722dc5438215f1ed039e60f0ea3f7355be1c22e1 May 17 00:45:42.317569 env[1327]: time="2025-05-17T00:45:42.316639443Z" level=warning msg="cleaning up after shim disconnected" id=07b975e22fda9e46a7f0426e722dc5438215f1ed039e60f0ea3f7355be1c22e1 namespace=k8s.io May 17 00:45:42.317569 env[1327]: time="2025-05-17T00:45:42.316660221Z" level=info msg="cleaning up dead shim" May 17 00:45:42.331201 env[1327]: time="2025-05-17T00:45:42.331138727Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:42Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4311 runtime=io.containerd.runc.v2\n" May 17 00:45:42.796558 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-07b975e22fda9e46a7f0426e722dc5438215f1ed039e60f0ea3f7355be1c22e1-rootfs.mount: Deactivated successfully. May 17 00:45:43.131489 env[1327]: time="2025-05-17T00:45:43.122665204Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" May 17 00:45:43.155953 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1080386098.mount: Deactivated successfully. May 17 00:45:43.164028 env[1327]: time="2025-05-17T00:45:43.163935605Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb\"" May 17 00:45:43.167464 env[1327]: time="2025-05-17T00:45:43.167408460Z" level=info msg="StartContainer for \"acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb\"" May 17 00:45:43.289721 env[1327]: time="2025-05-17T00:45:43.289651285Z" level=info msg="StartContainer for \"acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb\" returns successfully" May 17 00:45:43.328615 env[1327]: time="2025-05-17T00:45:43.328524730Z" level=info msg="shim disconnected" id=acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb May 17 00:45:43.328615 env[1327]: time="2025-05-17T00:45:43.328609073Z" level=warning msg="cleaning up after shim disconnected" id=acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb namespace=k8s.io May 17 00:45:43.329159 env[1327]: time="2025-05-17T00:45:43.328629296Z" level=info msg="cleaning up dead shim" May 17 00:45:43.342557 env[1327]: time="2025-05-17T00:45:43.342480733Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:43Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4371 runtime=io.containerd.runc.v2\n" May 17 00:45:43.796955 systemd[1]: run-containerd-runc-k8s.io-acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb-runc.pdeUwN.mount: Deactivated successfully. May 17 00:45:43.797266 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-acb09a37f40af4457d3398fcd52f817386c484b9a3c798633dca1b685d75e0fb-rootfs.mount: Deactivated successfully. May 17 00:45:44.124814 env[1327]: time="2025-05-17T00:45:44.124381633Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" May 17 00:45:44.153937 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1808650632.mount: Deactivated successfully. May 17 00:45:44.168397 env[1327]: time="2025-05-17T00:45:44.168292138Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"b890c9755343725f03de395a3236e0ae884a8bd299fd2062de2292f7bace7a68\"" May 17 00:45:44.170108 env[1327]: time="2025-05-17T00:45:44.170057658Z" level=info msg="StartContainer for \"b890c9755343725f03de395a3236e0ae884a8bd299fd2062de2292f7bace7a68\"" May 17 00:45:44.295726 env[1327]: time="2025-05-17T00:45:44.295659065Z" level=info msg="StartContainer for \"b890c9755343725f03de395a3236e0ae884a8bd299fd2062de2292f7bace7a68\" returns successfully" May 17 00:45:44.327722 env[1327]: time="2025-05-17T00:45:44.327637806Z" level=info msg="shim disconnected" id=b890c9755343725f03de395a3236e0ae884a8bd299fd2062de2292f7bace7a68 May 17 00:45:44.327722 env[1327]: time="2025-05-17T00:45:44.327708278Z" level=warning msg="cleaning up after shim disconnected" id=b890c9755343725f03de395a3236e0ae884a8bd299fd2062de2292f7bace7a68 namespace=k8s.io May 17 00:45:44.327722 env[1327]: time="2025-05-17T00:45:44.327729349Z" level=info msg="cleaning up dead shim" May 17 00:45:44.341933 env[1327]: time="2025-05-17T00:45:44.341857273Z" level=warning msg="cleanup warnings time=\"2025-05-17T00:45:44Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4426 runtime=io.containerd.runc.v2\n" May 17 00:45:44.796766 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b890c9755343725f03de395a3236e0ae884a8bd299fd2062de2292f7bace7a68-rootfs.mount: Deactivated successfully. May 17 00:45:45.131231 env[1327]: time="2025-05-17T00:45:45.130865942Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" May 17 00:45:45.172133 env[1327]: time="2025-05-17T00:45:45.169634962Z" level=info msg="CreateContainer within sandbox \"ea6ad70618cc4a536a67658594214fe0c8a4e41c032aa9aa3844c7d281299f91\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"7cf74c89b48238f1bffd7fae5448a4efad926d833916989f69fcb78f39815cc4\"" May 17 00:45:45.171034 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1280963322.mount: Deactivated successfully. May 17 00:45:45.175687 env[1327]: time="2025-05-17T00:45:45.175617679Z" level=info msg="StartContainer for \"7cf74c89b48238f1bffd7fae5448a4efad926d833916989f69fcb78f39815cc4\"" May 17 00:45:45.302380 env[1327]: time="2025-05-17T00:45:45.299869973Z" level=info msg="StartContainer for \"7cf74c89b48238f1bffd7fae5448a4efad926d833916989f69fcb78f39815cc4\" returns successfully" May 17 00:45:45.865411 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) May 17 00:45:46.164421 kubelet[2224]: I0517 00:45:46.164177 2224 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-vb4r2" podStartSLOduration=6.164141617 podStartE2EDuration="6.164141617s" podCreationTimestamp="2025-05-17 00:45:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-05-17 00:45:46.163547483 +0000 UTC m=+129.856108154" watchObservedRunningTime="2025-05-17 00:45:46.164141617 +0000 UTC m=+129.856702313" May 17 00:45:49.333730 systemd-networkd[1079]: lxc_health: Link UP May 17 00:45:49.368453 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready May 17 00:45:49.375588 systemd-networkd[1079]: lxc_health: Gained carrier May 17 00:45:50.282998 systemd[1]: run-containerd-runc-k8s.io-7cf74c89b48238f1bffd7fae5448a4efad926d833916989f69fcb78f39815cc4-runc.EOydoU.mount: Deactivated successfully. May 17 00:45:50.796138 systemd-networkd[1079]: lxc_health: Gained IPv6LL May 17 00:45:52.635702 systemd[1]: run-containerd-runc-k8s.io-7cf74c89b48238f1bffd7fae5448a4efad926d833916989f69fcb78f39815cc4-runc.sBcTZZ.mount: Deactivated successfully. May 17 00:45:54.919823 systemd[1]: run-containerd-runc-k8s.io-7cf74c89b48238f1bffd7fae5448a4efad926d833916989f69fcb78f39815cc4-runc.JuDX8W.mount: Deactivated successfully. May 17 00:45:55.073148 sshd[4113]: pam_unix(sshd:session): session closed for user core May 17 00:45:55.080038 systemd-logind[1310]: Session 27 logged out. Waiting for processes to exit. May 17 00:45:55.082190 systemd[1]: sshd@27-10.128.0.10:22-139.178.89.65:55322.service: Deactivated successfully. May 17 00:45:55.083680 systemd[1]: session-27.scope: Deactivated successfully. May 17 00:45:55.086098 systemd-logind[1310]: Removed session 27.