Jul 2 07:48:28.080152 kernel: Linux version 5.15.161-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Mon Jul 1 23:45:21 -00 2024 Jul 2 07:48:28.080202 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=d29251fe942de56b08103b03939b6e5af4108e76dc6080fe2498c5db43f16e82 Jul 2 07:48:28.080222 kernel: BIOS-provided physical RAM map: Jul 2 07:48:28.080237 kernel: BIOS-e820: [mem 0x0000000000000000-0x0000000000000fff] reserved Jul 2 07:48:28.080251 kernel: BIOS-e820: [mem 0x0000000000001000-0x0000000000054fff] usable Jul 2 07:48:28.080266 kernel: BIOS-e820: [mem 0x0000000000055000-0x000000000005ffff] reserved Jul 2 07:48:28.080289 kernel: BIOS-e820: [mem 0x0000000000060000-0x0000000000097fff] usable Jul 2 07:48:28.080305 kernel: BIOS-e820: [mem 0x0000000000098000-0x000000000009ffff] reserved Jul 2 07:48:28.080320 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000bf8ecfff] usable Jul 2 07:48:28.080335 kernel: BIOS-e820: [mem 0x00000000bf8ed000-0x00000000bfb6cfff] reserved Jul 2 07:48:28.080350 kernel: BIOS-e820: [mem 0x00000000bfb6d000-0x00000000bfb7efff] ACPI data Jul 2 07:48:28.080366 kernel: BIOS-e820: [mem 0x00000000bfb7f000-0x00000000bfbfefff] ACPI NVS Jul 2 07:48:28.080381 kernel: BIOS-e820: [mem 0x00000000bfbff000-0x00000000bffdffff] usable Jul 2 07:48:28.080396 kernel: BIOS-e820: [mem 0x00000000bffe0000-0x00000000bfffffff] reserved Jul 2 07:48:28.080420 kernel: BIOS-e820: [mem 0x0000000100000000-0x000000021fffffff] usable Jul 2 07:48:28.080436 kernel: NX (Execute Disable) protection: active Jul 2 07:48:28.080453 kernel: efi: EFI v2.70 by EDK II Jul 2 07:48:28.080469 kernel: efi: TPMFinalLog=0xbfbf7000 ACPI=0xbfb7e000 ACPI 2.0=0xbfb7e014 SMBIOS=0xbf9e8000 RNG=0xbfb73018 TPMEventLog=0xbd2d2018 Jul 2 07:48:28.080486 kernel: random: crng init done Jul 2 07:48:28.080502 kernel: SMBIOS 2.4 present. Jul 2 07:48:28.080519 kernel: DMI: Google Google Compute Engine/Google Compute Engine, BIOS Google 04/02/2024 Jul 2 07:48:28.080535 kernel: Hypervisor detected: KVM Jul 2 07:48:28.080556 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jul 2 07:48:28.080572 kernel: kvm-clock: cpu 0, msr 168192001, primary cpu clock Jul 2 07:48:28.080588 kernel: kvm-clock: using sched offset of 12416543775 cycles Jul 2 07:48:28.080606 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jul 2 07:48:28.080623 kernel: tsc: Detected 2299.998 MHz processor Jul 2 07:48:28.080640 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jul 2 07:48:28.080657 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jul 2 07:48:28.080673 kernel: last_pfn = 0x220000 max_arch_pfn = 0x400000000 Jul 2 07:48:28.080690 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jul 2 07:48:28.080706 kernel: last_pfn = 0xbffe0 max_arch_pfn = 0x400000000 Jul 2 07:48:28.080727 kernel: Using GB pages for direct mapping Jul 2 07:48:28.080744 kernel: Secure boot disabled Jul 2 07:48:28.080760 kernel: ACPI: Early table checksum verification disabled Jul 2 07:48:28.080796 kernel: ACPI: RSDP 0x00000000BFB7E014 000024 (v02 Google) Jul 2 07:48:28.080811 kernel: ACPI: XSDT 0x00000000BFB7D0E8 00005C (v01 Google GOOGFACP 00000001 01000013) Jul 2 07:48:28.080825 kernel: ACPI: FACP 0x00000000BFB78000 0000F4 (v02 Google GOOGFACP 00000001 GOOG 00000001) Jul 2 07:48:28.080840 kernel: ACPI: DSDT 0x00000000BFB79000 001A64 (v01 Google GOOGDSDT 00000001 GOOG 00000001) Jul 2 07:48:28.080854 kernel: ACPI: FACS 0x00000000BFBF2000 000040 Jul 2 07:48:28.080882 kernel: ACPI: SSDT 0x00000000BFB7C000 000316 (v02 GOOGLE Tpm2Tabl 00001000 INTL 20211217) Jul 2 07:48:28.080900 kernel: ACPI: TPM2 0x00000000BFB7B000 000034 (v04 GOOGLE 00000001 GOOG 00000001) Jul 2 07:48:28.080916 kernel: ACPI: SRAT 0x00000000BFB77000 0000C8 (v03 Google GOOGSRAT 00000001 GOOG 00000001) Jul 2 07:48:28.080933 kernel: ACPI: APIC 0x00000000BFB76000 000076 (v05 Google GOOGAPIC 00000001 GOOG 00000001) Jul 2 07:48:28.080950 kernel: ACPI: SSDT 0x00000000BFB75000 000980 (v01 Google GOOGSSDT 00000001 GOOG 00000001) Jul 2 07:48:28.080969 kernel: ACPI: WAET 0x00000000BFB74000 000028 (v01 Google GOOGWAET 00000001 GOOG 00000001) Jul 2 07:48:28.080992 kernel: ACPI: Reserving FACP table memory at [mem 0xbfb78000-0xbfb780f3] Jul 2 07:48:28.081009 kernel: ACPI: Reserving DSDT table memory at [mem 0xbfb79000-0xbfb7aa63] Jul 2 07:48:28.081027 kernel: ACPI: Reserving FACS table memory at [mem 0xbfbf2000-0xbfbf203f] Jul 2 07:48:28.081045 kernel: ACPI: Reserving SSDT table memory at [mem 0xbfb7c000-0xbfb7c315] Jul 2 07:48:28.081064 kernel: ACPI: Reserving TPM2 table memory at [mem 0xbfb7b000-0xbfb7b033] Jul 2 07:48:28.081082 kernel: ACPI: Reserving SRAT table memory at [mem 0xbfb77000-0xbfb770c7] Jul 2 07:48:28.081098 kernel: ACPI: Reserving APIC table memory at [mem 0xbfb76000-0xbfb76075] Jul 2 07:48:28.081116 kernel: ACPI: Reserving SSDT table memory at [mem 0xbfb75000-0xbfb7597f] Jul 2 07:48:28.081135 kernel: ACPI: Reserving WAET table memory at [mem 0xbfb74000-0xbfb74027] Jul 2 07:48:28.081164 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Jul 2 07:48:28.081182 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Jul 2 07:48:28.081196 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Jul 2 07:48:28.081212 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0xbfffffff] Jul 2 07:48:28.081227 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x21fffffff] Jul 2 07:48:28.081245 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0xbfffffff] -> [mem 0x00000000-0xbfffffff] Jul 2 07:48:28.081262 kernel: NUMA: Node 0 [mem 0x00000000-0xbfffffff] + [mem 0x100000000-0x21fffffff] -> [mem 0x00000000-0x21fffffff] Jul 2 07:48:28.081280 kernel: NODE_DATA(0) allocated [mem 0x21fffa000-0x21fffffff] Jul 2 07:48:28.081298 kernel: Zone ranges: Jul 2 07:48:28.081321 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jul 2 07:48:28.081338 kernel: DMA32 [mem 0x0000000001000000-0x00000000ffffffff] Jul 2 07:48:28.081356 kernel: Normal [mem 0x0000000100000000-0x000000021fffffff] Jul 2 07:48:28.081374 kernel: Movable zone start for each node Jul 2 07:48:28.081392 kernel: Early memory node ranges Jul 2 07:48:28.081410 kernel: node 0: [mem 0x0000000000001000-0x0000000000054fff] Jul 2 07:48:28.081428 kernel: node 0: [mem 0x0000000000060000-0x0000000000097fff] Jul 2 07:48:28.081445 kernel: node 0: [mem 0x0000000000100000-0x00000000bf8ecfff] Jul 2 07:48:28.081462 kernel: node 0: [mem 0x00000000bfbff000-0x00000000bffdffff] Jul 2 07:48:28.081483 kernel: node 0: [mem 0x0000000100000000-0x000000021fffffff] Jul 2 07:48:28.081500 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000021fffffff] Jul 2 07:48:28.081518 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jul 2 07:48:28.081536 kernel: On node 0, zone DMA: 11 pages in unavailable ranges Jul 2 07:48:28.081553 kernel: On node 0, zone DMA: 104 pages in unavailable ranges Jul 2 07:48:28.081572 kernel: On node 0, zone DMA32: 786 pages in unavailable ranges Jul 2 07:48:28.081589 kernel: On node 0, zone Normal: 32 pages in unavailable ranges Jul 2 07:48:28.081607 kernel: ACPI: PM-Timer IO Port: 0xb008 Jul 2 07:48:28.081625 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jul 2 07:48:28.081648 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jul 2 07:48:28.081665 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jul 2 07:48:28.081683 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jul 2 07:48:28.081701 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jul 2 07:48:28.081718 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jul 2 07:48:28.081737 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jul 2 07:48:28.081754 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Jul 2 07:48:28.092349 kernel: [mem 0xc0000000-0xffffffff] available for PCI devices Jul 2 07:48:28.092374 kernel: Booting paravirtualized kernel on KVM Jul 2 07:48:28.092400 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jul 2 07:48:28.092418 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Jul 2 07:48:28.092435 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u1048576 Jul 2 07:48:28.092452 kernel: pcpu-alloc: s188696 r8192 d32488 u1048576 alloc=1*2097152 Jul 2 07:48:28.092469 kernel: pcpu-alloc: [0] 0 1 Jul 2 07:48:28.092486 kernel: kvm-guest: PV spinlocks enabled Jul 2 07:48:28.092503 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jul 2 07:48:28.092519 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1932280 Jul 2 07:48:28.092536 kernel: Policy zone: Normal Jul 2 07:48:28.092559 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=d29251fe942de56b08103b03939b6e5af4108e76dc6080fe2498c5db43f16e82 Jul 2 07:48:28.092576 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jul 2 07:48:28.092593 kernel: Dentry cache hash table entries: 1048576 (order: 11, 8388608 bytes, linear) Jul 2 07:48:28.092609 kernel: Inode-cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jul 2 07:48:28.092627 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jul 2 07:48:28.092644 kernel: Memory: 7516812K/7860584K available (12294K kernel code, 2276K rwdata, 13712K rodata, 47444K init, 4144K bss, 343512K reserved, 0K cma-reserved) Jul 2 07:48:28.092661 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jul 2 07:48:28.092678 kernel: Kernel/User page tables isolation: enabled Jul 2 07:48:28.092698 kernel: ftrace: allocating 34514 entries in 135 pages Jul 2 07:48:28.092715 kernel: ftrace: allocated 135 pages with 4 groups Jul 2 07:48:28.092731 kernel: rcu: Hierarchical RCU implementation. Jul 2 07:48:28.092749 kernel: rcu: RCU event tracing is enabled. Jul 2 07:48:28.092783 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jul 2 07:48:28.092800 kernel: Rude variant of Tasks RCU enabled. Jul 2 07:48:28.092817 kernel: Tracing variant of Tasks RCU enabled. Jul 2 07:48:28.092834 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jul 2 07:48:28.092849 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jul 2 07:48:28.092871 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Jul 2 07:48:28.092901 kernel: Console: colour dummy device 80x25 Jul 2 07:48:28.092919 kernel: printk: console [ttyS0] enabled Jul 2 07:48:28.092940 kernel: ACPI: Core revision 20210730 Jul 2 07:48:28.092958 kernel: APIC: Switch to symmetric I/O mode setup Jul 2 07:48:28.092976 kernel: x2apic enabled Jul 2 07:48:28.092993 kernel: Switched APIC routing to physical x2apic. Jul 2 07:48:28.093011 kernel: ..TIMER: vector=0x30 apic1=0 pin1=0 apic2=-1 pin2=-1 Jul 2 07:48:28.093029 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x212733415c7, max_idle_ns: 440795236380 ns Jul 2 07:48:28.093048 kernel: Calibrating delay loop (skipped) preset value.. 4599.99 BogoMIPS (lpj=2299998) Jul 2 07:48:28.093069 kernel: Last level iTLB entries: 4KB 1024, 2MB 1024, 4MB 1024 Jul 2 07:48:28.093087 kernel: Last level dTLB entries: 4KB 1024, 2MB 1024, 4MB 1024, 1GB 4 Jul 2 07:48:28.093105 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jul 2 07:48:28.093123 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on vm exit Jul 2 07:48:28.093140 kernel: Spectre V2 : Spectre BHI mitigation: SW BHB clearing on syscall Jul 2 07:48:28.093158 kernel: Spectre V2 : Mitigation: IBRS Jul 2 07:48:28.093175 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jul 2 07:48:28.093197 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Jul 2 07:48:28.093214 kernel: RETBleed: Mitigation: IBRS Jul 2 07:48:28.093232 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jul 2 07:48:28.093250 kernel: Spectre V2 : User space: Mitigation: STIBP via seccomp and prctl Jul 2 07:48:28.093274 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Jul 2 07:48:28.093292 kernel: MDS: Mitigation: Clear CPU buffers Jul 2 07:48:28.093309 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Jul 2 07:48:28.093328 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jul 2 07:48:28.093349 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jul 2 07:48:28.093367 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jul 2 07:48:28.093385 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jul 2 07:48:28.093403 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Jul 2 07:48:28.093420 kernel: Freeing SMP alternatives memory: 32K Jul 2 07:48:28.093437 kernel: pid_max: default: 32768 minimum: 301 Jul 2 07:48:28.093454 kernel: LSM: Security Framework initializing Jul 2 07:48:28.093472 kernel: SELinux: Initializing. Jul 2 07:48:28.093490 kernel: Mount-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Jul 2 07:48:28.093511 kernel: Mountpoint-cache hash table entries: 16384 (order: 5, 131072 bytes, linear) Jul 2 07:48:28.093529 kernel: smpboot: CPU0: Intel(R) Xeon(R) CPU @ 2.30GHz (family: 0x6, model: 0x3f, stepping: 0x0) Jul 2 07:48:28.093547 kernel: Performance Events: unsupported p6 CPU model 63 no PMU driver, software events only. Jul 2 07:48:28.093564 kernel: signal: max sigframe size: 1776 Jul 2 07:48:28.093582 kernel: rcu: Hierarchical SRCU implementation. Jul 2 07:48:28.093599 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Jul 2 07:48:28.093617 kernel: smp: Bringing up secondary CPUs ... Jul 2 07:48:28.093635 kernel: x86: Booting SMP configuration: Jul 2 07:48:28.093652 kernel: .... node #0, CPUs: #1 Jul 2 07:48:28.093674 kernel: kvm-clock: cpu 1, msr 168192041, secondary cpu clock Jul 2 07:48:28.093693 kernel: MDS CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/mds.html for more details. Jul 2 07:48:28.093712 kernel: MMIO Stale Data CPU bug present and SMT on, data leak possible. See https://www.kernel.org/doc/html/latest/admin-guide/hw-vuln/processor_mmio_stale_data.html for more details. Jul 2 07:48:28.093729 kernel: smp: Brought up 1 node, 2 CPUs Jul 2 07:48:28.093747 kernel: smpboot: Max logical packages: 1 Jul 2 07:48:28.093775 kernel: smpboot: Total of 2 processors activated (9199.99 BogoMIPS) Jul 2 07:48:28.093793 kernel: devtmpfs: initialized Jul 2 07:48:28.093810 kernel: x86/mm: Memory block size: 128MB Jul 2 07:48:28.093828 kernel: ACPI: PM: Registering ACPI NVS region [mem 0xbfb7f000-0xbfbfefff] (524288 bytes) Jul 2 07:48:28.093850 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jul 2 07:48:28.093868 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jul 2 07:48:28.093886 kernel: pinctrl core: initialized pinctrl subsystem Jul 2 07:48:28.093904 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jul 2 07:48:28.093922 kernel: audit: initializing netlink subsys (disabled) Jul 2 07:48:28.093939 kernel: audit: type=2000 audit(1719906507.066:1): state=initialized audit_enabled=0 res=1 Jul 2 07:48:28.093957 kernel: thermal_sys: Registered thermal governor 'step_wise' Jul 2 07:48:28.093975 kernel: thermal_sys: Registered thermal governor 'user_space' Jul 2 07:48:28.093992 kernel: cpuidle: using governor menu Jul 2 07:48:28.094012 kernel: ACPI: bus type PCI registered Jul 2 07:48:28.094029 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jul 2 07:48:28.094046 kernel: dca service started, version 1.12.1 Jul 2 07:48:28.094064 kernel: PCI: Using configuration type 1 for base access Jul 2 07:48:28.094082 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jul 2 07:48:28.094099 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Jul 2 07:48:28.094117 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Jul 2 07:48:28.094135 kernel: ACPI: Added _OSI(Module Device) Jul 2 07:48:28.094152 kernel: ACPI: Added _OSI(Processor Device) Jul 2 07:48:28.094174 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jul 2 07:48:28.094191 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jul 2 07:48:28.094210 kernel: ACPI: Added _OSI(Linux-Dell-Video) Jul 2 07:48:28.094227 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Jul 2 07:48:28.094245 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Jul 2 07:48:28.094268 kernel: ACPI: 3 ACPI AML tables successfully acquired and loaded Jul 2 07:48:28.094285 kernel: ACPI: Interpreter enabled Jul 2 07:48:28.094303 kernel: ACPI: PM: (supports S0 S3 S5) Jul 2 07:48:28.094320 kernel: ACPI: Using IOAPIC for interrupt routing Jul 2 07:48:28.094342 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jul 2 07:48:28.094360 kernel: ACPI: Enabled 16 GPEs in block 00 to 0F Jul 2 07:48:28.094378 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jul 2 07:48:28.094612 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Jul 2 07:48:28.096014 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. Jul 2 07:48:28.096045 kernel: PCI host bridge to bus 0000:00 Jul 2 07:48:28.096467 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jul 2 07:48:28.096875 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jul 2 07:48:28.097032 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jul 2 07:48:28.097182 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfefff window] Jul 2 07:48:28.097336 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jul 2 07:48:28.097518 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Jul 2 07:48:28.097705 kernel: pci 0000:00:01.0: [8086:7110] type 00 class 0x060100 Jul 2 07:48:28.097905 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Jul 2 07:48:28.098075 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Jul 2 07:48:28.098250 kernel: pci 0000:00:03.0: [1af4:1004] type 00 class 0x000000 Jul 2 07:48:28.098426 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc040-0xc07f] Jul 2 07:48:28.098594 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc0001000-0xc000107f] Jul 2 07:48:28.098795 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Jul 2 07:48:28.098969 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc03f] Jul 2 07:48:28.099144 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc0000000-0xc000007f] Jul 2 07:48:28.099326 kernel: pci 0000:00:05.0: [1af4:1005] type 00 class 0x00ff00 Jul 2 07:48:28.099492 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc080-0xc09f] Jul 2 07:48:28.099650 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xc0002000-0xc000203f] Jul 2 07:48:28.099672 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jul 2 07:48:28.099690 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jul 2 07:48:28.099708 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jul 2 07:48:28.099731 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jul 2 07:48:28.099748 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Jul 2 07:48:28.099779 kernel: iommu: Default domain type: Translated Jul 2 07:48:28.099797 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jul 2 07:48:28.099815 kernel: vgaarb: loaded Jul 2 07:48:28.099832 kernel: pps_core: LinuxPPS API ver. 1 registered Jul 2 07:48:28.099850 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Jul 2 07:48:28.099868 kernel: PTP clock support registered Jul 2 07:48:28.099886 kernel: Registered efivars operations Jul 2 07:48:28.099908 kernel: PCI: Using ACPI for IRQ routing Jul 2 07:48:28.099926 kernel: PCI: pci_cache_line_size set to 64 bytes Jul 2 07:48:28.099943 kernel: e820: reserve RAM buffer [mem 0x00055000-0x0005ffff] Jul 2 07:48:28.099961 kernel: e820: reserve RAM buffer [mem 0x00098000-0x0009ffff] Jul 2 07:48:28.099978 kernel: e820: reserve RAM buffer [mem 0xbf8ed000-0xbfffffff] Jul 2 07:48:28.099996 kernel: e820: reserve RAM buffer [mem 0xbffe0000-0xbfffffff] Jul 2 07:48:28.100012 kernel: clocksource: Switched to clocksource kvm-clock Jul 2 07:48:28.100029 kernel: VFS: Disk quotas dquot_6.6.0 Jul 2 07:48:28.100047 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jul 2 07:48:28.100069 kernel: pnp: PnP ACPI init Jul 2 07:48:28.100087 kernel: pnp: PnP ACPI: found 7 devices Jul 2 07:48:28.100104 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jul 2 07:48:28.100122 kernel: NET: Registered PF_INET protocol family Jul 2 07:48:28.100139 kernel: IP idents hash table entries: 131072 (order: 8, 1048576 bytes, linear) Jul 2 07:48:28.100157 kernel: tcp_listen_portaddr_hash hash table entries: 4096 (order: 4, 65536 bytes, linear) Jul 2 07:48:28.100175 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jul 2 07:48:28.100193 kernel: TCP established hash table entries: 65536 (order: 7, 524288 bytes, linear) Jul 2 07:48:28.100210 kernel: TCP bind hash table entries: 65536 (order: 8, 1048576 bytes, linear) Jul 2 07:48:28.100231 kernel: TCP: Hash tables configured (established 65536 bind 65536) Jul 2 07:48:28.100248 kernel: UDP hash table entries: 4096 (order: 5, 131072 bytes, linear) Jul 2 07:48:28.100273 kernel: UDP-Lite hash table entries: 4096 (order: 5, 131072 bytes, linear) Jul 2 07:48:28.100291 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jul 2 07:48:28.100309 kernel: NET: Registered PF_XDP protocol family Jul 2 07:48:28.100465 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jul 2 07:48:28.100618 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jul 2 07:48:28.111330 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jul 2 07:48:28.111536 kernel: pci_bus 0000:00: resource 7 [mem 0xc0000000-0xfebfefff window] Jul 2 07:48:28.111720 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jul 2 07:48:28.111746 kernel: PCI: CLS 0 bytes, default 64 Jul 2 07:48:28.111779 kernel: PCI-DMA: Using software bounce buffering for IO (SWIOTLB) Jul 2 07:48:28.111796 kernel: software IO TLB: mapped [mem 0x00000000b7ff7000-0x00000000bbff7000] (64MB) Jul 2 07:48:28.111812 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Jul 2 07:48:28.111829 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x212733415c7, max_idle_ns: 440795236380 ns Jul 2 07:48:28.111844 kernel: clocksource: Switched to clocksource tsc Jul 2 07:48:28.111866 kernel: Initialise system trusted keyrings Jul 2 07:48:28.111882 kernel: workingset: timestamp_bits=39 max_order=21 bucket_order=0 Jul 2 07:48:28.111898 kernel: Key type asymmetric registered Jul 2 07:48:28.111914 kernel: Asymmetric key parser 'x509' registered Jul 2 07:48:28.111930 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Jul 2 07:48:28.111946 kernel: io scheduler mq-deadline registered Jul 2 07:48:28.111961 kernel: io scheduler kyber registered Jul 2 07:48:28.111976 kernel: io scheduler bfq registered Jul 2 07:48:28.113286 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jul 2 07:48:28.113312 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Jul 2 07:48:28.113520 kernel: virtio-pci 0000:00:03.0: virtio_pci: leaving for legacy driver Jul 2 07:48:28.113543 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 10 Jul 2 07:48:28.113709 kernel: virtio-pci 0000:00:04.0: virtio_pci: leaving for legacy driver Jul 2 07:48:28.113733 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Jul 2 07:48:28.113924 kernel: virtio-pci 0000:00:05.0: virtio_pci: leaving for legacy driver Jul 2 07:48:28.113946 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jul 2 07:48:28.113962 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jul 2 07:48:28.113979 kernel: 00:04: ttyS1 at I/O 0x2f8 (irq = 3, base_baud = 115200) is a 16550A Jul 2 07:48:28.114000 kernel: 00:05: ttyS2 at I/O 0x3e8 (irq = 6, base_baud = 115200) is a 16550A Jul 2 07:48:28.114015 kernel: 00:06: ttyS3 at I/O 0x2e8 (irq = 7, base_baud = 115200) is a 16550A Jul 2 07:48:28.114202 kernel: tpm_tis MSFT0101:00: 2.0 TPM (device-id 0x9009, rev-id 0) Jul 2 07:48:28.114228 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jul 2 07:48:28.114245 kernel: i8042: Warning: Keylock active Jul 2 07:48:28.114279 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jul 2 07:48:28.114296 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jul 2 07:48:28.114468 kernel: rtc_cmos 00:00: RTC can wake from S4 Jul 2 07:48:28.114635 kernel: rtc_cmos 00:00: registered as rtc0 Jul 2 07:48:28.114867 kernel: rtc_cmos 00:00: setting system clock to 2024-07-02T07:48:27 UTC (1719906507) Jul 2 07:48:28.115029 kernel: rtc_cmos 00:00: alarms up to one day, 114 bytes nvram Jul 2 07:48:28.115051 kernel: intel_pstate: CPU model not supported Jul 2 07:48:28.115069 kernel: pstore: Registered efi as persistent store backend Jul 2 07:48:28.115086 kernel: NET: Registered PF_INET6 protocol family Jul 2 07:48:28.115104 kernel: Segment Routing with IPv6 Jul 2 07:48:28.115121 kernel: In-situ OAM (IOAM) with IPv6 Jul 2 07:48:28.115144 kernel: NET: Registered PF_PACKET protocol family Jul 2 07:48:28.115161 kernel: Key type dns_resolver registered Jul 2 07:48:28.115179 kernel: IPI shorthand broadcast: enabled Jul 2 07:48:28.115196 kernel: sched_clock: Marking stable (699195131, 123571462)->(838363614, -15597021) Jul 2 07:48:28.115213 kernel: registered taskstats version 1 Jul 2 07:48:28.115231 kernel: Loading compiled-in X.509 certificates Jul 2 07:48:28.115248 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jul 2 07:48:28.115273 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.161-flatcar: a1ce693884775675566f1ed116e36d15950b9a42' Jul 2 07:48:28.115290 kernel: Key type .fscrypt registered Jul 2 07:48:28.115312 kernel: Key type fscrypt-provisioning registered Jul 2 07:48:28.115330 kernel: pstore: Using crash dump compression: deflate Jul 2 07:48:28.115347 kernel: ima: Allocated hash algorithm: sha1 Jul 2 07:48:28.115364 kernel: ima: No architecture policies found Jul 2 07:48:28.115381 kernel: clk: Disabling unused clocks Jul 2 07:48:28.115398 kernel: Freeing unused kernel image (initmem) memory: 47444K Jul 2 07:48:28.115416 kernel: Write protecting the kernel read-only data: 28672k Jul 2 07:48:28.115433 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Jul 2 07:48:28.115455 kernel: Freeing unused kernel image (rodata/data gap) memory: 624K Jul 2 07:48:28.115472 kernel: Run /init as init process Jul 2 07:48:28.115490 kernel: with arguments: Jul 2 07:48:28.115506 kernel: /init Jul 2 07:48:28.115522 kernel: with environment: Jul 2 07:48:28.115539 kernel: HOME=/ Jul 2 07:48:28.115555 kernel: TERM=linux Jul 2 07:48:28.115573 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jul 2 07:48:28.115594 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Jul 2 07:48:28.115620 systemd[1]: Detected virtualization kvm. Jul 2 07:48:28.115639 systemd[1]: Detected architecture x86-64. Jul 2 07:48:28.115657 systemd[1]: Running in initrd. Jul 2 07:48:28.115674 systemd[1]: No hostname configured, using default hostname. Jul 2 07:48:28.115691 systemd[1]: Hostname set to . Jul 2 07:48:28.115710 systemd[1]: Initializing machine ID from VM UUID. Jul 2 07:48:28.115728 systemd[1]: Queued start job for default target initrd.target. Jul 2 07:48:28.115750 systemd[1]: Started systemd-ask-password-console.path. Jul 2 07:48:28.116811 systemd[1]: Reached target cryptsetup.target. Jul 2 07:48:28.116836 systemd[1]: Reached target paths.target. Jul 2 07:48:28.116856 systemd[1]: Reached target slices.target. Jul 2 07:48:28.116875 systemd[1]: Reached target swap.target. Jul 2 07:48:28.116893 systemd[1]: Reached target timers.target. Jul 2 07:48:28.116912 systemd[1]: Listening on iscsid.socket. Jul 2 07:48:28.116929 systemd[1]: Listening on iscsiuio.socket. Jul 2 07:48:28.116954 systemd[1]: Listening on systemd-journald-audit.socket. Jul 2 07:48:28.116974 systemd[1]: Listening on systemd-journald-dev-log.socket. Jul 2 07:48:28.116994 systemd[1]: Listening on systemd-journald.socket. Jul 2 07:48:28.117012 systemd[1]: Listening on systemd-networkd.socket. Jul 2 07:48:28.117029 systemd[1]: Listening on systemd-udevd-control.socket. Jul 2 07:48:28.117045 systemd[1]: Listening on systemd-udevd-kernel.socket. Jul 2 07:48:28.117062 systemd[1]: Reached target sockets.target. Jul 2 07:48:28.117081 systemd[1]: Starting kmod-static-nodes.service... Jul 2 07:48:28.117099 systemd[1]: Finished network-cleanup.service. Jul 2 07:48:28.117122 systemd[1]: Starting systemd-fsck-usr.service... Jul 2 07:48:28.117141 systemd[1]: Starting systemd-journald.service... Jul 2 07:48:28.117177 systemd[1]: Starting systemd-modules-load.service... Jul 2 07:48:28.117197 systemd[1]: Starting systemd-resolved.service... Jul 2 07:48:28.117214 systemd[1]: Starting systemd-vconsole-setup.service... Jul 2 07:48:28.117233 systemd[1]: Finished kmod-static-nodes.service. Jul 2 07:48:28.117264 kernel: audit: type=1130 audit(1719906508.080:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.117282 systemd[1]: Finished systemd-fsck-usr.service. Jul 2 07:48:28.117300 kernel: audit: type=1130 audit(1719906508.086:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.117329 systemd[1]: Finished systemd-vconsole-setup.service. Jul 2 07:48:28.117347 kernel: audit: type=1130 audit(1719906508.100:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.117364 systemd[1]: Starting dracut-cmdline-ask.service... Jul 2 07:48:28.117382 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Jul 2 07:48:28.117404 systemd-journald[189]: Journal started Jul 2 07:48:28.117493 systemd-journald[189]: Runtime Journal (/run/log/journal/2e199e2f4ec92ce0f0f20c9eac44439b) is 8.0M, max 148.8M, 140.8M free. Jul 2 07:48:28.080000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.086000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.100000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.094851 systemd-modules-load[190]: Inserted module 'overlay' Jul 2 07:48:28.125552 systemd[1]: Started systemd-journald.service. Jul 2 07:48:28.125596 kernel: audit: type=1130 audit(1719906508.121:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.121000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.139000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.140980 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Jul 2 07:48:28.144891 kernel: audit: type=1130 audit(1719906508.139:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.151536 systemd[1]: Finished dracut-cmdline-ask.service. Jul 2 07:48:28.160555 kernel: audit: type=1130 audit(1719906508.149:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.149000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.155853 systemd[1]: Starting dracut-cmdline.service... Jul 2 07:48:28.163313 systemd-resolved[191]: Positive Trust Anchors: Jul 2 07:48:28.163333 systemd-resolved[191]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 2 07:48:28.163391 systemd-resolved[191]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Jul 2 07:48:28.169362 systemd-resolved[191]: Defaulting to hostname 'linux'. Jul 2 07:48:28.191319 dracut-cmdline[206]: dracut-dracut-053 Jul 2 07:48:28.191319 dracut-cmdline[206]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 flatcar.first_boot=detected flatcar.oem.id=gce verity.usrhash=d29251fe942de56b08103b03939b6e5af4108e76dc6080fe2498c5db43f16e82 Jul 2 07:48:28.203874 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jul 2 07:48:28.203912 kernel: Bridge firewalling registered Jul 2 07:48:28.170951 systemd[1]: Started systemd-resolved.service. Jul 2 07:48:28.201063 systemd-modules-load[190]: Inserted module 'br_netfilter' Jul 2 07:48:28.214000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.216047 systemd[1]: Reached target nss-lookup.target. Jul 2 07:48:28.228922 kernel: audit: type=1130 audit(1719906508.214:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.233788 kernel: SCSI subsystem initialized Jul 2 07:48:28.252037 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jul 2 07:48:28.252091 kernel: device-mapper: uevent: version 1.0.3 Jul 2 07:48:28.252126 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Jul 2 07:48:28.258442 systemd-modules-load[190]: Inserted module 'dm_multipath' Jul 2 07:48:28.259565 systemd[1]: Finished systemd-modules-load.service. Jul 2 07:48:28.267988 kernel: audit: type=1130 audit(1719906508.262:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.262000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.267576 systemd[1]: Starting systemd-sysctl.service... Jul 2 07:48:28.281504 systemd[1]: Finished systemd-sysctl.service. Jul 2 07:48:28.291903 kernel: audit: type=1130 audit(1719906508.283:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.283000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.298782 kernel: Loading iSCSI transport class v2.0-870. Jul 2 07:48:28.319802 kernel: iscsi: registered transport (tcp) Jul 2 07:48:28.346204 kernel: iscsi: registered transport (qla4xxx) Jul 2 07:48:28.346277 kernel: QLogic iSCSI HBA Driver Jul 2 07:48:28.391776 systemd[1]: Finished dracut-cmdline.service. Jul 2 07:48:28.390000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.393967 systemd[1]: Starting dracut-pre-udev.service... Jul 2 07:48:28.449821 kernel: raid6: avx2x4 gen() 18597 MB/s Jul 2 07:48:28.467808 kernel: raid6: avx2x4 xor() 7600 MB/s Jul 2 07:48:28.484802 kernel: raid6: avx2x2 gen() 18524 MB/s Jul 2 07:48:28.501800 kernel: raid6: avx2x2 xor() 18627 MB/s Jul 2 07:48:28.518803 kernel: raid6: avx2x1 gen() 14463 MB/s Jul 2 07:48:28.535800 kernel: raid6: avx2x1 xor() 16188 MB/s Jul 2 07:48:28.552801 kernel: raid6: sse2x4 gen() 11099 MB/s Jul 2 07:48:28.569798 kernel: raid6: sse2x4 xor() 6888 MB/s Jul 2 07:48:28.586800 kernel: raid6: sse2x2 gen() 12124 MB/s Jul 2 07:48:28.603800 kernel: raid6: sse2x2 xor() 7461 MB/s Jul 2 07:48:28.620801 kernel: raid6: sse2x1 gen() 10574 MB/s Jul 2 07:48:28.638270 kernel: raid6: sse2x1 xor() 5207 MB/s Jul 2 07:48:28.638304 kernel: raid6: using algorithm avx2x4 gen() 18597 MB/s Jul 2 07:48:28.638334 kernel: raid6: .... xor() 7600 MB/s, rmw enabled Jul 2 07:48:28.639054 kernel: raid6: using avx2x2 recovery algorithm Jul 2 07:48:28.653798 kernel: xor: automatically using best checksumming function avx Jul 2 07:48:28.758801 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Jul 2 07:48:28.770633 systemd[1]: Finished dracut-pre-udev.service. Jul 2 07:48:28.772000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.772000 audit: BPF prog-id=7 op=LOAD Jul 2 07:48:28.772000 audit: BPF prog-id=8 op=LOAD Jul 2 07:48:28.775168 systemd[1]: Starting systemd-udevd.service... Jul 2 07:48:28.792330 systemd-udevd[388]: Using default interface naming scheme 'v252'. Jul 2 07:48:28.799736 systemd[1]: Started systemd-udevd.service. Jul 2 07:48:28.799000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.802526 systemd[1]: Starting dracut-pre-trigger.service... Jul 2 07:48:28.821732 dracut-pre-trigger[396]: rd.md=0: removing MD RAID activation Jul 2 07:48:28.857230 systemd[1]: Finished dracut-pre-trigger.service. Jul 2 07:48:28.855000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.859249 systemd[1]: Starting systemd-udev-trigger.service... Jul 2 07:48:28.924477 systemd[1]: Finished systemd-udev-trigger.service. Jul 2 07:48:28.926000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:28.991787 kernel: cryptd: max_cpu_qlen set to 1000 Jul 2 07:48:29.043458 kernel: AVX2 version of gcm_enc/dec engaged. Jul 2 07:48:29.043538 kernel: AES CTR mode by8 optimization enabled Jul 2 07:48:29.045790 kernel: scsi host0: Virtio SCSI HBA Jul 2 07:48:29.053004 kernel: scsi 0:0:1:0: Direct-Access Google PersistentDisk 1 PQ: 0 ANSI: 6 Jul 2 07:48:29.136236 kernel: sd 0:0:1:0: [sda] 25165824 512-byte logical blocks: (12.9 GB/12.0 GiB) Jul 2 07:48:29.136540 kernel: sd 0:0:1:0: [sda] 4096-byte physical blocks Jul 2 07:48:29.136732 kernel: sd 0:0:1:0: [sda] Write Protect is off Jul 2 07:48:29.137785 kernel: sd 0:0:1:0: [sda] Mode Sense: 1f 00 00 08 Jul 2 07:48:29.138027 kernel: sd 0:0:1:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Jul 2 07:48:29.149616 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jul 2 07:48:29.149670 kernel: GPT:17805311 != 25165823 Jul 2 07:48:29.149692 kernel: GPT:Alternate GPT header not at the end of the disk. Jul 2 07:48:29.149722 kernel: GPT:17805311 != 25165823 Jul 2 07:48:29.150286 kernel: GPT: Use GNU Parted to correct GPT errors. Jul 2 07:48:29.151853 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 2 07:48:29.153429 kernel: sd 0:0:1:0: [sda] Attached SCSI disk Jul 2 07:48:29.203788 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by (udev-worker) (431) Jul 2 07:48:29.217392 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Jul 2 07:48:29.226119 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Jul 2 07:48:29.231297 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Jul 2 07:48:29.231512 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Jul 2 07:48:29.242982 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Jul 2 07:48:29.245023 systemd[1]: Starting disk-uuid.service... Jul 2 07:48:29.256975 disk-uuid[511]: Primary Header is updated. Jul 2 07:48:29.256975 disk-uuid[511]: Secondary Entries is updated. Jul 2 07:48:29.256975 disk-uuid[511]: Secondary Header is updated. Jul 2 07:48:29.268788 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 2 07:48:29.287797 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 2 07:48:29.295969 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 2 07:48:30.295422 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jul 2 07:48:30.295510 disk-uuid[512]: The operation has completed successfully. Jul 2 07:48:30.360307 systemd[1]: disk-uuid.service: Deactivated successfully. Jul 2 07:48:30.366000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.366000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.360465 systemd[1]: Finished disk-uuid.service. Jul 2 07:48:30.378890 systemd[1]: Starting verity-setup.service... Jul 2 07:48:30.406803 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Jul 2 07:48:30.478192 systemd[1]: Found device dev-mapper-usr.device. Jul 2 07:48:30.489221 systemd[1]: Mounting sysusr-usr.mount... Jul 2 07:48:30.502230 systemd[1]: Finished verity-setup.service. Jul 2 07:48:30.516000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.587787 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Jul 2 07:48:30.588562 systemd[1]: Mounted sysusr-usr.mount. Jul 2 07:48:30.596062 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Jul 2 07:48:30.596986 systemd[1]: Starting ignition-setup.service... Jul 2 07:48:30.650913 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 2 07:48:30.650958 kernel: BTRFS info (device sda6): using free space tree Jul 2 07:48:30.650982 kernel: BTRFS info (device sda6): has skinny extents Jul 2 07:48:30.651003 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 2 07:48:30.645278 systemd[1]: Starting parse-ip-for-networkd.service... Jul 2 07:48:30.660246 systemd[1]: mnt-oem.mount: Deactivated successfully. Jul 2 07:48:30.680378 systemd[1]: Finished ignition-setup.service. Jul 2 07:48:30.678000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.682065 systemd[1]: Starting ignition-fetch-offline.service... Jul 2 07:48:30.759083 systemd[1]: Finished parse-ip-for-networkd.service. Jul 2 07:48:30.757000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.758000 audit: BPF prog-id=9 op=LOAD Jul 2 07:48:30.761439 systemd[1]: Starting systemd-networkd.service... Jul 2 07:48:30.795519 systemd-networkd[686]: lo: Link UP Jul 2 07:48:30.795532 systemd-networkd[686]: lo: Gained carrier Jul 2 07:48:30.796329 systemd-networkd[686]: Enumeration completed Jul 2 07:48:30.796710 systemd-networkd[686]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 2 07:48:30.796900 systemd[1]: Started systemd-networkd.service. Jul 2 07:48:30.798913 systemd-networkd[686]: eth0: Link UP Jul 2 07:48:30.798920 systemd-networkd[686]: eth0: Gained carrier Jul 2 07:48:30.849000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.808880 systemd-networkd[686]: eth0: DHCPv4 address 10.128.0.12/32, gateway 10.128.0.1 acquired from 169.254.169.254 Jul 2 07:48:30.879000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.851041 systemd[1]: Reached target network.target. Jul 2 07:48:30.908061 iscsid[695]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Jul 2 07:48:30.908061 iscsid[695]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log Jul 2 07:48:30.908061 iscsid[695]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Jul 2 07:48:30.908061 iscsid[695]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Jul 2 07:48:30.908061 iscsid[695]: If using hardware iscsi like qla4xxx this message can be ignored. Jul 2 07:48:30.908061 iscsid[695]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Jul 2 07:48:30.908061 iscsid[695]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Jul 2 07:48:30.913000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.959000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.985000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.859963 systemd[1]: Starting iscsiuio.service... Jul 2 07:48:30.942101 ignition[603]: Ignition 2.14.0 Jul 2 07:48:31.084000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.868620 systemd[1]: Started iscsiuio.service. Jul 2 07:48:30.942114 ignition[603]: Stage: fetch-offline Jul 2 07:48:30.882091 systemd[1]: Starting iscsid.service... Jul 2 07:48:31.112000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.942186 ignition[603]: reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:30.901000 systemd[1]: Started iscsid.service. Jul 2 07:48:30.942226 ignition[603]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:30.916265 systemd[1]: Starting dracut-initqueue.service... Jul 2 07:48:30.958996 ignition[603]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:31.160000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.936024 systemd[1]: Finished dracut-initqueue.service. Jul 2 07:48:30.959180 ignition[603]: parsed url from cmdline: "" Jul 2 07:48:30.961330 systemd[1]: Finished ignition-fetch-offline.service. Jul 2 07:48:30.959185 ignition[603]: no config URL provided Jul 2 07:48:30.987216 systemd[1]: Reached target remote-fs-pre.target. Jul 2 07:48:31.200000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:30.959192 ignition[603]: reading system config file "/usr/lib/ignition/user.ign" Jul 2 07:48:31.003037 systemd[1]: Reached target remote-cryptsetup.target. Jul 2 07:48:30.959202 ignition[603]: no config at "/usr/lib/ignition/user.ign" Jul 2 07:48:31.020040 systemd[1]: Reached target remote-fs.target. Jul 2 07:48:30.959210 ignition[603]: failed to fetch config: resource requires networking Jul 2 07:48:31.039149 systemd[1]: Starting dracut-pre-mount.service... Jul 2 07:48:30.959525 ignition[603]: Ignition finished successfully Jul 2 07:48:31.064096 systemd[1]: Starting ignition-fetch.service... Jul 2 07:48:31.076006 ignition[710]: Ignition 2.14.0 Jul 2 07:48:31.077371 systemd[1]: Finished dracut-pre-mount.service. Jul 2 07:48:31.076014 ignition[710]: Stage: fetch Jul 2 07:48:31.103342 unknown[710]: fetched base config from "system" Jul 2 07:48:31.076130 ignition[710]: reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:31.103424 unknown[710]: fetched base config from "system" Jul 2 07:48:31.076161 ignition[710]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:31.103432 unknown[710]: fetched user config from "gcp" Jul 2 07:48:31.086265 ignition[710]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:31.105705 systemd[1]: Finished ignition-fetch.service. Jul 2 07:48:31.086449 ignition[710]: parsed url from cmdline: "" Jul 2 07:48:31.114943 systemd[1]: Starting ignition-kargs.service... Jul 2 07:48:31.086459 ignition[710]: no config URL provided Jul 2 07:48:31.146253 systemd[1]: Finished ignition-kargs.service. Jul 2 07:48:31.086471 ignition[710]: reading system config file "/usr/lib/ignition/user.ign" Jul 2 07:48:31.163081 systemd[1]: Starting ignition-disks.service... Jul 2 07:48:31.086490 ignition[710]: no config at "/usr/lib/ignition/user.ign" Jul 2 07:48:31.195295 systemd[1]: Finished ignition-disks.service. Jul 2 07:48:31.086530 ignition[710]: GET http://169.254.169.254/computeMetadata/v1/instance/attributes/user-data: attempt #1 Jul 2 07:48:31.202327 systemd[1]: Reached target initrd-root-device.target. Jul 2 07:48:31.096274 ignition[710]: GET result: OK Jul 2 07:48:31.224051 systemd[1]: Reached target local-fs-pre.target. Jul 2 07:48:31.096355 ignition[710]: parsing config with SHA512: aebcf11d39c80b07c57e5dde27fe16969651cb23c6f6762fde18f63c30096d6183318d7b34a5060df04695c83b5f7cf41cc22183d71324c4fe676a8a81fd7c76 Jul 2 07:48:31.241888 systemd[1]: Reached target local-fs.target. Jul 2 07:48:31.104110 ignition[710]: fetch: fetch complete Jul 2 07:48:31.255889 systemd[1]: Reached target sysinit.target. Jul 2 07:48:31.104116 ignition[710]: fetch: fetch passed Jul 2 07:48:31.269880 systemd[1]: Reached target basic.target. Jul 2 07:48:31.104161 ignition[710]: Ignition finished successfully Jul 2 07:48:31.284231 systemd[1]: Starting systemd-fsck-root.service... Jul 2 07:48:31.127895 ignition[716]: Ignition 2.14.0 Jul 2 07:48:31.127906 ignition[716]: Stage: kargs Jul 2 07:48:31.128037 ignition[716]: reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:31.128073 ignition[716]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:31.135242 ignition[716]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:31.136583 ignition[716]: kargs: kargs passed Jul 2 07:48:31.136632 ignition[716]: Ignition finished successfully Jul 2 07:48:31.173736 ignition[722]: Ignition 2.14.0 Jul 2 07:48:31.173745 ignition[722]: Stage: disks Jul 2 07:48:31.173903 ignition[722]: reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:31.173939 ignition[722]: parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:31.181512 ignition[722]: no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:31.182847 ignition[722]: disks: disks passed Jul 2 07:48:31.182893 ignition[722]: Ignition finished successfully Jul 2 07:48:31.322712 systemd-fsck[730]: ROOT: clean, 614/1628000 files, 124057/1617920 blocks Jul 2 07:48:31.544695 systemd[1]: Finished systemd-fsck-root.service. Jul 2 07:48:31.551000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:31.554178 systemd[1]: Mounting sysroot.mount... Jul 2 07:48:31.582914 kernel: EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Jul 2 07:48:31.579024 systemd[1]: Mounted sysroot.mount. Jul 2 07:48:31.590144 systemd[1]: Reached target initrd-root-fs.target. Jul 2 07:48:31.611122 systemd[1]: Mounting sysroot-usr.mount... Jul 2 07:48:31.628364 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Jul 2 07:48:31.628530 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jul 2 07:48:31.628583 systemd[1]: Reached target ignition-diskful.target. Jul 2 07:48:31.649288 systemd[1]: Mounted sysroot-usr.mount. Jul 2 07:48:31.717267 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (736) Jul 2 07:48:31.717300 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 2 07:48:31.717316 kernel: BTRFS info (device sda6): using free space tree Jul 2 07:48:31.717330 kernel: BTRFS info (device sda6): has skinny extents Jul 2 07:48:31.674058 systemd[1]: Mounting sysroot-usr-share-oem.mount... Jul 2 07:48:31.738907 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 2 07:48:31.685998 systemd[1]: Starting initrd-setup-root.service... Jul 2 07:48:31.742311 systemd[1]: Mounted sysroot-usr-share-oem.mount. Jul 2 07:48:31.756944 initrd-setup-root[741]: cut: /sysroot/etc/passwd: No such file or directory Jul 2 07:48:31.766886 initrd-setup-root[765]: cut: /sysroot/etc/group: No such file or directory Jul 2 07:48:31.784896 initrd-setup-root[775]: cut: /sysroot/etc/shadow: No such file or directory Jul 2 07:48:31.794897 initrd-setup-root[783]: cut: /sysroot/etc/gshadow: No such file or directory Jul 2 07:48:31.809372 systemd[1]: Finished initrd-setup-root.service. Jul 2 07:48:31.807000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:31.810553 systemd[1]: Starting ignition-mount.service... Jul 2 07:48:31.838759 systemd[1]: Starting sysroot-boot.service... Jul 2 07:48:31.846939 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. Jul 2 07:48:31.847044 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. Jul 2 07:48:31.872878 ignition[801]: INFO : Ignition 2.14.0 Jul 2 07:48:31.872878 ignition[801]: INFO : Stage: mount Jul 2 07:48:31.872878 ignition[801]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:31.872878 ignition[801]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:31.896000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:31.884345 systemd[1]: Finished sysroot-boot.service. Jul 2 07:48:31.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:31.938028 ignition[801]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:31.938028 ignition[801]: INFO : mount: mount passed Jul 2 07:48:31.938028 ignition[801]: INFO : Ignition finished successfully Jul 2 07:48:32.010878 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (811) Jul 2 07:48:32.010925 kernel: BTRFS info (device sda6): using crc32c (crc32c-intel) checksum algorithm Jul 2 07:48:32.010948 kernel: BTRFS info (device sda6): using free space tree Jul 2 07:48:32.010969 kernel: BTRFS info (device sda6): has skinny extents Jul 2 07:48:32.010989 kernel: BTRFS info (device sda6): enabling ssd optimizations Jul 2 07:48:31.898340 systemd[1]: Finished ignition-mount.service. Jul 2 07:48:31.932035 systemd[1]: Starting ignition-files.service... Jul 2 07:48:31.949685 systemd[1]: Mounting sysroot-usr-share-oem.mount... Jul 2 07:48:32.043943 ignition[830]: INFO : Ignition 2.14.0 Jul 2 07:48:32.043943 ignition[830]: INFO : Stage: files Jul 2 07:48:32.043943 ignition[830]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:32.043943 ignition[830]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:32.043943 ignition[830]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:32.107868 kernel: BTRFS info: devid 1 device path /dev/sda6 changed to /dev/disk/by-label/OEM scanned by ignition (833) Jul 2 07:48:32.005530 systemd[1]: Mounted sysroot-usr-share-oem.mount. Jul 2 07:48:32.115889 ignition[830]: DEBUG : files: compiled without relabeling support, skipping Jul 2 07:48:32.115889 ignition[830]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jul 2 07:48:32.115889 ignition[830]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jul 2 07:48:32.115889 ignition[830]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jul 2 07:48:32.115889 ignition[830]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jul 2 07:48:32.115889 ignition[830]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/hosts" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): oem config not found in "/usr/share/oem", looking on oem partition Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(4): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1987197591" Jul 2 07:48:32.115889 ignition[830]: CRITICAL : files: createFilesystemsFiles: createFiles: op(3): op(4): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1987197591": device or resource busy Jul 2 07:48:32.115889 ignition[830]: ERROR : files: createFilesystemsFiles: createFiles: op(3): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem1987197591", trying btrfs: device or resource busy Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(5): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1987197591" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(5): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem1987197591" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(6): [started] unmounting "/mnt/oem1987197591" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): op(6): [finished] unmounting "/mnt/oem1987197591" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/hosts" Jul 2 07:48:32.115889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 2 07:48:32.057416 unknown[830]: wrote ssh authorized keys file for user: core Jul 2 07:48:32.373889 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jul 2 07:48:32.521023 systemd-networkd[686]: eth0: Gained IPv6LL Jul 2 07:48:35.242112 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Jul 2 07:48:35.387122 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jul 2 07:48:35.403899 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 2 07:48:35.403899 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Jul 2 07:48:35.638647 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET result: OK Jul 2 07:48:35.785554 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/profile.d/google-cloud-sdk.sh" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): oem config not found in "/usr/share/oem", looking on oem partition Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3465745116" Jul 2 07:48:35.800925 ignition[830]: CRITICAL : files: createFilesystemsFiles: createFiles: op(9): op(a): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3465745116": device or resource busy Jul 2 07:48:35.800925 ignition[830]: ERROR : files: createFilesystemsFiles: createFiles: op(9): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem3465745116", trying btrfs: device or resource busy Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(b): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3465745116" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(b): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3465745116" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(c): [started] unmounting "/mnt/oem3465745116" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): op(c): [finished] unmounting "/mnt/oem3465745116" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/profile.d/google-cloud-sdk.sh" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/home/core/install.sh" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/home/core/install.sh" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/home/core/nginx.yaml" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/home/core/nginx.yaml" Jul 2 07:48:35.800925 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(10): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(10): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(11): [started] writing file "/sysroot/etc/flatcar/update.conf" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(11): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(12): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(12): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): [started] writing file "/sysroot/etc/systemd/system/oem-gce-enable-oslogin.service" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): oem config not found in "/usr/share/oem", looking on oem partition Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(14): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3810732030" Jul 2 07:48:36.045909 ignition[830]: CRITICAL : files: createFilesystemsFiles: createFiles: op(13): op(14): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3810732030": device or resource busy Jul 2 07:48:36.045909 ignition[830]: ERROR : files: createFilesystemsFiles: createFiles: op(13): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem3810732030", trying btrfs: device or resource busy Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(15): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3810732030" Jul 2 07:48:36.045909 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(15): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3810732030" Jul 2 07:48:35.819061 systemd[1]: mnt-oem3810732030.mount: Deactivated successfully. Jul 2 07:48:36.301003 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(16): [started] unmounting "/mnt/oem3810732030" Jul 2 07:48:36.301003 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): op(16): [finished] unmounting "/mnt/oem3810732030" Jul 2 07:48:36.301003 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(13): [finished] writing file "/sysroot/etc/systemd/system/oem-gce-enable-oslogin.service" Jul 2 07:48:36.301003 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(17): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jul 2 07:48:36.301003 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(17): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Jul 2 07:48:36.301003 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(17): GET result: OK Jul 2 07:48:36.430940 kernel: kauditd_printk_skb: 26 callbacks suppressed Jul 2 07:48:36.430982 kernel: audit: type=1130 audit(1719906516.386:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.386000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.381561 systemd[1]: Finished ignition-files.service. Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(17): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): [started] writing file "/sysroot/etc/systemd/system/oem-gce.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): oem config not found in "/usr/share/oem", looking on oem partition Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(19): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem641481073" Jul 2 07:48:36.446031 ignition[830]: CRITICAL : files: createFilesystemsFiles: createFiles: op(18): op(19): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem641481073": device or resource busy Jul 2 07:48:36.446031 ignition[830]: ERROR : files: createFilesystemsFiles: createFiles: op(18): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem641481073", trying btrfs: device or resource busy Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1a): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem641481073" Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1a): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem641481073" Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1b): [started] unmounting "/mnt/oem641481073" Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): op(1b): [finished] unmounting "/mnt/oem641481073" Jul 2 07:48:36.446031 ignition[830]: INFO : files: createFilesystemsFiles: createFiles: op(18): [finished] writing file "/sysroot/etc/systemd/system/oem-gce.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: op(1c): [started] processing unit "coreos-metadata-sshkeys@.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: op(1c): [finished] processing unit "coreos-metadata-sshkeys@.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: op(1d): [started] processing unit "oem-gce.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: op(1d): [finished] processing unit "oem-gce.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: op(1e): [started] processing unit "oem-gce-enable-oslogin.service" Jul 2 07:48:36.446031 ignition[830]: INFO : files: op(1e): [finished] processing unit "oem-gce-enable-oslogin.service" Jul 2 07:48:36.956894 kernel: audit: type=1130 audit(1719906516.479:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.956956 kernel: audit: type=1130 audit(1719906516.518:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.956981 kernel: audit: type=1131 audit(1719906516.518:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.957003 kernel: audit: type=1130 audit(1719906516.638:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.957021 kernel: audit: type=1131 audit(1719906516.638:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.957036 kernel: audit: type=1130 audit(1719906516.772:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.957054 kernel: audit: type=1131 audit(1719906516.882:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.479000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.518000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.518000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.638000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.638000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.772000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.882000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.398052 systemd[1]: Starting initrd-setup-root-after-ignition.service... Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(1f): [started] processing unit "prepare-helm.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(1f): op(20): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(1f): op(20): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(1f): [finished] processing unit "prepare-helm.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(21): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(21): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(22): [started] setting preset to enabled for "oem-gce.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(22): [finished] setting preset to enabled for "oem-gce.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(23): [started] setting preset to enabled for "oem-gce-enable-oslogin.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(23): [finished] setting preset to enabled for "oem-gce-enable-oslogin.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(24): [started] setting preset to enabled for "prepare-helm.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: op(24): [finished] setting preset to enabled for "prepare-helm.service" Jul 2 07:48:36.974927 ignition[830]: INFO : files: createResultFile: createFiles: op(25): [started] writing file "/sysroot/etc/.ignition-result.json" Jul 2 07:48:36.974927 ignition[830]: INFO : files: createResultFile: createFiles: op(25): [finished] writing file "/sysroot/etc/.ignition-result.json" Jul 2 07:48:36.974927 ignition[830]: INFO : files: files passed Jul 2 07:48:36.974927 ignition[830]: INFO : Ignition finished successfully Jul 2 07:48:37.305881 kernel: audit: type=1131 audit(1719906517.129:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.305929 kernel: audit: type=1131 audit(1719906517.187:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.129000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.187000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.225000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.244000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.438941 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Jul 2 07:48:37.312000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.330990 iscsid[695]: iscsid shutting down. Jul 2 07:48:37.338032 initrd-setup-root-after-ignition[853]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jul 2 07:48:36.440075 systemd[1]: Starting ignition-quench.service... Jul 2 07:48:37.357000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.453224 systemd[1]: Finished initrd-setup-root-after-ignition.service. Jul 2 07:48:37.379000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.481272 systemd[1]: ignition-quench.service: Deactivated successfully. Jul 2 07:48:37.395000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.481411 systemd[1]: Finished ignition-quench.service. Jul 2 07:48:37.411000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.520244 systemd[1]: Reached target ignition-complete.target. Jul 2 07:48:37.426000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.583246 systemd[1]: Starting initrd-parse-etc.service... Jul 2 07:48:37.442000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.626220 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jul 2 07:48:37.458000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.468108 ignition[868]: INFO : Ignition 2.14.0 Jul 2 07:48:37.468108 ignition[868]: INFO : Stage: umount Jul 2 07:48:37.468108 ignition[868]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Jul 2 07:48:37.468108 ignition[868]: DEBUG : parsing config with SHA512: 28536912712fffc63406b6accf8759a9de2528d78fa3e153de6c4a0ac81102f9876238326a650eaef6ce96ba6e26bae8fbbfe85a3f956a15fdad11da447b6af6 Jul 2 07:48:37.468108 ignition[868]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/gcp" Jul 2 07:48:37.468108 ignition[868]: INFO : umount: umount passed Jul 2 07:48:37.468108 ignition[868]: INFO : Ignition finished successfully Jul 2 07:48:37.474000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.481000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.626338 systemd[1]: Finished initrd-parse-etc.service. Jul 2 07:48:36.640084 systemd[1]: Reached target initrd-fs.target. Jul 2 07:48:37.598000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.701106 systemd[1]: Reached target initrd.target. Jul 2 07:48:37.613000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.723107 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Jul 2 07:48:36.724235 systemd[1]: Starting dracut-pre-pivot.service... Jul 2 07:48:37.646000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.752169 systemd[1]: Finished dracut-pre-pivot.service. Jul 2 07:48:37.662000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.662000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.775205 systemd[1]: Starting initrd-cleanup.service... Jul 2 07:48:36.818491 systemd[1]: Stopped target nss-lookup.target. Jul 2 07:48:36.829180 systemd[1]: Stopped target remote-cryptsetup.target. Jul 2 07:48:36.847213 systemd[1]: Stopped target timers.target. Jul 2 07:48:36.867157 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jul 2 07:48:36.867334 systemd[1]: Stopped dracut-pre-pivot.service. Jul 2 07:48:37.735000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.884378 systemd[1]: Stopped target initrd.target. Jul 2 07:48:37.757000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.757000 audit: BPF prog-id=6 op=UNLOAD Jul 2 07:48:36.921252 systemd[1]: Stopped target basic.target. Jul 2 07:48:36.939202 systemd[1]: Stopped target ignition-complete.target. Jul 2 07:48:36.965125 systemd[1]: Stopped target ignition-diskful.target. Jul 2 07:48:37.800000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.975191 systemd[1]: Stopped target initrd-root-device.target. Jul 2 07:48:37.816000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:36.993165 systemd[1]: Stopped target remote-fs.target. Jul 2 07:48:37.832000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.017276 systemd[1]: Stopped target remote-fs-pre.target. Jul 2 07:48:37.039198 systemd[1]: Stopped target sysinit.target. Jul 2 07:48:37.854000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.057196 systemd[1]: Stopped target local-fs.target. Jul 2 07:48:37.076195 systemd[1]: Stopped target local-fs-pre.target. Jul 2 07:48:37.095207 systemd[1]: Stopped target swap.target. Jul 2 07:48:37.905000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.113124 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jul 2 07:48:37.920000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.113509 systemd[1]: Stopped dracut-pre-mount.service. Jul 2 07:48:37.936000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.131388 systemd[1]: Stopped target cryptsetup.target. Jul 2 07:48:37.170156 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jul 2 07:48:37.170341 systemd[1]: Stopped dracut-initqueue.service. Jul 2 07:48:37.967000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.189395 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jul 2 07:48:37.989000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.189655 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Jul 2 07:48:38.007000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:38.007000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:37.227324 systemd[1]: ignition-files.service: Deactivated successfully. Jul 2 07:48:37.227533 systemd[1]: Stopped ignition-files.service. Jul 2 07:48:37.247630 systemd[1]: Stopping ignition-mount.service... Jul 2 07:48:37.282359 systemd[1]: Stopping iscsid.service... Jul 2 07:48:37.294009 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jul 2 07:48:37.294195 systemd[1]: Stopped kmod-static-nodes.service. Jul 2 07:48:38.082895 systemd-journald[189]: Received SIGTERM from PID 1 (systemd). Jul 2 07:48:37.315467 systemd[1]: Stopping sysroot-boot.service... Jul 2 07:48:37.344951 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jul 2 07:48:37.345178 systemd[1]: Stopped systemd-udev-trigger.service. Jul 2 07:48:37.359213 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jul 2 07:48:37.359369 systemd[1]: Stopped dracut-pre-trigger.service. Jul 2 07:48:37.384972 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jul 2 07:48:37.385915 systemd[1]: iscsid.service: Deactivated successfully. Jul 2 07:48:37.386020 systemd[1]: Stopped iscsid.service. Jul 2 07:48:37.397680 systemd[1]: ignition-mount.service: Deactivated successfully. Jul 2 07:48:37.397800 systemd[1]: Stopped ignition-mount.service. Jul 2 07:48:37.413553 systemd[1]: sysroot-boot.service: Deactivated successfully. Jul 2 07:48:37.413660 systemd[1]: Stopped sysroot-boot.service. Jul 2 07:48:37.428494 systemd[1]: ignition-disks.service: Deactivated successfully. Jul 2 07:48:37.428637 systemd[1]: Stopped ignition-disks.service. Jul 2 07:48:37.443948 systemd[1]: ignition-kargs.service: Deactivated successfully. Jul 2 07:48:37.444016 systemd[1]: Stopped ignition-kargs.service. Jul 2 07:48:37.459954 systemd[1]: ignition-fetch.service: Deactivated successfully. Jul 2 07:48:37.460024 systemd[1]: Stopped ignition-fetch.service. Jul 2 07:48:37.475982 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jul 2 07:48:37.476071 systemd[1]: Stopped ignition-fetch-offline.service. Jul 2 07:48:37.483117 systemd[1]: Stopped target paths.target. Jul 2 07:48:37.497059 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jul 2 07:48:37.500860 systemd[1]: Stopped systemd-ask-password-console.path. Jul 2 07:48:37.514036 systemd[1]: Stopped target slices.target. Jul 2 07:48:37.556897 systemd[1]: Stopped target sockets.target. Jul 2 07:48:37.564066 systemd[1]: iscsid.socket: Deactivated successfully. Jul 2 07:48:37.564119 systemd[1]: Closed iscsid.socket. Jul 2 07:48:37.585019 systemd[1]: ignition-setup.service: Deactivated successfully. Jul 2 07:48:37.585090 systemd[1]: Stopped ignition-setup.service. Jul 2 07:48:37.600061 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jul 2 07:48:37.600122 systemd[1]: Stopped initrd-setup-root.service. Jul 2 07:48:37.615129 systemd[1]: Stopping iscsiuio.service... Jul 2 07:48:37.630365 systemd[1]: iscsiuio.service: Deactivated successfully. Jul 2 07:48:37.630474 systemd[1]: Stopped iscsiuio.service. Jul 2 07:48:37.648257 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jul 2 07:48:37.648368 systemd[1]: Finished initrd-cleanup.service. Jul 2 07:48:37.664887 systemd[1]: Stopped target network.target. Jul 2 07:48:37.678896 systemd[1]: iscsiuio.socket: Deactivated successfully. Jul 2 07:48:37.678958 systemd[1]: Closed iscsiuio.socket. Jul 2 07:48:37.693081 systemd[1]: Stopping systemd-networkd.service... Jul 2 07:48:37.696819 systemd-networkd[686]: eth0: DHCPv6 lease lost Jul 2 07:48:37.707991 systemd[1]: Stopping systemd-resolved.service... Jul 2 07:48:37.723275 systemd[1]: systemd-resolved.service: Deactivated successfully. Jul 2 07:48:37.723393 systemd[1]: Stopped systemd-resolved.service. Jul 2 07:48:38.090000 audit: BPF prog-id=9 op=UNLOAD Jul 2 07:48:37.737631 systemd[1]: systemd-networkd.service: Deactivated successfully. Jul 2 07:48:37.737745 systemd[1]: Stopped systemd-networkd.service. Jul 2 07:48:37.759590 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jul 2 07:48:37.759639 systemd[1]: Closed systemd-networkd.socket. Jul 2 07:48:37.774843 systemd[1]: Stopping network-cleanup.service... Jul 2 07:48:37.787866 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jul 2 07:48:37.787950 systemd[1]: Stopped parse-ip-for-networkd.service. Jul 2 07:48:37.801965 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 2 07:48:37.802023 systemd[1]: Stopped systemd-sysctl.service. Jul 2 07:48:37.818134 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jul 2 07:48:37.818192 systemd[1]: Stopped systemd-modules-load.service. Jul 2 07:48:37.834117 systemd[1]: Stopping systemd-udevd.service... Jul 2 07:48:37.849834 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Jul 2 07:48:37.850462 systemd[1]: systemd-udevd.service: Deactivated successfully. Jul 2 07:48:37.850610 systemd[1]: Stopped systemd-udevd.service. Jul 2 07:48:37.857586 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jul 2 07:48:37.857689 systemd[1]: Closed systemd-udevd-control.socket. Jul 2 07:48:37.876989 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jul 2 07:48:37.877053 systemd[1]: Closed systemd-udevd-kernel.socket. Jul 2 07:48:37.891957 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jul 2 07:48:37.892046 systemd[1]: Stopped dracut-pre-udev.service. Jul 2 07:48:37.906988 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jul 2 07:48:37.907057 systemd[1]: Stopped dracut-cmdline.service. Jul 2 07:48:37.921961 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jul 2 07:48:37.922028 systemd[1]: Stopped dracut-cmdline-ask.service. Jul 2 07:48:37.938944 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Jul 2 07:48:37.960948 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jul 2 07:48:37.961030 systemd[1]: Stopped systemd-vconsole-setup.service. Jul 2 07:48:37.969661 systemd[1]: network-cleanup.service: Deactivated successfully. Jul 2 07:48:37.969805 systemd[1]: Stopped network-cleanup.service. Jul 2 07:48:37.991315 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jul 2 07:48:37.991442 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Jul 2 07:48:38.009237 systemd[1]: Reached target initrd-switch-root.target. Jul 2 07:48:38.025950 systemd[1]: Starting initrd-switch-root.service... Jul 2 07:48:38.047953 systemd[1]: Switching root. Jul 2 07:48:38.093796 systemd-journald[189]: Journal stopped Jul 2 07:48:42.653020 kernel: SELinux: Class mctp_socket not defined in policy. Jul 2 07:48:42.653102 kernel: SELinux: Class anon_inode not defined in policy. Jul 2 07:48:42.653123 kernel: SELinux: the above unknown classes and permissions will be allowed Jul 2 07:48:42.653137 kernel: SELinux: policy capability network_peer_controls=1 Jul 2 07:48:42.653156 kernel: SELinux: policy capability open_perms=1 Jul 2 07:48:42.653174 kernel: SELinux: policy capability extended_socket_class=1 Jul 2 07:48:42.653192 kernel: SELinux: policy capability always_check_network=0 Jul 2 07:48:42.653207 kernel: SELinux: policy capability cgroup_seclabel=1 Jul 2 07:48:42.653221 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jul 2 07:48:42.653236 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jul 2 07:48:42.653253 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jul 2 07:48:42.653276 systemd[1]: Successfully loaded SELinux policy in 109.072ms. Jul 2 07:48:42.653307 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.229ms. Jul 2 07:48:42.653323 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Jul 2 07:48:42.653342 systemd[1]: Detected virtualization kvm. Jul 2 07:48:42.653358 systemd[1]: Detected architecture x86-64. Jul 2 07:48:42.653373 systemd[1]: Detected first boot. Jul 2 07:48:42.653392 systemd[1]: Initializing machine ID from VM UUID. Jul 2 07:48:42.653411 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Jul 2 07:48:42.653427 systemd[1]: Populated /etc with preset unit settings. Jul 2 07:48:42.653442 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Jul 2 07:48:42.653462 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Jul 2 07:48:42.653478 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 2 07:48:42.653495 kernel: kauditd_printk_skb: 48 callbacks suppressed Jul 2 07:48:42.653509 kernel: audit: type=1334 audit(1719906521.736:88): prog-id=12 op=LOAD Jul 2 07:48:42.653526 kernel: audit: type=1334 audit(1719906521.736:89): prog-id=3 op=UNLOAD Jul 2 07:48:42.653540 kernel: audit: type=1334 audit(1719906521.742:90): prog-id=13 op=LOAD Jul 2 07:48:42.653553 kernel: audit: type=1334 audit(1719906521.749:91): prog-id=14 op=LOAD Jul 2 07:48:42.653567 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jul 2 07:48:42.653582 kernel: audit: type=1334 audit(1719906521.749:92): prog-id=4 op=UNLOAD Jul 2 07:48:42.653595 kernel: audit: type=1334 audit(1719906521.749:93): prog-id=5 op=UNLOAD Jul 2 07:48:42.653609 kernel: audit: type=1131 audit(1719906521.751:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.653623 systemd[1]: Stopped initrd-switch-root.service. Jul 2 07:48:42.653641 kernel: audit: type=1334 audit(1719906521.819:95): prog-id=12 op=UNLOAD Jul 2 07:48:42.653655 kernel: audit: type=1130 audit(1719906521.834:96): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.653670 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jul 2 07:48:42.653686 kernel: audit: type=1131 audit(1719906521.834:97): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.653702 systemd[1]: Created slice system-addon\x2dconfig.slice. Jul 2 07:48:42.653718 systemd[1]: Created slice system-addon\x2drun.slice. Jul 2 07:48:42.653735 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Jul 2 07:48:42.653750 systemd[1]: Created slice system-getty.slice. Jul 2 07:48:42.653797 systemd[1]: Created slice system-modprobe.slice. Jul 2 07:48:42.653814 systemd[1]: Created slice system-serial\x2dgetty.slice. Jul 2 07:48:42.653829 systemd[1]: Created slice system-system\x2dcloudinit.slice. Jul 2 07:48:42.653844 systemd[1]: Created slice system-systemd\x2dfsck.slice. Jul 2 07:48:42.653860 systemd[1]: Created slice user.slice. Jul 2 07:48:42.653874 systemd[1]: Started systemd-ask-password-console.path. Jul 2 07:48:42.653889 systemd[1]: Started systemd-ask-password-wall.path. Jul 2 07:48:42.653904 systemd[1]: Set up automount boot.automount. Jul 2 07:48:42.653922 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Jul 2 07:48:42.653937 systemd[1]: Stopped target initrd-switch-root.target. Jul 2 07:48:42.653952 systemd[1]: Stopped target initrd-fs.target. Jul 2 07:48:42.653968 systemd[1]: Stopped target initrd-root-fs.target. Jul 2 07:48:42.653986 systemd[1]: Reached target integritysetup.target. Jul 2 07:48:42.654000 systemd[1]: Reached target remote-cryptsetup.target. Jul 2 07:48:42.654017 systemd[1]: Reached target remote-fs.target. Jul 2 07:48:42.654032 systemd[1]: Reached target slices.target. Jul 2 07:48:42.654047 systemd[1]: Reached target swap.target. Jul 2 07:48:42.654064 systemd[1]: Reached target torcx.target. Jul 2 07:48:42.654079 systemd[1]: Reached target veritysetup.target. Jul 2 07:48:42.654094 systemd[1]: Listening on systemd-coredump.socket. Jul 2 07:48:42.654109 systemd[1]: Listening on systemd-initctl.socket. Jul 2 07:48:42.654123 systemd[1]: Listening on systemd-networkd.socket. Jul 2 07:48:42.654138 systemd[1]: Listening on systemd-udevd-control.socket. Jul 2 07:48:42.654154 systemd[1]: Listening on systemd-udevd-kernel.socket. Jul 2 07:48:42.654169 systemd[1]: Listening on systemd-userdbd.socket. Jul 2 07:48:42.654183 systemd[1]: Mounting dev-hugepages.mount... Jul 2 07:48:42.654198 systemd[1]: Mounting dev-mqueue.mount... Jul 2 07:48:42.654217 systemd[1]: Mounting media.mount... Jul 2 07:48:42.654231 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:42.654246 systemd[1]: Mounting sys-kernel-debug.mount... Jul 2 07:48:42.654261 systemd[1]: Mounting sys-kernel-tracing.mount... Jul 2 07:48:42.654281 systemd[1]: Mounting tmp.mount... Jul 2 07:48:42.654295 systemd[1]: Starting flatcar-tmpfiles.service... Jul 2 07:48:42.654311 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Jul 2 07:48:42.654326 systemd[1]: Starting kmod-static-nodes.service... Jul 2 07:48:42.654340 systemd[1]: Starting modprobe@configfs.service... Jul 2 07:48:42.654358 systemd[1]: Starting modprobe@dm_mod.service... Jul 2 07:48:42.654373 systemd[1]: Starting modprobe@drm.service... Jul 2 07:48:42.654388 systemd[1]: Starting modprobe@efi_pstore.service... Jul 2 07:48:42.654402 systemd[1]: Starting modprobe@fuse.service... Jul 2 07:48:42.654417 systemd[1]: Starting modprobe@loop.service... Jul 2 07:48:42.654432 kernel: fuse: init (API version 7.34) Jul 2 07:48:42.654446 kernel: loop: module loaded Jul 2 07:48:42.654460 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jul 2 07:48:42.654477 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jul 2 07:48:42.654496 systemd[1]: Stopped systemd-fsck-root.service. Jul 2 07:48:42.654512 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jul 2 07:48:42.654526 systemd[1]: Stopped systemd-fsck-usr.service. Jul 2 07:48:42.654541 systemd[1]: Stopped systemd-journald.service. Jul 2 07:48:42.654559 systemd[1]: Starting systemd-journald.service... Jul 2 07:48:42.654573 systemd[1]: Starting systemd-modules-load.service... Jul 2 07:48:42.654588 systemd[1]: Starting systemd-network-generator.service... Jul 2 07:48:42.654608 systemd-journald[992]: Journal started Jul 2 07:48:42.654673 systemd-journald[992]: Runtime Journal (/run/log/journal/2e199e2f4ec92ce0f0f20c9eac44439b) is 8.0M, max 148.8M, 140.8M free. Jul 2 07:48:38.363000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Jul 2 07:48:38.513000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Jul 2 07:48:38.513000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Jul 2 07:48:38.513000 audit: BPF prog-id=10 op=LOAD Jul 2 07:48:38.513000 audit: BPF prog-id=10 op=UNLOAD Jul 2 07:48:38.513000 audit: BPF prog-id=11 op=LOAD Jul 2 07:48:38.513000 audit: BPF prog-id=11 op=UNLOAD Jul 2 07:48:38.668000 audit[901]: AVC avc: denied { associate } for pid=901 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Jul 2 07:48:38.668000 audit[901]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c0001178e2 a1=c00002ae40 a2=c000029100 a3=32 items=0 ppid=884 pid=901 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Jul 2 07:48:38.668000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Jul 2 07:48:38.679000 audit[901]: AVC avc: denied { associate } for pid=901 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Jul 2 07:48:38.679000 audit[901]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c0001179b9 a2=1ed a3=0 items=2 ppid=884 pid=901 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Jul 2 07:48:38.679000 audit: CWD cwd="/" Jul 2 07:48:38.679000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:38.679000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:38.679000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Jul 2 07:48:41.736000 audit: BPF prog-id=12 op=LOAD Jul 2 07:48:41.736000 audit: BPF prog-id=3 op=UNLOAD Jul 2 07:48:41.742000 audit: BPF prog-id=13 op=LOAD Jul 2 07:48:41.749000 audit: BPF prog-id=14 op=LOAD Jul 2 07:48:41.749000 audit: BPF prog-id=4 op=UNLOAD Jul 2 07:48:41.749000 audit: BPF prog-id=5 op=UNLOAD Jul 2 07:48:41.751000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:41.819000 audit: BPF prog-id=12 op=UNLOAD Jul 2 07:48:41.834000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:41.834000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.570000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.591000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.605000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.605000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.606000 audit: BPF prog-id=15 op=LOAD Jul 2 07:48:42.606000 audit: BPF prog-id=16 op=LOAD Jul 2 07:48:42.606000 audit: BPF prog-id=17 op=LOAD Jul 2 07:48:42.606000 audit: BPF prog-id=13 op=UNLOAD Jul 2 07:48:42.606000 audit: BPF prog-id=14 op=UNLOAD Jul 2 07:48:42.647000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Jul 2 07:48:42.647000 audit[992]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=4 a1=7fffa10247c0 a2=4000 a3=7fffa102485c items=0 ppid=1 pid=992 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Jul 2 07:48:42.647000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Jul 2 07:48:38.665564 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.5 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.5 /var/lib/torcx/store]" Jul 2 07:48:41.736088 systemd[1]: Queued start job for default target multi-user.target. Jul 2 07:48:38.666482 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Jul 2 07:48:41.752395 systemd[1]: systemd-journald.service: Deactivated successfully. Jul 2 07:48:38.666516 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Jul 2 07:48:38.666575 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Jul 2 07:48:38.666596 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="skipped missing lower profile" missing profile=oem Jul 2 07:48:38.666651 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Jul 2 07:48:38.666675 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Jul 2 07:48:38.667014 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Jul 2 07:48:38.667080 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Jul 2 07:48:38.667105 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Jul 2 07:48:38.669382 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Jul 2 07:48:38.669429 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Jul 2 07:48:38.669452 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.5: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.5 Jul 2 07:48:38.669470 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Jul 2 07:48:38.669491 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.5: no such file or directory" path=/var/lib/torcx/store/3510.3.5 Jul 2 07:48:38.669507 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:38Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Jul 2 07:48:41.134593 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:41Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Jul 2 07:48:41.134956 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:41Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Jul 2 07:48:41.135130 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:41Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Jul 2 07:48:41.135394 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:41Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Jul 2 07:48:41.135458 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:41Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Jul 2 07:48:41.135538 /usr/lib/systemd/system-generators/torcx-generator[901]: time="2024-07-02T07:48:41Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Jul 2 07:48:42.662804 systemd[1]: Starting systemd-remount-fs.service... Jul 2 07:48:42.677808 systemd[1]: Starting systemd-udev-trigger.service... Jul 2 07:48:42.691782 systemd[1]: verity-setup.service: Deactivated successfully. Jul 2 07:48:42.697795 systemd[1]: Stopped verity-setup.service. Jul 2 07:48:42.702000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.716959 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:42.725801 systemd[1]: Started systemd-journald.service. Jul 2 07:48:42.734114 systemd[1]: Mounted dev-hugepages.mount. Jul 2 07:48:42.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.741050 systemd[1]: Mounted dev-mqueue.mount. Jul 2 07:48:42.748029 systemd[1]: Mounted media.mount. Jul 2 07:48:42.755029 systemd[1]: Mounted sys-kernel-debug.mount. Jul 2 07:48:42.763110 systemd[1]: Mounted sys-kernel-tracing.mount. Jul 2 07:48:42.771983 systemd[1]: Mounted tmp.mount. Jul 2 07:48:42.779185 systemd[1]: Finished flatcar-tmpfiles.service. Jul 2 07:48:42.786000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.788209 systemd[1]: Finished kmod-static-nodes.service. Jul 2 07:48:42.795000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.797214 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jul 2 07:48:42.797419 systemd[1]: Finished modprobe@configfs.service. Jul 2 07:48:42.804000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.804000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.806215 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 2 07:48:42.806428 systemd[1]: Finished modprobe@dm_mod.service. Jul 2 07:48:42.813000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.813000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.815210 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 2 07:48:42.815408 systemd[1]: Finished modprobe@drm.service. Jul 2 07:48:42.822000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.822000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.824202 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 2 07:48:42.824407 systemd[1]: Finished modprobe@efi_pstore.service. Jul 2 07:48:42.831000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.831000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.833203 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jul 2 07:48:42.833402 systemd[1]: Finished modprobe@fuse.service. Jul 2 07:48:42.840000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.840000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.842202 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 2 07:48:42.842408 systemd[1]: Finished modprobe@loop.service. Jul 2 07:48:42.849000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.849000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.851208 systemd[1]: Finished systemd-modules-load.service. Jul 2 07:48:42.858000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.860162 systemd[1]: Finished systemd-network-generator.service. Jul 2 07:48:42.867000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.869173 systemd[1]: Finished systemd-remount-fs.service. Jul 2 07:48:42.876000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.878254 systemd[1]: Finished systemd-udev-trigger.service. Jul 2 07:48:42.885000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.887594 systemd[1]: Reached target network-pre.target. Jul 2 07:48:42.897484 systemd[1]: Mounting sys-fs-fuse-connections.mount... Jul 2 07:48:42.907487 systemd[1]: Mounting sys-kernel-config.mount... Jul 2 07:48:42.914901 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jul 2 07:48:42.919279 systemd[1]: Starting systemd-hwdb-update.service... Jul 2 07:48:42.927591 systemd[1]: Starting systemd-journal-flush.service... Jul 2 07:48:42.936027 systemd-journald[992]: Time spent on flushing to /var/log/journal/2e199e2f4ec92ce0f0f20c9eac44439b is 59.813ms for 1146 entries. Jul 2 07:48:42.936027 systemd-journald[992]: System Journal (/var/log/journal/2e199e2f4ec92ce0f0f20c9eac44439b) is 8.0M, max 584.8M, 576.8M free. Jul 2 07:48:43.024780 systemd-journald[992]: Received client request to flush runtime journal. Jul 2 07:48:43.005000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.014000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:42.935919 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 2 07:48:42.937415 systemd[1]: Starting systemd-random-seed.service... Jul 2 07:48:42.950936 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Jul 2 07:48:42.952605 systemd[1]: Starting systemd-sysctl.service... Jul 2 07:48:42.962005 systemd[1]: Starting systemd-sysusers.service... Jul 2 07:48:42.970538 systemd[1]: Starting systemd-udev-settle.service... Jul 2 07:48:42.981245 systemd[1]: Mounted sys-fs-fuse-connections.mount. Jul 2 07:48:42.990024 systemd[1]: Mounted sys-kernel-config.mount. Jul 2 07:48:42.998218 systemd[1]: Finished systemd-random-seed.service. Jul 2 07:48:43.007311 systemd[1]: Finished systemd-sysctl.service. Jul 2 07:48:43.019321 systemd[1]: Reached target first-boot-complete.target. Jul 2 07:48:43.029639 systemd[1]: Finished systemd-journal-flush.service. Jul 2 07:48:43.036000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.041013 udevadm[1006]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jul 2 07:48:43.046042 systemd[1]: Finished systemd-sysusers.service. Jul 2 07:48:43.053000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.615951 systemd[1]: Finished systemd-hwdb-update.service. Jul 2 07:48:43.622000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.622000 audit: BPF prog-id=18 op=LOAD Jul 2 07:48:43.622000 audit: BPF prog-id=19 op=LOAD Jul 2 07:48:43.622000 audit: BPF prog-id=7 op=UNLOAD Jul 2 07:48:43.623000 audit: BPF prog-id=8 op=UNLOAD Jul 2 07:48:43.625639 systemd[1]: Starting systemd-udevd.service... Jul 2 07:48:43.647625 systemd-udevd[1009]: Using default interface naming scheme 'v252'. Jul 2 07:48:43.690397 systemd[1]: Started systemd-udevd.service. Jul 2 07:48:43.697000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.699000 audit: BPF prog-id=20 op=LOAD Jul 2 07:48:43.702118 systemd[1]: Starting systemd-networkd.service... Jul 2 07:48:43.714000 audit: BPF prog-id=21 op=LOAD Jul 2 07:48:43.714000 audit: BPF prog-id=22 op=LOAD Jul 2 07:48:43.714000 audit: BPF prog-id=23 op=LOAD Jul 2 07:48:43.717096 systemd[1]: Starting systemd-userdbd.service... Jul 2 07:48:43.769273 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Jul 2 07:48:43.781905 systemd[1]: Started systemd-userdbd.service. Jul 2 07:48:43.788000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.875786 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Jul 2 07:48:43.920068 systemd-networkd[1022]: lo: Link UP Jul 2 07:48:43.920082 systemd-networkd[1022]: lo: Gained carrier Jul 2 07:48:43.920821 systemd-networkd[1022]: Enumeration completed Jul 2 07:48:43.920954 systemd[1]: Started systemd-networkd.service. Jul 2 07:48:43.921209 systemd-networkd[1022]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jul 2 07:48:43.923204 systemd-networkd[1022]: eth0: Link UP Jul 2 07:48:43.923219 systemd-networkd[1022]: eth0: Gained carrier Jul 2 07:48:43.927000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:43.933941 systemd-networkd[1022]: eth0: DHCPv4 address 10.128.0.12/32, gateway 10.128.0.1 acquired from 169.254.169.254 Jul 2 07:48:43.940000 audit[1017]: AVC avc: denied { confidentiality } for pid=1017 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Jul 2 07:48:43.975791 kernel: BTRFS info: devid 1 device path /dev/disk/by-label/OEM changed to /dev/sda6 scanned by (udev-worker) (1028) Jul 2 07:48:43.984830 kernel: ACPI: button: Power Button [PWRF] Jul 2 07:48:44.000933 kernel: EDAC MC: Ver: 3.0.0 Jul 2 07:48:44.001030 kernel: input: Sleep Button as /devices/LNXSYSTM:00/LNXSLPBN:00/input/input3 Jul 2 07:48:44.019806 kernel: ACPI: button: Sleep Button [SLPF] Jul 2 07:48:43.940000 audit[1017]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=55f78f932d10 a1=3207c a2=7fd13359bbc5 a3=5 items=108 ppid=1009 pid=1017 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Jul 2 07:48:43.940000 audit: CWD cwd="/" Jul 2 07:48:43.940000 audit: PATH item=0 name=(null) inode=1041 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=1 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=2 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=3 name=(null) inode=14552 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=4 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=5 name=(null) inode=14553 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=6 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=7 name=(null) inode=14554 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=8 name=(null) inode=14554 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=9 name=(null) inode=14555 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=10 name=(null) inode=14554 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=11 name=(null) inode=14556 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=12 name=(null) inode=14554 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=13 name=(null) inode=14557 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=14 name=(null) inode=14554 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=15 name=(null) inode=14558 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=16 name=(null) inode=14554 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=17 name=(null) inode=14559 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=18 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=19 name=(null) inode=14560 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=20 name=(null) inode=14560 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=21 name=(null) inode=14561 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=22 name=(null) inode=14560 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=23 name=(null) inode=14562 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=24 name=(null) inode=14560 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=25 name=(null) inode=14563 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=26 name=(null) inode=14560 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=27 name=(null) inode=14564 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=28 name=(null) inode=14560 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=29 name=(null) inode=14565 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=30 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=31 name=(null) inode=14566 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=32 name=(null) inode=14566 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=33 name=(null) inode=14567 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=34 name=(null) inode=14566 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=35 name=(null) inode=14568 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=36 name=(null) inode=14566 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=37 name=(null) inode=14569 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=38 name=(null) inode=14566 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=39 name=(null) inode=14570 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=40 name=(null) inode=14566 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=41 name=(null) inode=14571 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=42 name=(null) inode=14551 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=43 name=(null) inode=14572 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=44 name=(null) inode=14572 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=45 name=(null) inode=14573 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=46 name=(null) inode=14572 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=47 name=(null) inode=14574 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=48 name=(null) inode=14572 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=49 name=(null) inode=14575 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=50 name=(null) inode=14572 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=51 name=(null) inode=14576 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=52 name=(null) inode=14572 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=53 name=(null) inode=14577 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=54 name=(null) inode=1041 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=55 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=56 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=57 name=(null) inode=14579 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=58 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=59 name=(null) inode=14580 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=60 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=61 name=(null) inode=14581 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=62 name=(null) inode=14581 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=63 name=(null) inode=14582 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=64 name=(null) inode=14581 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=65 name=(null) inode=14583 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=66 name=(null) inode=14581 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=67 name=(null) inode=14584 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=68 name=(null) inode=14581 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=69 name=(null) inode=14585 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=70 name=(null) inode=14581 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=71 name=(null) inode=14586 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=72 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=73 name=(null) inode=14587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=74 name=(null) inode=14587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=75 name=(null) inode=14588 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=76 name=(null) inode=14587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=77 name=(null) inode=14589 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=78 name=(null) inode=14587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=79 name=(null) inode=14590 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=80 name=(null) inode=14587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=81 name=(null) inode=14591 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=82 name=(null) inode=14587 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=83 name=(null) inode=14592 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=84 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=85 name=(null) inode=14593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=86 name=(null) inode=14593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=87 name=(null) inode=14594 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=88 name=(null) inode=14593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=89 name=(null) inode=14595 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=90 name=(null) inode=14593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=91 name=(null) inode=14596 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=92 name=(null) inode=14593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=93 name=(null) inode=14597 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=94 name=(null) inode=14593 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=95 name=(null) inode=14598 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=96 name=(null) inode=14578 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=97 name=(null) inode=14599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=98 name=(null) inode=14599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=99 name=(null) inode=14600 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=100 name=(null) inode=14599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=101 name=(null) inode=14601 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=102 name=(null) inode=14599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=103 name=(null) inode=14602 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=104 name=(null) inode=14599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=105 name=(null) inode=14603 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=106 name=(null) inode=14599 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PATH item=107 name=(null) inode=14604 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Jul 2 07:48:43.940000 audit: PROCTITLE proctitle="(udev-worker)" Jul 2 07:48:44.066800 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input4 Jul 2 07:48:44.077077 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Jul 2 07:48:44.088794 kernel: piix4_smbus 0000:00:01.3: SMBus base address uninitialized - upgrade BIOS or use force_addr=0xaddr Jul 2 07:48:44.100788 kernel: mousedev: PS/2 mouse device common for all mice Jul 2 07:48:44.118521 systemd[1]: Finished systemd-udev-settle.service. Jul 2 07:48:44.125000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.128359 systemd[1]: Starting lvm2-activation-early.service... Jul 2 07:48:44.155449 lvm[1046]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 2 07:48:44.184041 systemd[1]: Finished lvm2-activation-early.service. Jul 2 07:48:44.191000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.193104 systemd[1]: Reached target cryptsetup.target. Jul 2 07:48:44.203331 systemd[1]: Starting lvm2-activation.service... Jul 2 07:48:44.209652 lvm[1047]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jul 2 07:48:44.238054 systemd[1]: Finished lvm2-activation.service. Jul 2 07:48:44.245000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.247100 systemd[1]: Reached target local-fs-pre.target. Jul 2 07:48:44.255883 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jul 2 07:48:44.255960 systemd[1]: Reached target local-fs.target. Jul 2 07:48:44.263889 systemd[1]: Reached target machines.target. Jul 2 07:48:44.273432 systemd[1]: Starting ldconfig.service... Jul 2 07:48:44.281750 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Jul 2 07:48:44.281861 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:44.283534 systemd[1]: Starting systemd-boot-update.service... Jul 2 07:48:44.292491 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Jul 2 07:48:44.303379 systemd[1]: Starting systemd-machine-id-commit.service... Jul 2 07:48:44.305232 systemd[1]: Starting systemd-sysext.service... Jul 2 07:48:44.307238 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1049 (bootctl) Jul 2 07:48:44.309381 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Jul 2 07:48:44.329000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.331216 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Jul 2 07:48:44.334582 systemd[1]: Unmounting usr-share-oem.mount... Jul 2 07:48:44.342030 systemd[1]: usr-share-oem.mount: Deactivated successfully. Jul 2 07:48:44.342323 systemd[1]: Unmounted usr-share-oem.mount. Jul 2 07:48:44.370828 kernel: loop0: detected capacity change from 0 to 211296 Jul 2 07:48:44.451508 systemd-fsck[1059]: fsck.fat 4.2 (2021-01-31) Jul 2 07:48:44.451508 systemd-fsck[1059]: /dev/sda1: 789 files, 119238/258078 clusters Jul 2 07:48:44.454276 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Jul 2 07:48:44.462000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.466175 systemd[1]: Mounting boot.mount... Jul 2 07:48:44.516041 systemd[1]: Mounted boot.mount. Jul 2 07:48:44.538017 systemd[1]: Finished systemd-boot-update.service. Jul 2 07:48:44.545000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.840307 ldconfig[1048]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jul 2 07:48:44.853441 systemd[1]: Finished ldconfig.service. Jul 2 07:48:44.859000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.862049 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jul 2 07:48:44.862852 systemd[1]: Finished systemd-machine-id-commit.service. Jul 2 07:48:44.869000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:44.881783 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jul 2 07:48:44.907801 kernel: loop1: detected capacity change from 0 to 211296 Jul 2 07:48:44.930491 (sd-sysext)[1064]: Using extensions 'kubernetes'. Jul 2 07:48:44.931191 (sd-sysext)[1064]: Merged extensions into '/usr'. Jul 2 07:48:44.952311 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:44.954178 systemd[1]: Mounting usr-share-oem.mount... Jul 2 07:48:44.961096 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Jul 2 07:48:44.962822 systemd[1]: Starting modprobe@dm_mod.service... Jul 2 07:48:44.971468 systemd[1]: Starting modprobe@efi_pstore.service... Jul 2 07:48:44.980636 systemd[1]: Starting modprobe@loop.service... Jul 2 07:48:44.988005 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Jul 2 07:48:44.988244 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:44.988475 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:44.993020 systemd[1]: Mounted usr-share-oem.mount. Jul 2 07:48:45.000435 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 2 07:48:45.000665 systemd[1]: Finished modprobe@dm_mod.service. Jul 2 07:48:45.007000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.007000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.009467 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 2 07:48:45.009662 systemd[1]: Finished modprobe@efi_pstore.service. Jul 2 07:48:45.016000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.016000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.018431 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 2 07:48:45.018621 systemd[1]: Finished modprobe@loop.service. Jul 2 07:48:45.025000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.025000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.027516 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 2 07:48:45.027706 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.029133 systemd[1]: Finished systemd-sysext.service. Jul 2 07:48:45.036000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.039392 systemd[1]: Starting ensure-sysext.service... Jul 2 07:48:45.048355 systemd[1]: Starting systemd-tmpfiles-setup.service... Jul 2 07:48:45.060498 systemd[1]: Reloading. Jul 2 07:48:45.079031 systemd-tmpfiles[1071]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Jul 2 07:48:45.085892 systemd-tmpfiles[1071]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jul 2 07:48:45.095488 systemd-tmpfiles[1071]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jul 2 07:48:45.163425 /usr/lib/systemd/system-generators/torcx-generator[1091]: time="2024-07-02T07:48:45Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.5 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.5 /var/lib/torcx/store]" Jul 2 07:48:45.163471 /usr/lib/systemd/system-generators/torcx-generator[1091]: time="2024-07-02T07:48:45Z" level=info msg="torcx already run" Jul 2 07:48:45.311902 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Jul 2 07:48:45.311934 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Jul 2 07:48:45.352578 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 2 07:48:45.430000 audit: BPF prog-id=24 op=LOAD Jul 2 07:48:45.430000 audit: BPF prog-id=15 op=UNLOAD Jul 2 07:48:45.430000 audit: BPF prog-id=25 op=LOAD Jul 2 07:48:45.430000 audit: BPF prog-id=26 op=LOAD Jul 2 07:48:45.430000 audit: BPF prog-id=16 op=UNLOAD Jul 2 07:48:45.430000 audit: BPF prog-id=17 op=UNLOAD Jul 2 07:48:45.431000 audit: BPF prog-id=27 op=LOAD Jul 2 07:48:45.431000 audit: BPF prog-id=20 op=UNLOAD Jul 2 07:48:45.432000 audit: BPF prog-id=28 op=LOAD Jul 2 07:48:45.432000 audit: BPF prog-id=21 op=UNLOAD Jul 2 07:48:45.432000 audit: BPF prog-id=29 op=LOAD Jul 2 07:48:45.432000 audit: BPF prog-id=30 op=LOAD Jul 2 07:48:45.432000 audit: BPF prog-id=22 op=UNLOAD Jul 2 07:48:45.432000 audit: BPF prog-id=23 op=UNLOAD Jul 2 07:48:45.433000 audit: BPF prog-id=31 op=LOAD Jul 2 07:48:45.433000 audit: BPF prog-id=32 op=LOAD Jul 2 07:48:45.433000 audit: BPF prog-id=18 op=UNLOAD Jul 2 07:48:45.433000 audit: BPF prog-id=19 op=UNLOAD Jul 2 07:48:45.444527 systemd[1]: Finished systemd-tmpfiles-setup.service. Jul 2 07:48:45.451000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.458938 systemd[1]: Starting audit-rules.service... Jul 2 07:48:45.468815 systemd[1]: Starting clean-ca-certificates.service... Jul 2 07:48:45.479025 systemd[1]: Starting oem-gce-enable-oslogin.service... Jul 2 07:48:45.490359 systemd[1]: Starting systemd-journal-catalog-update.service... Jul 2 07:48:45.498000 audit: BPF prog-id=33 op=LOAD Jul 2 07:48:45.501834 systemd[1]: Starting systemd-resolved.service... Jul 2 07:48:45.508000 audit: BPF prog-id=34 op=LOAD Jul 2 07:48:45.511730 systemd[1]: Starting systemd-timesyncd.service... Jul 2 07:48:45.521473 systemd[1]: Starting systemd-update-utmp.service... Jul 2 07:48:45.528000 audit[1162]: SYSTEM_BOOT pid=1162 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.530919 systemd[1]: Finished clean-ca-certificates.service. Jul 2 07:48:45.538000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Jul 2 07:48:45.540412 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Jul 2 07:48:45.540572 systemd[1]: Finished oem-gce-enable-oslogin.service. Jul 2 07:48:45.540000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Jul 2 07:48:45.540000 audit[1165]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe62435b70 a2=420 a3=0 items=0 ppid=1135 pid=1165 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Jul 2 07:48:45.542152 augenrules[1165]: No rules Jul 2 07:48:45.540000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Jul 2 07:48:45.549529 systemd[1]: Finished audit-rules.service. Jul 2 07:48:45.557490 systemd[1]: Finished systemd-journal-catalog-update.service. Jul 2 07:48:45.572991 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:45.573545 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.576120 systemd[1]: Starting modprobe@dm_mod.service... Jul 2 07:48:45.576873 systemd-networkd[1022]: eth0: Gained IPv6LL Jul 2 07:48:45.585974 systemd[1]: Starting modprobe@efi_pstore.service... Jul 2 07:48:45.594892 systemd[1]: Starting modprobe@loop.service... Jul 2 07:48:45.603510 systemd[1]: Starting oem-gce-enable-oslogin.service... Jul 2 07:48:45.608946 enable-oslogin[1173]: /etc/pam.d/sshd already exists. Not enabling OS Login Jul 2 07:48:45.611971 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.612306 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:45.614515 systemd[1]: Starting systemd-update-done.service... Jul 2 07:48:45.621874 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 2 07:48:45.622205 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:45.625927 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 2 07:48:45.626164 systemd[1]: Finished modprobe@dm_mod.service. Jul 2 07:48:45.636589 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 2 07:48:45.636803 systemd[1]: Finished modprobe@efi_pstore.service. Jul 2 07:48:45.645619 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 2 07:48:45.645886 systemd[1]: Finished modprobe@loop.service. Jul 2 07:48:45.654573 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Jul 2 07:48:45.654807 systemd[1]: Finished oem-gce-enable-oslogin.service. Jul 2 07:48:45.665387 systemd[1]: Finished systemd-update-done.service. Jul 2 07:48:45.676408 systemd[1]: Finished systemd-update-utmp.service. Jul 2 07:48:45.687492 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:45.688053 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.690391 systemd[1]: Starting modprobe@dm_mod.service... Jul 2 07:48:45.699904 systemd[1]: Starting modprobe@efi_pstore.service... Jul 2 07:48:45.708818 systemd[1]: Starting modprobe@loop.service... Jul 2 07:48:45.717759 systemd[1]: Starting oem-gce-enable-oslogin.service... Jul 2 07:48:45.723176 enable-oslogin[1179]: /etc/pam.d/sshd already exists. Not enabling OS Login Jul 2 07:48:45.725939 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.726170 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:45.726362 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 2 07:48:45.726506 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:45.728607 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 2 07:48:45.728862 systemd[1]: Finished modprobe@dm_mod.service. Jul 2 07:48:45.737545 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 2 07:48:45.737754 systemd[1]: Finished modprobe@efi_pstore.service. Jul 2 07:48:45.746445 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 2 07:48:45.746637 systemd[1]: Finished modprobe@loop.service. Jul 2 07:48:45.753289 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Jul 2 07:48:45.753511 systemd[1]: Finished oem-gce-enable-oslogin.service. Jul 2 07:48:45.762554 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 2 07:48:45.762747 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.767565 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:45.768080 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.770638 systemd[1]: Starting modprobe@dm_mod.service... Jul 2 07:48:45.773873 systemd-resolved[1154]: Positive Trust Anchors: Jul 2 07:48:45.773890 systemd-resolved[1154]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jul 2 07:48:45.773951 systemd-resolved[1154]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Jul 2 07:48:45.775575 systemd-timesyncd[1160]: Contacted time server 169.254.169.254:123 (169.254.169.254). Jul 2 07:48:45.775660 systemd-timesyncd[1160]: Initial clock synchronization to Tue 2024-07-02 07:48:45.887454 UTC. Jul 2 07:48:45.779540 systemd[1]: Starting modprobe@drm.service... Jul 2 07:48:45.786244 systemd-resolved[1154]: Defaulting to hostname 'linux'. Jul 2 07:48:45.789115 systemd[1]: Starting modprobe@efi_pstore.service... Jul 2 07:48:45.797590 systemd[1]: Starting modprobe@loop.service... Jul 2 07:48:45.806497 systemd[1]: Starting oem-gce-enable-oslogin.service... Jul 2 07:48:45.810423 enable-oslogin[1185]: /etc/pam.d/sshd already exists. Not enabling OS Login Jul 2 07:48:45.814977 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Jul 2 07:48:45.815144 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:45.816850 systemd[1]: Starting systemd-networkd-wait-online.service... Jul 2 07:48:45.824965 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jul 2 07:48:45.825191 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Jul 2 07:48:45.826745 systemd[1]: Started systemd-timesyncd.service. Jul 2 07:48:45.835838 systemd[1]: Started systemd-resolved.service. Jul 2 07:48:45.844527 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jul 2 07:48:45.844740 systemd[1]: Finished modprobe@dm_mod.service. Jul 2 07:48:45.853399 systemd[1]: modprobe@drm.service: Deactivated successfully. Jul 2 07:48:45.853600 systemd[1]: Finished modprobe@drm.service. Jul 2 07:48:45.862390 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jul 2 07:48:45.862595 systemd[1]: Finished modprobe@efi_pstore.service. Jul 2 07:48:45.871461 systemd[1]: modprobe@loop.service: Deactivated successfully. Jul 2 07:48:45.871687 systemd[1]: Finished modprobe@loop.service. Jul 2 07:48:45.880370 systemd[1]: oem-gce-enable-oslogin.service: Deactivated successfully. Jul 2 07:48:45.880592 systemd[1]: Finished oem-gce-enable-oslogin.service. Jul 2 07:48:45.889446 systemd[1]: Finished systemd-networkd-wait-online.service. Jul 2 07:48:45.899939 systemd[1]: Reached target network.target. Jul 2 07:48:45.907932 systemd[1]: Reached target network-online.target. Jul 2 07:48:45.916884 systemd[1]: Reached target nss-lookup.target. Jul 2 07:48:45.924876 systemd[1]: Reached target time-set.target. Jul 2 07:48:45.932910 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jul 2 07:48:45.932964 systemd[1]: Reached target sysinit.target. Jul 2 07:48:45.941014 systemd[1]: Started motdgen.path. Jul 2 07:48:45.947962 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Jul 2 07:48:45.958062 systemd[1]: Started logrotate.timer. Jul 2 07:48:45.965014 systemd[1]: Started mdadm.timer. Jul 2 07:48:45.971900 systemd[1]: Started systemd-tmpfiles-clean.timer. Jul 2 07:48:45.979881 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jul 2 07:48:45.979932 systemd[1]: Reached target paths.target. Jul 2 07:48:45.986890 systemd[1]: Reached target timers.target. Jul 2 07:48:45.994350 systemd[1]: Listening on dbus.socket. Jul 2 07:48:46.002293 systemd[1]: Starting docker.socket... Jul 2 07:48:46.013064 systemd[1]: Listening on sshd.socket. Jul 2 07:48:46.020081 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:46.020168 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Jul 2 07:48:46.021138 systemd[1]: Finished ensure-sysext.service. Jul 2 07:48:46.030118 systemd[1]: Listening on docker.socket. Jul 2 07:48:46.038000 systemd[1]: Reached target sockets.target. Jul 2 07:48:46.046883 systemd[1]: Reached target basic.target. Jul 2 07:48:46.053969 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Jul 2 07:48:46.054011 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Jul 2 07:48:46.055593 systemd[1]: Starting containerd.service... Jul 2 07:48:46.064254 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Jul 2 07:48:46.074693 systemd[1]: Starting dbus.service... Jul 2 07:48:46.084166 systemd[1]: Starting enable-oem-cloudinit.service... Jul 2 07:48:46.095098 systemd[1]: Starting extend-filesystems.service... Jul 2 07:48:46.104498 jq[1192]: false Jul 2 07:48:46.101952 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Jul 2 07:48:46.104204 systemd[1]: Starting kubelet.service... Jul 2 07:48:46.113202 systemd[1]: Starting motdgen.service... Jul 2 07:48:46.120030 systemd[1]: Starting oem-gce.service... Jul 2 07:48:46.127455 systemd[1]: Starting prepare-helm.service... Jul 2 07:48:46.137052 systemd[1]: Starting ssh-key-proc-cmdline.service... Jul 2 07:48:46.145977 systemd[1]: Starting sshd-keygen.service... Jul 2 07:48:46.157377 systemd[1]: Starting systemd-logind.service... Jul 2 07:48:46.164960 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Jul 2 07:48:46.165100 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionSecurity=!tpm2). Jul 2 07:48:46.165924 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jul 2 07:48:46.167295 systemd[1]: Starting update-engine.service... Jul 2 07:48:46.177237 systemd[1]: Starting update-ssh-keys-after-ignition.service... Jul 2 07:48:46.184865 jq[1214]: true Jul 2 07:48:46.189872 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jul 2 07:48:46.190236 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Jul 2 07:48:46.196406 systemd[1]: motdgen.service: Deactivated successfully. Jul 2 07:48:46.196698 systemd[1]: Finished motdgen.service. Jul 2 07:48:46.201573 extend-filesystems[1193]: Found loop1 Jul 2 07:48:46.201573 extend-filesystems[1193]: Found sda Jul 2 07:48:46.201573 extend-filesystems[1193]: Found sda1 Jul 2 07:48:46.201573 extend-filesystems[1193]: Found sda2 Jul 2 07:48:46.201573 extend-filesystems[1193]: Found sda3 Jul 2 07:48:46.201573 extend-filesystems[1193]: Found usr Jul 2 07:48:46.201573 extend-filesystems[1193]: Found sda4 Jul 2 07:48:46.201573 extend-filesystems[1193]: Found sda6 Jul 2 07:48:46.294235 extend-filesystems[1193]: Found sda7 Jul 2 07:48:46.294235 extend-filesystems[1193]: Found sda9 Jul 2 07:48:46.294235 extend-filesystems[1193]: Checking size of /dev/sda9 Jul 2 07:48:46.268552 dbus-daemon[1191]: [system] SELinux support is enabled Jul 2 07:48:46.207671 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jul 2 07:48:46.332190 extend-filesystems[1193]: Resized partition /dev/sda9 Jul 2 07:48:46.318631 dbus-daemon[1191]: [system] Activating systemd to hand-off: service name='org.freedesktop.hostname1' unit='dbus-org.freedesktop.hostname1.service' requested by ':1.0' (uid=244 pid=1022 comm="/usr/lib/systemd/systemd-networkd" label="system_u:system_r:kernel_t:s0") Jul 2 07:48:46.208046 systemd[1]: Finished ssh-key-proc-cmdline.service. Jul 2 07:48:46.337013 dbus-daemon[1191]: [system] Successfully activated service 'org.freedesktop.systemd1' Jul 2 07:48:46.274221 systemd[1]: Started dbus.service. Jul 2 07:48:46.289829 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jul 2 07:48:46.340662 mkfs.ext4[1226]: mke2fs 1.46.5 (30-Dec-2021) Jul 2 07:48:46.340662 mkfs.ext4[1226]: Discarding device blocks: 0/262144\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008 \u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008\u0008done Jul 2 07:48:46.340662 mkfs.ext4[1226]: Creating filesystem with 262144 4k blocks and 65536 inodes Jul 2 07:48:46.340662 mkfs.ext4[1226]: Filesystem UUID: a57670d2-5573-4f4f-a63d-f33097d6ce89 Jul 2 07:48:46.340662 mkfs.ext4[1226]: Superblock backups stored on blocks: Jul 2 07:48:46.340662 mkfs.ext4[1226]: 32768, 98304, 163840, 229376 Jul 2 07:48:46.340662 mkfs.ext4[1226]: Allocating group tables: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done Jul 2 07:48:46.340662 mkfs.ext4[1226]: Writing inode tables: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done Jul 2 07:48:46.340662 mkfs.ext4[1226]: Creating journal (8192 blocks): done Jul 2 07:48:46.340662 mkfs.ext4[1226]: Writing superblocks and filesystem accounting information: 0/8\u0008\u0008\u0008 \u0008\u0008\u0008done Jul 2 07:48:46.289969 systemd[1]: Reached target system-config.target. Jul 2 07:48:46.341343 jq[1222]: true Jul 2 07:48:46.303003 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jul 2 07:48:46.303047 systemd[1]: Reached target user-config.target. Jul 2 07:48:46.343899 systemd[1]: Starting systemd-hostnamed.service... Jul 2 07:48:46.367671 extend-filesystems[1236]: resize2fs 1.46.5 (30-Dec-2021) Jul 2 07:48:46.383979 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 2538491 blocks Jul 2 07:48:46.390492 tar[1220]: linux-amd64/helm Jul 2 07:48:46.390935 umount[1241]: umount: /var/lib/flatcar-oem-gce.img: not mounted. Jul 2 07:48:46.420451 update_engine[1212]: I0702 07:48:46.420298 1212 main.cc:92] Flatcar Update Engine starting Jul 2 07:48:46.426812 systemd[1]: Started update-engine.service. Jul 2 07:48:46.427679 update_engine[1212]: I0702 07:48:46.426856 1212 update_check_scheduler.cc:74] Next update check in 2m23s Jul 2 07:48:46.438588 systemd[1]: Started locksmithd.service. Jul 2 07:48:46.447141 kernel: loop2: detected capacity change from 0 to 2097152 Jul 2 07:48:46.465821 kernel: EXT4-fs (sda9): resized filesystem to 2538491 Jul 2 07:48:46.478196 extend-filesystems[1236]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Jul 2 07:48:46.478196 extend-filesystems[1236]: old_desc_blocks = 1, new_desc_blocks = 2 Jul 2 07:48:46.478196 extend-filesystems[1236]: The filesystem on /dev/sda9 is now 2538491 (4k) blocks long. Jul 2 07:48:46.554938 kernel: EXT4-fs (loop2): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.487 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/sshKeys: Attempt #1 Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.496 INFO Fetch failed with 404: resource not found Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.496 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/ssh-keys: Attempt #1 Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.499 INFO Fetch successful Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.499 INFO Fetching http://169.254.169.254/computeMetadata/v1/instance/attributes/block-project-ssh-keys: Attempt #1 Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.500 INFO Fetch failed with 404: resource not found Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.500 INFO Fetching http://169.254.169.254/computeMetadata/v1/project/attributes/sshKeys: Attempt #1 Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.501 INFO Fetch failed with 404: resource not found Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.501 INFO Fetching http://169.254.169.254/computeMetadata/v1/project/attributes/ssh-keys: Attempt #1 Jul 2 07:48:46.578001 coreos-metadata[1190]: Jul 02 07:48:46.502 INFO Fetch successful Jul 2 07:48:46.578594 extend-filesystems[1193]: Resized filesystem in /dev/sda9 Jul 2 07:48:46.479594 systemd[1]: extend-filesystems.service: Deactivated successfully. Jul 2 07:48:46.587250 bash[1255]: Updated "/home/core/.ssh/authorized_keys" Jul 2 07:48:46.587507 update-ssh-keys[1260]: Updated "/home/core/.ssh/authorized_keys" Jul 2 07:48:46.479960 systemd[1]: Finished extend-filesystems.service. Jul 2 07:48:46.504057 unknown[1190]: wrote ssh authorized keys file for user: core Jul 2 07:48:46.505466 systemd[1]: Finished update-ssh-keys-after-ignition.service. Jul 2 07:48:46.585306 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Jul 2 07:48:46.644752 env[1224]: time="2024-07-02T07:48:46.644571064Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Jul 2 07:48:46.655262 dbus-daemon[1191]: [system] Successfully activated service 'org.freedesktop.hostname1' Jul 2 07:48:46.656039 dbus-daemon[1191]: [system] Activating via systemd: service name='org.freedesktop.PolicyKit1' unit='polkit.service' requested by ':1.5' (uid=0 pid=1239 comm="/usr/lib/systemd/systemd-hostnamed" label="system_u:system_r:kernel_t:s0") Jul 2 07:48:46.657106 systemd[1]: Started systemd-hostnamed.service. Jul 2 07:48:46.670200 systemd[1]: Starting polkit.service... Jul 2 07:48:46.718998 systemd-logind[1211]: Watching system buttons on /dev/input/event1 (Power Button) Jul 2 07:48:46.719041 systemd-logind[1211]: Watching system buttons on /dev/input/event2 (Sleep Button) Jul 2 07:48:46.719073 systemd-logind[1211]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jul 2 07:48:46.726824 systemd-logind[1211]: New seat seat0. Jul 2 07:48:46.735865 env[1224]: time="2024-07-02T07:48:46.735821072Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jul 2 07:48:46.736177 env[1224]: time="2024-07-02T07:48:46.736143365Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jul 2 07:48:46.738830 env[1224]: time="2024-07-02T07:48:46.738758815Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.161-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jul 2 07:48:46.738928 env[1224]: time="2024-07-02T07:48:46.738827665Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jul 2 07:48:46.739170 env[1224]: time="2024-07-02T07:48:46.739123285Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 2 07:48:46.739170 env[1224]: time="2024-07-02T07:48:46.739152975Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jul 2 07:48:46.739297 env[1224]: time="2024-07-02T07:48:46.739175453Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Jul 2 07:48:46.739297 env[1224]: time="2024-07-02T07:48:46.739192770Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jul 2 07:48:46.739403 env[1224]: time="2024-07-02T07:48:46.739312109Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jul 2 07:48:46.739651 env[1224]: time="2024-07-02T07:48:46.739606197Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jul 2 07:48:46.740272 env[1224]: time="2024-07-02T07:48:46.740219814Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jul 2 07:48:46.740272 env[1224]: time="2024-07-02T07:48:46.740251103Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jul 2 07:48:46.740416 env[1224]: time="2024-07-02T07:48:46.740336312Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Jul 2 07:48:46.740416 env[1224]: time="2024-07-02T07:48:46.740360845Z" level=info msg="metadata content store policy set" policy=shared Jul 2 07:48:46.740916 systemd[1]: Started systemd-logind.service. Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752380202Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752425693Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752457944Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752517839Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752541208Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752609474Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752632774Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752655932Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752676913Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752699148Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752719657Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.752802 env[1224]: time="2024-07-02T07:48:46.752740340Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jul 2 07:48:46.753385 env[1224]: time="2024-07-02T07:48:46.752898295Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jul 2 07:48:46.753385 env[1224]: time="2024-07-02T07:48:46.753019725Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753580388Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753628255Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753651910Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753725680Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753747971Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753853623Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753877801Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753898871Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753919772Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753950158Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753972186Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.753996622Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.754160824Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.754194078Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.754807 env[1224]: time="2024-07-02T07:48:46.754216083Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.755532 env[1224]: time="2024-07-02T07:48:46.754238163Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jul 2 07:48:46.755532 env[1224]: time="2024-07-02T07:48:46.754264480Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Jul 2 07:48:46.755532 env[1224]: time="2024-07-02T07:48:46.754283843Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jul 2 07:48:46.755532 env[1224]: time="2024-07-02T07:48:46.754311739Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Jul 2 07:48:46.755532 env[1224]: time="2024-07-02T07:48:46.754360813Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jul 2 07:48:46.755771 env[1224]: time="2024-07-02T07:48:46.754712302Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jul 2 07:48:46.759324 env[1224]: time="2024-07-02T07:48:46.756018452Z" level=info msg="Connect containerd service" Jul 2 07:48:46.759324 env[1224]: time="2024-07-02T07:48:46.756089493Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jul 2 07:48:46.781190 env[1224]: time="2024-07-02T07:48:46.781146792Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 2 07:48:46.781486 env[1224]: time="2024-07-02T07:48:46.781449209Z" level=info msg="Start subscribing containerd event" Jul 2 07:48:46.781634 env[1224]: time="2024-07-02T07:48:46.781614788Z" level=info msg="Start recovering state" Jul 2 07:48:46.781853 env[1224]: time="2024-07-02T07:48:46.781832186Z" level=info msg="Start event monitor" Jul 2 07:48:46.781964 env[1224]: time="2024-07-02T07:48:46.781946707Z" level=info msg="Start snapshots syncer" Jul 2 07:48:46.782077 env[1224]: time="2024-07-02T07:48:46.782060518Z" level=info msg="Start cni network conf syncer for default" Jul 2 07:48:46.782184 env[1224]: time="2024-07-02T07:48:46.782167779Z" level=info msg="Start streaming server" Jul 2 07:48:46.782896 env[1224]: time="2024-07-02T07:48:46.782870491Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jul 2 07:48:46.783155 env[1224]: time="2024-07-02T07:48:46.783135496Z" level=info msg=serving... address=/run/containerd/containerd.sock Jul 2 07:48:46.783432 systemd[1]: Started containerd.service. Jul 2 07:48:46.795262 polkitd[1267]: Started polkitd version 121 Jul 2 07:48:46.833980 polkitd[1267]: Loading rules from directory /etc/polkit-1/rules.d Jul 2 07:48:46.858894 polkitd[1267]: Loading rules from directory /usr/share/polkit-1/rules.d Jul 2 07:48:46.860749 env[1224]: time="2024-07-02T07:48:46.860708642Z" level=info msg="containerd successfully booted in 0.233512s" Jul 2 07:48:46.862606 polkitd[1267]: Finished loading, compiling and executing 2 rules Jul 2 07:48:46.863307 dbus-daemon[1191]: [system] Successfully activated service 'org.freedesktop.PolicyKit1' Jul 2 07:48:46.863557 systemd[1]: Started polkit.service. Jul 2 07:48:46.863959 polkitd[1267]: Acquired the name org.freedesktop.PolicyKit1 on the system bus Jul 2 07:48:46.881378 systemd-hostnamed[1239]: Hostname set to (transient) Jul 2 07:48:46.884242 systemd-resolved[1154]: System hostname changed to 'ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal'. Jul 2 07:48:48.182576 tar[1220]: linux-amd64/LICENSE Jul 2 07:48:48.182576 tar[1220]: linux-amd64/README.md Jul 2 07:48:48.197721 systemd[1]: Finished prepare-helm.service. Jul 2 07:48:48.284902 systemd[1]: Started kubelet.service. Jul 2 07:48:49.093452 locksmithd[1256]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jul 2 07:48:49.752202 kubelet[1290]: E0702 07:48:49.752113 1290 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 2 07:48:49.755226 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 2 07:48:49.755429 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 2 07:48:49.755804 systemd[1]: kubelet.service: Consumed 1.445s CPU time. Jul 2 07:48:50.869477 sshd_keygen[1223]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jul 2 07:48:50.922294 systemd[1]: Finished sshd-keygen.service. Jul 2 07:48:50.933426 systemd[1]: Starting issuegen.service... Jul 2 07:48:50.944398 systemd[1]: issuegen.service: Deactivated successfully. Jul 2 07:48:50.944644 systemd[1]: Finished issuegen.service. Jul 2 07:48:50.954044 systemd[1]: Starting systemd-user-sessions.service... Jul 2 07:48:50.964699 systemd[1]: Finished systemd-user-sessions.service. Jul 2 07:48:50.975147 systemd[1]: Started getty@tty1.service. Jul 2 07:48:50.984130 systemd[1]: Started serial-getty@ttyS0.service. Jul 2 07:48:50.993296 systemd[1]: Reached target getty.target. Jul 2 07:48:52.693301 systemd[1]: var-lib-flatcar\x2doem\x2dgce.mount: Deactivated successfully. Jul 2 07:48:54.669816 kernel: loop2: detected capacity change from 0 to 2097152 Jul 2 07:48:54.689140 systemd-nspawn[1317]: Spawning container oem-gce on /var/lib/flatcar-oem-gce.img. Jul 2 07:48:54.689140 systemd-nspawn[1317]: Press ^] three times within 1s to kill container. Jul 2 07:48:54.705813 kernel: EXT4-fs (loop2): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Jul 2 07:48:54.710627 systemd[1]: Created slice system-sshd.slice. Jul 2 07:48:54.720814 systemd[1]: Started sshd@0-10.128.0.12:22-147.75.109.163:43072.service. Jul 2 07:48:54.798837 systemd[1]: Started oem-gce.service. Jul 2 07:48:54.806388 systemd[1]: Reached target multi-user.target. Jul 2 07:48:54.817005 systemd[1]: Starting systemd-update-utmp-runlevel.service... Jul 2 07:48:54.830049 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Jul 2 07:48:54.830301 systemd[1]: Finished systemd-update-utmp-runlevel.service. Jul 2 07:48:54.840059 systemd[1]: Startup finished in 969ms (kernel) + 10.454s (initrd) + 16.598s (userspace) = 28.023s. Jul 2 07:48:54.865642 systemd-nspawn[1317]: + '[' -e /etc/default/instance_configs.cfg.template ']' Jul 2 07:48:54.865642 systemd-nspawn[1317]: + echo -e '[InstanceSetup]\nset_host_keys = false' Jul 2 07:48:54.866048 systemd-nspawn[1317]: + /usr/bin/google_instance_setup Jul 2 07:48:55.056958 sshd[1323]: Accepted publickey for core from 147.75.109.163 port 43072 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:48:55.060572 sshd[1323]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:48:55.078843 systemd[1]: Created slice user-500.slice. Jul 2 07:48:55.082997 systemd[1]: Starting user-runtime-dir@500.service... Jul 2 07:48:55.088638 systemd-logind[1211]: New session 1 of user core. Jul 2 07:48:55.098748 systemd[1]: Finished user-runtime-dir@500.service. Jul 2 07:48:55.101248 systemd[1]: Starting user@500.service... Jul 2 07:48:55.118187 (systemd)[1328]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:48:55.266348 systemd[1328]: Queued start job for default target default.target. Jul 2 07:48:55.267211 systemd[1328]: Reached target paths.target. Jul 2 07:48:55.267246 systemd[1328]: Reached target sockets.target. Jul 2 07:48:55.267269 systemd[1328]: Reached target timers.target. Jul 2 07:48:55.267290 systemd[1328]: Reached target basic.target. Jul 2 07:48:55.267363 systemd[1328]: Reached target default.target. Jul 2 07:48:55.267422 systemd[1328]: Startup finished in 138ms. Jul 2 07:48:55.267460 systemd[1]: Started user@500.service. Jul 2 07:48:55.268636 systemd[1]: Started session-1.scope. Jul 2 07:48:55.501044 systemd[1]: Started sshd@1-10.128.0.12:22-147.75.109.163:43074.service. Jul 2 07:48:55.565705 instance-setup[1326]: INFO Running google_set_multiqueue. Jul 2 07:48:55.583154 instance-setup[1326]: INFO Set channels for eth0 to 2. Jul 2 07:48:55.586433 instance-setup[1326]: INFO Setting /proc/irq/31/smp_affinity_list to 0 for device virtio1. Jul 2 07:48:55.588156 instance-setup[1326]: INFO /proc/irq/31/smp_affinity_list: real affinity 0 Jul 2 07:48:55.588716 instance-setup[1326]: INFO Setting /proc/irq/32/smp_affinity_list to 0 for device virtio1. Jul 2 07:48:55.590289 instance-setup[1326]: INFO /proc/irq/32/smp_affinity_list: real affinity 0 Jul 2 07:48:55.590819 instance-setup[1326]: INFO Setting /proc/irq/33/smp_affinity_list to 1 for device virtio1. Jul 2 07:48:55.592757 instance-setup[1326]: INFO /proc/irq/33/smp_affinity_list: real affinity 1 Jul 2 07:48:55.593356 instance-setup[1326]: INFO Setting /proc/irq/34/smp_affinity_list to 1 for device virtio1. Jul 2 07:48:55.595216 instance-setup[1326]: INFO /proc/irq/34/smp_affinity_list: real affinity 1 Jul 2 07:48:55.610719 instance-setup[1326]: INFO Queue 0 XPS=1 for /sys/class/net/eth0/queues/tx-0/xps_cpus Jul 2 07:48:55.610923 instance-setup[1326]: INFO Queue 1 XPS=2 for /sys/class/net/eth0/queues/tx-1/xps_cpus Jul 2 07:48:55.650125 systemd-nspawn[1317]: + /usr/bin/google_metadata_script_runner --script-type startup Jul 2 07:48:55.811841 sshd[1339]: Accepted publickey for core from 147.75.109.163 port 43074 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:48:55.813353 sshd[1339]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:48:55.822042 systemd[1]: Started session-2.scope. Jul 2 07:48:55.823841 systemd-logind[1211]: New session 2 of user core. Jul 2 07:48:55.996818 startup-script[1369]: INFO Starting startup scripts. Jul 2 07:48:56.009243 startup-script[1369]: INFO No startup scripts found in metadata. Jul 2 07:48:56.009407 startup-script[1369]: INFO Finished running startup scripts. Jul 2 07:48:56.031068 sshd[1339]: pam_unix(sshd:session): session closed for user core Jul 2 07:48:56.036586 systemd-logind[1211]: Session 2 logged out. Waiting for processes to exit. Jul 2 07:48:56.039206 systemd[1]: sshd@1-10.128.0.12:22-147.75.109.163:43074.service: Deactivated successfully. Jul 2 07:48:56.040305 systemd[1]: session-2.scope: Deactivated successfully. Jul 2 07:48:56.042483 systemd-logind[1211]: Removed session 2. Jul 2 07:48:56.044033 systemd-nspawn[1317]: + trap 'stopping=1 ; kill "${daemon_pids[@]}" || :' SIGTERM Jul 2 07:48:56.044507 systemd-nspawn[1317]: + daemon_pids=() Jul 2 07:48:56.044507 systemd-nspawn[1317]: + for d in accounts clock_skew network Jul 2 07:48:56.044507 systemd-nspawn[1317]: + daemon_pids+=($!) Jul 2 07:48:56.044678 systemd-nspawn[1317]: + for d in accounts clock_skew network Jul 2 07:48:56.044989 systemd-nspawn[1317]: + daemon_pids+=($!) Jul 2 07:48:56.045108 systemd-nspawn[1317]: + for d in accounts clock_skew network Jul 2 07:48:56.045278 systemd-nspawn[1317]: + /usr/bin/google_clock_skew_daemon Jul 2 07:48:56.045456 systemd-nspawn[1317]: + daemon_pids+=($!) Jul 2 07:48:56.045585 systemd-nspawn[1317]: + NOTIFY_SOCKET=/run/systemd/notify Jul 2 07:48:56.045655 systemd-nspawn[1317]: + /usr/bin/systemd-notify --ready Jul 2 07:48:56.046182 systemd-nspawn[1317]: + /usr/bin/google_network_daemon Jul 2 07:48:56.046508 systemd-nspawn[1317]: + /usr/bin/google_accounts_daemon Jul 2 07:48:56.077088 systemd[1]: Started sshd@2-10.128.0.12:22-147.75.109.163:43088.service. Jul 2 07:48:56.112865 systemd-nspawn[1317]: + wait -n 36 37 38 Jul 2 07:48:56.389343 sshd[1380]: Accepted publickey for core from 147.75.109.163 port 43088 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:48:56.390072 sshd[1380]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:48:56.398655 systemd[1]: Started session-3.scope. Jul 2 07:48:56.399555 systemd-logind[1211]: New session 3 of user core. Jul 2 07:48:56.589784 google-clock-skew[1376]: INFO Starting Google Clock Skew daemon. Jul 2 07:48:56.601694 sshd[1380]: pam_unix(sshd:session): session closed for user core Jul 2 07:48:56.605831 systemd[1]: sshd@2-10.128.0.12:22-147.75.109.163:43088.service: Deactivated successfully. Jul 2 07:48:56.606962 systemd[1]: session-3.scope: Deactivated successfully. Jul 2 07:48:56.608256 systemd-logind[1211]: Session 3 logged out. Waiting for processes to exit. Jul 2 07:48:56.609719 systemd-logind[1211]: Removed session 3. Jul 2 07:48:56.618634 google-clock-skew[1376]: INFO Clock drift token has changed: 0. Jul 2 07:48:56.628888 systemd-nspawn[1317]: hwclock: Cannot access the Hardware Clock via any known method. Jul 2 07:48:56.629188 systemd-nspawn[1317]: hwclock: Use the --verbose option to see the details of our search for an access method. Jul 2 07:48:56.630106 google-clock-skew[1376]: WARNING Failed to sync system time with hardware clock. Jul 2 07:48:56.646188 systemd[1]: Started sshd@3-10.128.0.12:22-147.75.109.163:43102.service. Jul 2 07:48:56.826131 google-networking[1377]: INFO Starting Google Networking daemon. Jul 2 07:48:56.846923 groupadd[1395]: group added to /etc/group: name=google-sudoers, GID=1000 Jul 2 07:48:56.850551 groupadd[1395]: group added to /etc/gshadow: name=google-sudoers Jul 2 07:48:56.857418 groupadd[1395]: new group: name=google-sudoers, GID=1000 Jul 2 07:48:56.872025 google-accounts[1375]: INFO Starting Google Accounts daemon. Jul 2 07:48:56.897417 google-accounts[1375]: WARNING OS Login not installed. Jul 2 07:48:56.898361 google-accounts[1375]: INFO Creating a new user account for 0. Jul 2 07:48:56.903745 systemd-nspawn[1317]: useradd: invalid user name '0': use --badname to ignore Jul 2 07:48:56.904637 google-accounts[1375]: WARNING Could not create user 0. Command '['useradd', '-m', '-s', '/bin/bash', '-p', '*', '0']' returned non-zero exit status 3.. Jul 2 07:48:56.953961 sshd[1389]: Accepted publickey for core from 147.75.109.163 port 43102 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:48:56.955760 sshd[1389]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:48:56.962291 systemd[1]: Started session-4.scope. Jul 2 07:48:56.963229 systemd-logind[1211]: New session 4 of user core. Jul 2 07:48:57.168272 sshd[1389]: pam_unix(sshd:session): session closed for user core Jul 2 07:48:57.172434 systemd[1]: sshd@3-10.128.0.12:22-147.75.109.163:43102.service: Deactivated successfully. Jul 2 07:48:57.173468 systemd[1]: session-4.scope: Deactivated successfully. Jul 2 07:48:57.174434 systemd-logind[1211]: Session 4 logged out. Waiting for processes to exit. Jul 2 07:48:57.175650 systemd-logind[1211]: Removed session 4. Jul 2 07:48:57.213580 systemd[1]: Started sshd@4-10.128.0.12:22-147.75.109.163:43108.service. Jul 2 07:48:57.502207 sshd[1410]: Accepted publickey for core from 147.75.109.163 port 43108 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:48:57.503964 sshd[1410]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:48:57.511094 systemd[1]: Started session-5.scope. Jul 2 07:48:57.511763 systemd-logind[1211]: New session 5 of user core. Jul 2 07:48:57.697991 sudo[1413]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jul 2 07:48:57.698416 sudo[1413]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Jul 2 07:48:57.732876 systemd[1]: Starting docker.service... Jul 2 07:48:57.782711 env[1423]: time="2024-07-02T07:48:57.782214938Z" level=info msg="Starting up" Jul 2 07:48:57.785439 env[1423]: time="2024-07-02T07:48:57.785367350Z" level=info msg="parsed scheme: \"unix\"" module=grpc Jul 2 07:48:57.785439 env[1423]: time="2024-07-02T07:48:57.785394026Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Jul 2 07:48:57.785439 env[1423]: time="2024-07-02T07:48:57.785422560Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Jul 2 07:48:57.785439 env[1423]: time="2024-07-02T07:48:57.785440535Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Jul 2 07:48:57.787471 env[1423]: time="2024-07-02T07:48:57.787417953Z" level=info msg="parsed scheme: \"unix\"" module=grpc Jul 2 07:48:57.787471 env[1423]: time="2024-07-02T07:48:57.787440663Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Jul 2 07:48:57.787471 env[1423]: time="2024-07-02T07:48:57.787460813Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Jul 2 07:48:57.787471 env[1423]: time="2024-07-02T07:48:57.787474190Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Jul 2 07:48:57.796394 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport131697089-merged.mount: Deactivated successfully. Jul 2 07:48:57.825981 env[1423]: time="2024-07-02T07:48:57.825906335Z" level=info msg="Loading containers: start." Jul 2 07:48:57.989799 kernel: Initializing XFRM netlink socket Jul 2 07:48:58.034146 env[1423]: time="2024-07-02T07:48:58.033357581Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Jul 2 07:48:58.110258 systemd-networkd[1022]: docker0: Link UP Jul 2 07:48:58.123979 env[1423]: time="2024-07-02T07:48:58.123930907Z" level=info msg="Loading containers: done." Jul 2 07:48:58.144463 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck327583582-merged.mount: Deactivated successfully. Jul 2 07:48:58.150562 env[1423]: time="2024-07-02T07:48:58.150505277Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jul 2 07:48:58.150796 env[1423]: time="2024-07-02T07:48:58.150744633Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Jul 2 07:48:58.150945 env[1423]: time="2024-07-02T07:48:58.150895527Z" level=info msg="Daemon has completed initialization" Jul 2 07:48:58.169276 systemd[1]: Started docker.service. Jul 2 07:48:58.181784 env[1423]: time="2024-07-02T07:48:58.181702162Z" level=info msg="API listen on /run/docker.sock" Jul 2 07:48:59.293594 env[1224]: time="2024-07-02T07:48:59.293519742Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.6\"" Jul 2 07:48:59.779331 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3789353599.mount: Deactivated successfully. Jul 2 07:48:59.781212 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jul 2 07:48:59.781494 systemd[1]: Stopped kubelet.service. Jul 2 07:48:59.781569 systemd[1]: kubelet.service: Consumed 1.445s CPU time. Jul 2 07:48:59.784433 systemd[1]: Starting kubelet.service... Jul 2 07:48:59.988190 systemd[1]: Started kubelet.service. Jul 2 07:49:00.055548 kubelet[1553]: E0702 07:49:00.055058 1553 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 2 07:49:00.061074 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 2 07:49:00.061287 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 2 07:49:01.982184 env[1224]: time="2024-07-02T07:49:01.982116003Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:01.984879 env[1224]: time="2024-07-02T07:49:01.984830750Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3af2ab51e136465590d968a2052e02e180fc7967a03724b269c1337e8f09d36f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:01.987384 env[1224]: time="2024-07-02T07:49:01.987339848Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:01.989591 env[1224]: time="2024-07-02T07:49:01.989552590Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:f4d993b3d73cc0d59558be584b5b40785b4a96874bc76873b69d1dd818485e70,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:01.990587 env[1224]: time="2024-07-02T07:49:01.990539191Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.6\" returns image reference \"sha256:3af2ab51e136465590d968a2052e02e180fc7967a03724b269c1337e8f09d36f\"" Jul 2 07:49:02.004339 env[1224]: time="2024-07-02T07:49:02.004280565Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.6\"" Jul 2 07:49:03.865630 env[1224]: time="2024-07-02T07:49:03.865561640Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:03.868056 env[1224]: time="2024-07-02T07:49:03.868006442Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:083b81fc09e858d3e0d4b42f567a9d44a2232b60bac396a94cbdd7ce1098235e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:03.870176 env[1224]: time="2024-07-02T07:49:03.870135102Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:03.872616 env[1224]: time="2024-07-02T07:49:03.872575705Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:692fc3f88a60b3afc76492ad347306d34042000f56f230959e9367fd59c48b1e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:03.873707 env[1224]: time="2024-07-02T07:49:03.873654791Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.6\" returns image reference \"sha256:083b81fc09e858d3e0d4b42f567a9d44a2232b60bac396a94cbdd7ce1098235e\"" Jul 2 07:49:03.887728 env[1224]: time="2024-07-02T07:49:03.887679401Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.6\"" Jul 2 07:49:05.057962 env[1224]: time="2024-07-02T07:49:05.057876601Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:05.060862 env[1224]: time="2024-07-02T07:49:05.060806672Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:49d9b8328a8fda6ebca6b3226c6d722d92ec7adffff18668511a88058444cf15,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:05.063875 env[1224]: time="2024-07-02T07:49:05.063837659Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:05.067074 env[1224]: time="2024-07-02T07:49:05.067030781Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:b91a4e45debd0d5336d9f533aefdf47d4b39b24071feb459e521709b9e4ec24f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:05.068636 env[1224]: time="2024-07-02T07:49:05.068590197Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.6\" returns image reference \"sha256:49d9b8328a8fda6ebca6b3226c6d722d92ec7adffff18668511a88058444cf15\"" Jul 2 07:49:05.083902 env[1224]: time="2024-07-02T07:49:05.083853296Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.6\"" Jul 2 07:49:06.087333 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1002939096.mount: Deactivated successfully. Jul 2 07:49:06.765326 env[1224]: time="2024-07-02T07:49:06.765254877Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:06.768048 env[1224]: time="2024-07-02T07:49:06.768003538Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:9c49592198fa15b509fe4ee4a538067866776e325d6dd33c77ad6647e1d3aac9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:06.770087 env[1224]: time="2024-07-02T07:49:06.770048541Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.29.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:06.771869 env[1224]: time="2024-07-02T07:49:06.771833488Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:88bacb3e1d6c0c37c6da95c6d6b8e30531d0b4d0ab540cc290b0af51fbfebd90,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:06.772494 env[1224]: time="2024-07-02T07:49:06.772456453Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.6\" returns image reference \"sha256:9c49592198fa15b509fe4ee4a538067866776e325d6dd33c77ad6647e1d3aac9\"" Jul 2 07:49:06.785738 env[1224]: time="2024-07-02T07:49:06.785701304Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Jul 2 07:49:07.182304 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3953920899.mount: Deactivated successfully. Jul 2 07:49:08.339100 env[1224]: time="2024-07-02T07:49:08.339025442Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.346576 env[1224]: time="2024-07-02T07:49:08.346523939Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.349169 env[1224]: time="2024-07-02T07:49:08.349126112Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.11.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.351718 env[1224]: time="2024-07-02T07:49:08.351676357Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.352855 env[1224]: time="2024-07-02T07:49:08.352761287Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:cbb01a7bd410dc08ba382018ab909a674fb0e48687f0c00797ed5bc34fcc6bb4\"" Jul 2 07:49:08.367553 env[1224]: time="2024-07-02T07:49:08.367508040Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Jul 2 07:49:08.683368 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4061662355.mount: Deactivated successfully. Jul 2 07:49:08.691977 env[1224]: time="2024-07-02T07:49:08.691921054Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.694846 env[1224]: time="2024-07-02T07:49:08.694008701Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.696061 env[1224]: time="2024-07-02T07:49:08.696021963Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.698031 env[1224]: time="2024-07-02T07:49:08.697990373Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:08.698828 env[1224]: time="2024-07-02T07:49:08.698787829Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Jul 2 07:49:08.713547 env[1224]: time="2024-07-02T07:49:08.713509556Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\"" Jul 2 07:49:09.117067 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3104099143.mount: Deactivated successfully. Jul 2 07:49:10.312516 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jul 2 07:49:10.312853 systemd[1]: Stopped kubelet.service. Jul 2 07:49:10.315148 systemd[1]: Starting kubelet.service... Jul 2 07:49:11.376197 systemd[1]: Started kubelet.service. Jul 2 07:49:11.493458 kubelet[1597]: E0702 07:49:11.493390 1597 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jul 2 07:49:11.496760 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jul 2 07:49:11.497049 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jul 2 07:49:11.784186 env[1224]: time="2024-07-02T07:49:11.783596958Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.10-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:11.787124 env[1224]: time="2024-07-02T07:49:11.787076522Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:11.789331 env[1224]: time="2024-07-02T07:49:11.789279925Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.10-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:11.790449 env[1224]: time="2024-07-02T07:49:11.790410441Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:11.791555 env[1224]: time="2024-07-02T07:49:11.791503446Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\"" Jul 2 07:49:15.710395 systemd[1]: Stopped kubelet.service. Jul 2 07:49:15.713888 systemd[1]: Starting kubelet.service... Jul 2 07:49:15.740952 systemd[1]: Reloading. Jul 2 07:49:15.872941 /usr/lib/systemd/system-generators/torcx-generator[1693]: time="2024-07-02T07:49:15Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.5 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.5 /var/lib/torcx/store]" Jul 2 07:49:15.874201 /usr/lib/systemd/system-generators/torcx-generator[1693]: time="2024-07-02T07:49:15Z" level=info msg="torcx already run" Jul 2 07:49:15.994809 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Jul 2 07:49:15.994837 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Jul 2 07:49:16.019417 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 2 07:49:16.161834 systemd[1]: Started kubelet.service. Jul 2 07:49:16.170588 systemd[1]: Stopping kubelet.service... Jul 2 07:49:16.171154 systemd[1]: kubelet.service: Deactivated successfully. Jul 2 07:49:16.171427 systemd[1]: Stopped kubelet.service. Jul 2 07:49:16.173980 systemd[1]: Starting kubelet.service... Jul 2 07:49:16.384082 systemd[1]: Started kubelet.service. Jul 2 07:49:16.449012 kubelet[1744]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 2 07:49:16.449012 kubelet[1744]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 2 07:49:16.449012 kubelet[1744]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 2 07:49:16.449600 kubelet[1744]: I0702 07:49:16.449089 1744 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 2 07:49:16.846706 kubelet[1744]: I0702 07:49:16.846631 1744 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Jul 2 07:49:16.846706 kubelet[1744]: I0702 07:49:16.846674 1744 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 2 07:49:16.847062 kubelet[1744]: I0702 07:49:16.847023 1744 server.go:919] "Client rotation is on, will bootstrap in background" Jul 2 07:49:16.880936 kubelet[1744]: E0702 07:49:16.880900 1744 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.128.0.12:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.882343 kubelet[1744]: I0702 07:49:16.882293 1744 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 2 07:49:16.900330 kubelet[1744]: I0702 07:49:16.900294 1744 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 2 07:49:16.900832 kubelet[1744]: I0702 07:49:16.900759 1744 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 2 07:49:16.901107 kubelet[1744]: I0702 07:49:16.901069 1744 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jul 2 07:49:16.902004 kubelet[1744]: I0702 07:49:16.901965 1744 topology_manager.go:138] "Creating topology manager with none policy" Jul 2 07:49:16.902004 kubelet[1744]: I0702 07:49:16.902003 1744 container_manager_linux.go:301] "Creating device plugin manager" Jul 2 07:49:16.902229 kubelet[1744]: I0702 07:49:16.902190 1744 state_mem.go:36] "Initialized new in-memory state store" Jul 2 07:49:16.902422 kubelet[1744]: I0702 07:49:16.902346 1744 kubelet.go:396] "Attempting to sync node with API server" Jul 2 07:49:16.902553 kubelet[1744]: I0702 07:49:16.902426 1744 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 2 07:49:16.902553 kubelet[1744]: I0702 07:49:16.902477 1744 kubelet.go:312] "Adding apiserver pod source" Jul 2 07:49:16.902553 kubelet[1744]: I0702 07:49:16.902503 1744 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 2 07:49:16.909709 kubelet[1744]: I0702 07:49:16.909684 1744 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Jul 2 07:49:16.914901 systemd[1]: systemd-hostnamed.service: Deactivated successfully. Jul 2 07:49:16.918383 kubelet[1744]: I0702 07:49:16.918347 1744 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 2 07:49:16.922794 kubelet[1744]: W0702 07:49:16.922749 1744 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jul 2 07:49:16.931519 kubelet[1744]: W0702 07:49:16.931451 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://10.128.0.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal&limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.931706 kubelet[1744]: E0702 07:49:16.931684 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.128.0.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal&limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.932005 kubelet[1744]: W0702 07:49:16.931961 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://10.128.0.12:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.932137 kubelet[1744]: E0702 07:49:16.932119 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.128.0.12:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.932352 kubelet[1744]: I0702 07:49:16.932331 1744 server.go:1256] "Started kubelet" Jul 2 07:49:16.941870 kubelet[1744]: E0702 07:49:16.941847 1744 event.go:355] "Unable to write event (may retry after sleeping)" err="Post \"https://10.128.0.12:6443/api/v1/namespaces/default/events\": dial tcp 10.128.0.12:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal.17de55e6acd047a4 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,UID:ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,},FirstTimestamp:2024-07-02 07:49:16.93227818 +0000 UTC m=+0.542005059,LastTimestamp:2024-07-02 07:49:16.93227818 +0000 UTC m=+0.542005059,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,}" Jul 2 07:49:16.942134 kubelet[1744]: I0702 07:49:16.942116 1744 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 2 07:49:16.942545 kubelet[1744]: I0702 07:49:16.942528 1744 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 2 07:49:16.942657 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Jul 2 07:49:16.942865 kubelet[1744]: I0702 07:49:16.942848 1744 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Jul 2 07:49:16.943395 kubelet[1744]: I0702 07:49:16.943363 1744 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 2 07:49:16.944368 kubelet[1744]: I0702 07:49:16.944343 1744 server.go:461] "Adding debug handlers to kubelet server" Jul 2 07:49:16.948751 kubelet[1744]: I0702 07:49:16.948713 1744 volume_manager.go:291] "Starting Kubelet Volume Manager" Jul 2 07:49:16.949810 kubelet[1744]: I0702 07:49:16.949784 1744 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Jul 2 07:49:16.951915 kubelet[1744]: I0702 07:49:16.951890 1744 reconciler_new.go:29] "Reconciler: start to sync state" Jul 2 07:49:16.953797 kubelet[1744]: W0702 07:49:16.953081 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://10.128.0.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.953797 kubelet[1744]: E0702 07:49:16.953162 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.128.0.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.955323 kubelet[1744]: I0702 07:49:16.955301 1744 factory.go:221] Registration of the systemd container factory successfully Jul 2 07:49:16.955580 kubelet[1744]: I0702 07:49:16.955554 1744 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 2 07:49:16.957903 kubelet[1744]: E0702 07:49:16.957878 1744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal?timeout=10s\": dial tcp 10.128.0.12:6443: connect: connection refused" interval="200ms" Jul 2 07:49:16.962692 kubelet[1744]: I0702 07:49:16.962458 1744 factory.go:221] Registration of the containerd container factory successfully Jul 2 07:49:16.974323 kubelet[1744]: E0702 07:49:16.974302 1744 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 2 07:49:16.985681 kubelet[1744]: I0702 07:49:16.985632 1744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 2 07:49:16.988227 kubelet[1744]: I0702 07:49:16.988198 1744 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 2 07:49:16.988337 kubelet[1744]: I0702 07:49:16.988233 1744 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 2 07:49:16.988337 kubelet[1744]: I0702 07:49:16.988274 1744 kubelet.go:2329] "Starting kubelet main sync loop" Jul 2 07:49:16.988459 kubelet[1744]: E0702 07:49:16.988357 1744 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 2 07:49:16.988614 kubelet[1744]: I0702 07:49:16.988594 1744 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 2 07:49:16.988741 kubelet[1744]: I0702 07:49:16.988725 1744 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 2 07:49:16.989265 kubelet[1744]: I0702 07:49:16.989208 1744 state_mem.go:36] "Initialized new in-memory state store" Jul 2 07:49:16.989604 kubelet[1744]: W0702 07:49:16.989576 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://10.128.0.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.989751 kubelet[1744]: E0702 07:49:16.989733 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.128.0.12:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:16.991906 kubelet[1744]: I0702 07:49:16.991875 1744 policy_none.go:49] "None policy: Start" Jul 2 07:49:16.993450 kubelet[1744]: I0702 07:49:16.993428 1744 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 2 07:49:16.993542 kubelet[1744]: I0702 07:49:16.993462 1744 state_mem.go:35] "Initializing new in-memory state store" Jul 2 07:49:17.004025 systemd[1]: Created slice kubepods.slice. Jul 2 07:49:17.012174 systemd[1]: Created slice kubepods-burstable.slice. Jul 2 07:49:17.017514 systemd[1]: Created slice kubepods-besteffort.slice. Jul 2 07:49:17.027747 kubelet[1744]: I0702 07:49:17.027719 1744 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 2 07:49:17.028139 kubelet[1744]: I0702 07:49:17.028100 1744 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 2 07:49:17.030194 kubelet[1744]: E0702 07:49:17.030137 1744 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" not found" Jul 2 07:49:17.054994 kubelet[1744]: I0702 07:49:17.054970 1744 kubelet_node_status.go:73] "Attempting to register node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.055473 kubelet[1744]: E0702 07:49:17.055437 1744 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.128.0.12:6443/api/v1/nodes\": dial tcp 10.128.0.12:6443: connect: connection refused" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.088892 kubelet[1744]: I0702 07:49:17.088846 1744 topology_manager.go:215] "Topology Admit Handler" podUID="9d9453c32d7a8fb57c60efc173598673" podNamespace="kube-system" podName="kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.094270 kubelet[1744]: I0702 07:49:17.094241 1744 topology_manager.go:215] "Topology Admit Handler" podUID="f9770990d6a30af493fdab6ba16d0b7c" podNamespace="kube-system" podName="kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.103669 kubelet[1744]: I0702 07:49:17.103385 1744 topology_manager.go:215] "Topology Admit Handler" podUID="58cd976d0cbee984c26678ef1d9b013b" podNamespace="kube-system" podName="kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.110599 systemd[1]: Created slice kubepods-burstable-pod9d9453c32d7a8fb57c60efc173598673.slice. Jul 2 07:49:17.123832 systemd[1]: Created slice kubepods-burstable-podf9770990d6a30af493fdab6ba16d0b7c.slice. Jul 2 07:49:17.130425 systemd[1]: Created slice kubepods-burstable-pod58cd976d0cbee984c26678ef1d9b013b.slice. Jul 2 07:49:17.158688 kubelet[1744]: E0702 07:49:17.158649 1744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal?timeout=10s\": dial tcp 10.128.0.12:6443: connect: connection refused" interval="400ms" Jul 2 07:49:17.253293 kubelet[1744]: I0702 07:49:17.253228 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9d9453c32d7a8fb57c60efc173598673-ca-certs\") pod \"kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"9d9453c32d7a8fb57c60efc173598673\") " pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.253293 kubelet[1744]: I0702 07:49:17.253294 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9d9453c32d7a8fb57c60efc173598673-k8s-certs\") pod \"kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"9d9453c32d7a8fb57c60efc173598673\") " pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.253574 kubelet[1744]: I0702 07:49:17.253331 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-ca-certs\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.253574 kubelet[1744]: I0702 07:49:17.253365 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-k8s-certs\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.253574 kubelet[1744]: I0702 07:49:17.253402 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-kubeconfig\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.254212 kubelet[1744]: I0702 07:49:17.254176 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.254372 kubelet[1744]: I0702 07:49:17.254253 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9d9453c32d7a8fb57c60efc173598673-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"9d9453c32d7a8fb57c60efc173598673\") " pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.254372 kubelet[1744]: I0702 07:49:17.254294 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-flexvolume-dir\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.254372 kubelet[1744]: I0702 07:49:17.254335 1744 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/58cd976d0cbee984c26678ef1d9b013b-kubeconfig\") pod \"kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"58cd976d0cbee984c26678ef1d9b013b\") " pod="kube-system/kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.260727 kubelet[1744]: I0702 07:49:17.260682 1744 kubelet_node_status.go:73] "Attempting to register node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.261319 kubelet[1744]: E0702 07:49:17.261266 1744 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.128.0.12:6443/api/v1/nodes\": dial tcp 10.128.0.12:6443: connect: connection refused" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.422206 env[1224]: time="2024-07-02T07:49:17.421202686Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,Uid:9d9453c32d7a8fb57c60efc173598673,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:17.430142 env[1224]: time="2024-07-02T07:49:17.430079309Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,Uid:f9770990d6a30af493fdab6ba16d0b7c,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:17.435438 env[1224]: time="2024-07-02T07:49:17.435376599Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,Uid:58cd976d0cbee984c26678ef1d9b013b,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:17.560257 kubelet[1744]: E0702 07:49:17.560220 1744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal?timeout=10s\": dial tcp 10.128.0.12:6443: connect: connection refused" interval="800ms" Jul 2 07:49:17.666544 kubelet[1744]: I0702 07:49:17.666484 1744 kubelet_node_status.go:73] "Attempting to register node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.666954 kubelet[1744]: E0702 07:49:17.666929 1744 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.128.0.12:6443/api/v1/nodes\": dial tcp 10.128.0.12:6443: connect: connection refused" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:17.777101 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4021009720.mount: Deactivated successfully. Jul 2 07:49:17.787342 env[1224]: time="2024-07-02T07:49:17.787290878Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.788721 env[1224]: time="2024-07-02T07:49:17.788676660Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.790704 env[1224]: time="2024-07-02T07:49:17.790665025Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.792434 env[1224]: time="2024-07-02T07:49:17.792369479Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.793839 env[1224]: time="2024-07-02T07:49:17.793804322Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.795448 env[1224]: time="2024-07-02T07:49:17.795414992Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.797888 env[1224]: time="2024-07-02T07:49:17.797851805Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.799034 env[1224]: time="2024-07-02T07:49:17.798987873Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.800359 env[1224]: time="2024-07-02T07:49:17.800309762Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.802281 env[1224]: time="2024-07-02T07:49:17.802244577Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.805609 env[1224]: time="2024-07-02T07:49:17.805544202Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.808906 env[1224]: time="2024-07-02T07:49:17.808820089Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:17.842713 env[1224]: time="2024-07-02T07:49:17.842625958Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:49:17.842992 env[1224]: time="2024-07-02T07:49:17.842685360Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:49:17.842992 env[1224]: time="2024-07-02T07:49:17.842705097Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:49:17.842992 env[1224]: time="2024-07-02T07:49:17.842922788Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9b82d723609d0dc280f1ca4d7e754f40b078abf80d591cd2f9bed958b0858c40 pid=1786 runtime=io.containerd.runc.v2 Jul 2 07:49:17.864571 env[1224]: time="2024-07-02T07:49:17.864340394Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:49:17.864571 env[1224]: time="2024-07-02T07:49:17.864397415Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:49:17.864571 env[1224]: time="2024-07-02T07:49:17.864417707Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:49:17.864905 env[1224]: time="2024-07-02T07:49:17.864626023Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0c9b6b516b2ceaa92c8e1b48cc0bd1eb5e133be68336bc2a683929e8be0ce6c5 pid=1805 runtime=io.containerd.runc.v2 Jul 2 07:49:17.873522 env[1224]: time="2024-07-02T07:49:17.873446585Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:49:17.873792 env[1224]: time="2024-07-02T07:49:17.873723291Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:49:17.874093 env[1224]: time="2024-07-02T07:49:17.873957830Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:49:17.874750 env[1224]: time="2024-07-02T07:49:17.874696686Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/4608ef841ffc0751f808a5868f46ee233692c65a25d894dbde69b76b3d3f5c7b pid=1823 runtime=io.containerd.runc.v2 Jul 2 07:49:17.880657 systemd[1]: Started cri-containerd-9b82d723609d0dc280f1ca4d7e754f40b078abf80d591cd2f9bed958b0858c40.scope. Jul 2 07:49:17.903804 systemd[1]: Started cri-containerd-0c9b6b516b2ceaa92c8e1b48cc0bd1eb5e133be68336bc2a683929e8be0ce6c5.scope. Jul 2 07:49:17.932240 systemd[1]: Started cri-containerd-4608ef841ffc0751f808a5868f46ee233692c65a25d894dbde69b76b3d3f5c7b.scope. Jul 2 07:49:17.988539 env[1224]: time="2024-07-02T07:49:17.988488317Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,Uid:9d9453c32d7a8fb57c60efc173598673,Namespace:kube-system,Attempt:0,} returns sandbox id \"9b82d723609d0dc280f1ca4d7e754f40b078abf80d591cd2f9bed958b0858c40\"" Jul 2 07:49:17.991451 kubelet[1744]: E0702 07:49:17.991026 1744 kubelet_pods.go:417] "Hostname for pod was too long, truncated it" podName="kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" hostnameMaxLen=63 truncatedHostname="kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-21291" Jul 2 07:49:17.994587 env[1224]: time="2024-07-02T07:49:17.994547322Z" level=info msg="CreateContainer within sandbox \"9b82d723609d0dc280f1ca4d7e754f40b078abf80d591cd2f9bed958b0858c40\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jul 2 07:49:18.023798 env[1224]: time="2024-07-02T07:49:18.022810522Z" level=info msg="CreateContainer within sandbox \"9b82d723609d0dc280f1ca4d7e754f40b078abf80d591cd2f9bed958b0858c40\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"a212eb1686dada6db5c58f934472971a2fc4762ad4df09e1980f1af81f375e2a\"" Jul 2 07:49:18.023798 env[1224]: time="2024-07-02T07:49:18.023563489Z" level=info msg="StartContainer for \"a212eb1686dada6db5c58f934472971a2fc4762ad4df09e1980f1af81f375e2a\"" Jul 2 07:49:18.025138 kubelet[1744]: W0702 07:49:18.025100 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://10.128.0.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:18.025276 kubelet[1744]: E0702 07:49:18.025168 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.128.0.12:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:18.032426 env[1224]: time="2024-07-02T07:49:18.032307381Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,Uid:f9770990d6a30af493fdab6ba16d0b7c,Namespace:kube-system,Attempt:0,} returns sandbox id \"0c9b6b516b2ceaa92c8e1b48cc0bd1eb5e133be68336bc2a683929e8be0ce6c5\"" Jul 2 07:49:18.036478 kubelet[1744]: E0702 07:49:18.036449 1744 kubelet_pods.go:417] "Hostname for pod was too long, truncated it" podName="kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" hostnameMaxLen=63 truncatedHostname="kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flat" Jul 2 07:49:18.038654 env[1224]: time="2024-07-02T07:49:18.038592083Z" level=info msg="CreateContainer within sandbox \"0c9b6b516b2ceaa92c8e1b48cc0bd1eb5e133be68336bc2a683929e8be0ce6c5\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jul 2 07:49:18.059375 env[1224]: time="2024-07-02T07:49:18.059323827Z" level=info msg="CreateContainer within sandbox \"0c9b6b516b2ceaa92c8e1b48cc0bd1eb5e133be68336bc2a683929e8be0ce6c5\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"12db4668d7aa070c757ab1ffbe79baf6a1f6a8e92c05d7ee57b182b4991b3a6b\"" Jul 2 07:49:18.060267 env[1224]: time="2024-07-02T07:49:18.060227514Z" level=info msg="StartContainer for \"12db4668d7aa070c757ab1ffbe79baf6a1f6a8e92c05d7ee57b182b4991b3a6b\"" Jul 2 07:49:18.078319 systemd[1]: Started cri-containerd-a212eb1686dada6db5c58f934472971a2fc4762ad4df09e1980f1af81f375e2a.scope. Jul 2 07:49:18.089017 env[1224]: time="2024-07-02T07:49:18.088960707Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal,Uid:58cd976d0cbee984c26678ef1d9b013b,Namespace:kube-system,Attempt:0,} returns sandbox id \"4608ef841ffc0751f808a5868f46ee233692c65a25d894dbde69b76b3d3f5c7b\"" Jul 2 07:49:18.091419 kubelet[1744]: E0702 07:49:18.091043 1744 kubelet_pods.go:417] "Hostname for pod was too long, truncated it" podName="kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" hostnameMaxLen=63 truncatedHostname="kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-21291" Jul 2 07:49:18.093952 env[1224]: time="2024-07-02T07:49:18.093892338Z" level=info msg="CreateContainer within sandbox \"4608ef841ffc0751f808a5868f46ee233692c65a25d894dbde69b76b3d3f5c7b\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jul 2 07:49:18.119640 systemd[1]: Started cri-containerd-12db4668d7aa070c757ab1ffbe79baf6a1f6a8e92c05d7ee57b182b4991b3a6b.scope. Jul 2 07:49:18.143092 env[1224]: time="2024-07-02T07:49:18.143035806Z" level=info msg="CreateContainer within sandbox \"4608ef841ffc0751f808a5868f46ee233692c65a25d894dbde69b76b3d3f5c7b\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"1d4318d1917832fd83e65a898726815be850fbb93858333edca8b21e22ff8a97\"" Jul 2 07:49:18.144223 env[1224]: time="2024-07-02T07:49:18.144184367Z" level=info msg="StartContainer for \"1d4318d1917832fd83e65a898726815be850fbb93858333edca8b21e22ff8a97\"" Jul 2 07:49:18.175730 kubelet[1744]: W0702 07:49:18.175662 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://10.128.0.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal&limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:18.177750 kubelet[1744]: E0702 07:49:18.177713 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.128.0.12:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal&limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:18.201778 systemd[1]: Started cri-containerd-1d4318d1917832fd83e65a898726815be850fbb93858333edca8b21e22ff8a97.scope. Jul 2 07:49:18.216856 env[1224]: time="2024-07-02T07:49:18.216806308Z" level=info msg="StartContainer for \"a212eb1686dada6db5c58f934472971a2fc4762ad4df09e1980f1af81f375e2a\" returns successfully" Jul 2 07:49:18.249934 env[1224]: time="2024-07-02T07:49:18.249833177Z" level=info msg="StartContainer for \"12db4668d7aa070c757ab1ffbe79baf6a1f6a8e92c05d7ee57b182b4991b3a6b\" returns successfully" Jul 2 07:49:18.287682 kubelet[1744]: W0702 07:49:18.286679 1744 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://10.128.0.12:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:18.287941 kubelet[1744]: E0702 07:49:18.287898 1744 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.128.0.12:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.128.0.12:6443: connect: connection refused Jul 2 07:49:18.299744 env[1224]: time="2024-07-02T07:49:18.299678232Z" level=info msg="StartContainer for \"1d4318d1917832fd83e65a898726815be850fbb93858333edca8b21e22ff8a97\" returns successfully" Jul 2 07:49:18.361756 kubelet[1744]: E0702 07:49:18.361662 1744 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.128.0.12:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal?timeout=10s\": dial tcp 10.128.0.12:6443: connect: connection refused" interval="1.6s" Jul 2 07:49:18.472849 kubelet[1744]: I0702 07:49:18.472814 1744 kubelet_node_status.go:73] "Attempting to register node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:22.300298 kubelet[1744]: I0702 07:49:22.300245 1744 kubelet_node_status.go:76] "Successfully registered node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:22.910174 kubelet[1744]: I0702 07:49:22.910121 1744 apiserver.go:52] "Watching apiserver" Jul 2 07:49:22.952201 kubelet[1744]: I0702 07:49:22.952152 1744 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Jul 2 07:49:23.265792 kubelet[1744]: W0702 07:49:23.265740 1744 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters must not contain dots] Jul 2 07:49:25.037254 systemd[1]: Reloading. Jul 2 07:49:25.167665 /usr/lib/systemd/system-generators/torcx-generator[2035]: time="2024-07-02T07:49:25Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.5 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.5 /var/lib/torcx/store]" Jul 2 07:49:25.172648 /usr/lib/systemd/system-generators/torcx-generator[2035]: time="2024-07-02T07:49:25Z" level=info msg="torcx already run" Jul 2 07:49:25.264590 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Jul 2 07:49:25.264618 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Jul 2 07:49:25.289640 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jul 2 07:49:25.445715 kubelet[1744]: I0702 07:49:25.445675 1744 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 2 07:49:25.448699 systemd[1]: Stopping kubelet.service... Jul 2 07:49:25.463839 systemd[1]: kubelet.service: Deactivated successfully. Jul 2 07:49:25.464091 systemd[1]: Stopped kubelet.service. Jul 2 07:49:25.464172 systemd[1]: kubelet.service: Consumed 1.028s CPU time. Jul 2 07:49:25.466533 systemd[1]: Starting kubelet.service... Jul 2 07:49:25.649723 systemd[1]: Started kubelet.service. Jul 2 07:49:25.736630 kubelet[2083]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 2 07:49:25.737185 kubelet[2083]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jul 2 07:49:25.737287 kubelet[2083]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jul 2 07:49:25.737506 kubelet[2083]: I0702 07:49:25.737446 2083 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jul 2 07:49:25.749391 kubelet[2083]: I0702 07:49:25.749357 2083 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Jul 2 07:49:25.749564 kubelet[2083]: I0702 07:49:25.749545 2083 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jul 2 07:49:25.749960 kubelet[2083]: I0702 07:49:25.749938 2083 server.go:919] "Client rotation is on, will bootstrap in background" Jul 2 07:49:25.752690 kubelet[2083]: I0702 07:49:25.752663 2083 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jul 2 07:49:25.756209 kubelet[2083]: I0702 07:49:25.756184 2083 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jul 2 07:49:25.766932 sudo[2096]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Jul 2 07:49:25.767384 sudo[2096]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Jul 2 07:49:25.774322 kubelet[2083]: I0702 07:49:25.774288 2083 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jul 2 07:49:25.774932 kubelet[2083]: I0702 07:49:25.774912 2083 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jul 2 07:49:25.775477 kubelet[2083]: I0702 07:49:25.775430 2083 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jul 2 07:49:25.775724 kubelet[2083]: I0702 07:49:25.775705 2083 topology_manager.go:138] "Creating topology manager with none policy" Jul 2 07:49:25.775881 kubelet[2083]: I0702 07:49:25.775865 2083 container_manager_linux.go:301] "Creating device plugin manager" Jul 2 07:49:25.776110 kubelet[2083]: I0702 07:49:25.776094 2083 state_mem.go:36] "Initialized new in-memory state store" Jul 2 07:49:25.776420 kubelet[2083]: I0702 07:49:25.776403 2083 kubelet.go:396] "Attempting to sync node with API server" Jul 2 07:49:25.776569 kubelet[2083]: I0702 07:49:25.776554 2083 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jul 2 07:49:25.776737 kubelet[2083]: I0702 07:49:25.776721 2083 kubelet.go:312] "Adding apiserver pod source" Jul 2 07:49:25.776918 kubelet[2083]: I0702 07:49:25.776894 2083 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jul 2 07:49:25.785343 kubelet[2083]: I0702 07:49:25.785214 2083 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Jul 2 07:49:25.786394 kubelet[2083]: I0702 07:49:25.786375 2083 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jul 2 07:49:25.787869 kubelet[2083]: I0702 07:49:25.787851 2083 server.go:1256] "Started kubelet" Jul 2 07:49:25.792758 kubelet[2083]: I0702 07:49:25.788136 2083 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jul 2 07:49:25.793294 kubelet[2083]: I0702 07:49:25.793277 2083 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jul 2 07:49:25.793442 kubelet[2083]: I0702 07:49:25.791808 2083 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Jul 2 07:49:25.797541 kubelet[2083]: I0702 07:49:25.797522 2083 server.go:461] "Adding debug handlers to kubelet server" Jul 2 07:49:25.807936 kubelet[2083]: I0702 07:49:25.807913 2083 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jul 2 07:49:25.815989 kubelet[2083]: E0702 07:49:25.815969 2083 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jul 2 07:49:25.836308 kubelet[2083]: I0702 07:49:25.836280 2083 volume_manager.go:291] "Starting Kubelet Volume Manager" Jul 2 07:49:25.838140 kubelet[2083]: I0702 07:49:25.838119 2083 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Jul 2 07:49:25.838898 kubelet[2083]: I0702 07:49:25.838879 2083 reconciler_new.go:29] "Reconciler: start to sync state" Jul 2 07:49:25.857736 kubelet[2083]: I0702 07:49:25.857660 2083 factory.go:221] Registration of the systemd container factory successfully Jul 2 07:49:25.858083 kubelet[2083]: I0702 07:49:25.858059 2083 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jul 2 07:49:25.862884 kubelet[2083]: I0702 07:49:25.862861 2083 factory.go:221] Registration of the containerd container factory successfully Jul 2 07:49:25.916242 kubelet[2083]: I0702 07:49:25.916148 2083 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jul 2 07:49:25.925889 kubelet[2083]: I0702 07:49:25.925864 2083 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jul 2 07:49:25.926141 kubelet[2083]: I0702 07:49:25.926125 2083 status_manager.go:217] "Starting to sync pod status with apiserver" Jul 2 07:49:25.926262 kubelet[2083]: I0702 07:49:25.926247 2083 kubelet.go:2329] "Starting kubelet main sync loop" Jul 2 07:49:25.926423 kubelet[2083]: E0702 07:49:25.926409 2083 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jul 2 07:49:25.943195 kubelet[2083]: I0702 07:49:25.943172 2083 kubelet_node_status.go:73] "Attempting to register node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:25.957112 kubelet[2083]: I0702 07:49:25.956560 2083 kubelet_node_status.go:112] "Node was previously registered" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:25.957112 kubelet[2083]: I0702 07:49:25.956721 2083 kubelet_node_status.go:76] "Successfully registered node" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:25.984913 kubelet[2083]: I0702 07:49:25.984887 2083 cpu_manager.go:214] "Starting CPU manager" policy="none" Jul 2 07:49:25.985093 kubelet[2083]: I0702 07:49:25.985077 2083 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jul 2 07:49:25.985203 kubelet[2083]: I0702 07:49:25.985191 2083 state_mem.go:36] "Initialized new in-memory state store" Jul 2 07:49:25.985483 kubelet[2083]: I0702 07:49:25.985468 2083 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jul 2 07:49:25.985621 kubelet[2083]: I0702 07:49:25.985605 2083 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jul 2 07:49:25.985724 kubelet[2083]: I0702 07:49:25.985711 2083 policy_none.go:49] "None policy: Start" Jul 2 07:49:25.986863 kubelet[2083]: I0702 07:49:25.986843 2083 memory_manager.go:170] "Starting memorymanager" policy="None" Jul 2 07:49:25.987028 kubelet[2083]: I0702 07:49:25.987013 2083 state_mem.go:35] "Initializing new in-memory state store" Jul 2 07:49:25.987299 kubelet[2083]: I0702 07:49:25.987285 2083 state_mem.go:75] "Updated machine memory state" Jul 2 07:49:25.992893 kubelet[2083]: I0702 07:49:25.992873 2083 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jul 2 07:49:25.996149 kubelet[2083]: I0702 07:49:25.996129 2083 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jul 2 07:49:26.026682 kubelet[2083]: I0702 07:49:26.026583 2083 topology_manager.go:215] "Topology Admit Handler" podUID="9d9453c32d7a8fb57c60efc173598673" podNamespace="kube-system" podName="kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.026682 kubelet[2083]: I0702 07:49:26.026675 2083 topology_manager.go:215] "Topology Admit Handler" podUID="f9770990d6a30af493fdab6ba16d0b7c" podNamespace="kube-system" podName="kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.026927 kubelet[2083]: I0702 07:49:26.026727 2083 topology_manager.go:215] "Topology Admit Handler" podUID="58cd976d0cbee984c26678ef1d9b013b" podNamespace="kube-system" podName="kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.038587 kubelet[2083]: W0702 07:49:26.037735 2083 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters must not contain dots] Jul 2 07:49:26.038906 kubelet[2083]: E0702 07:49:26.038880 2083 kubelet.go:1921] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" already exists" pod="kube-system/kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.039189 kubelet[2083]: W0702 07:49:26.038035 2083 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters must not contain dots] Jul 2 07:49:26.040817 kubelet[2083]: W0702 07:49:26.039516 2083 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must be no more than 63 characters must not contain dots] Jul 2 07:49:26.041782 kubelet[2083]: I0702 07:49:26.041747 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-flexvolume-dir\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042000 kubelet[2083]: I0702 07:49:26.041972 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-k8s-certs\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042159 kubelet[2083]: I0702 07:49:26.042145 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/58cd976d0cbee984c26678ef1d9b013b-kubeconfig\") pod \"kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"58cd976d0cbee984c26678ef1d9b013b\") " pod="kube-system/kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042324 kubelet[2083]: I0702 07:49:26.042310 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/9d9453c32d7a8fb57c60efc173598673-k8s-certs\") pod \"kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"9d9453c32d7a8fb57c60efc173598673\") " pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042486 kubelet[2083]: I0702 07:49:26.042472 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/9d9453c32d7a8fb57c60efc173598673-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"9d9453c32d7a8fb57c60efc173598673\") " pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042640 kubelet[2083]: I0702 07:49:26.042618 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-ca-certs\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042741 kubelet[2083]: I0702 07:49:26.042663 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/9d9453c32d7a8fb57c60efc173598673-ca-certs\") pod \"kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"9d9453c32d7a8fb57c60efc173598673\") " pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.042848 kubelet[2083]: I0702 07:49:26.042730 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-kubeconfig\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.045811 kubelet[2083]: I0702 07:49:26.042941 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/f9770990d6a30af493fdab6ba16d0b7c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" (UID: \"f9770990d6a30af493fdab6ba16d0b7c\") " pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" Jul 2 07:49:26.618841 sudo[2096]: pam_unix(sudo:session): session closed for user root Jul 2 07:49:26.795795 kubelet[2083]: I0702 07:49:26.795731 2083 apiserver.go:52] "Watching apiserver" Jul 2 07:49:26.838668 kubelet[2083]: I0702 07:49:26.838618 2083 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Jul 2 07:49:27.027427 kubelet[2083]: I0702 07:49:27.027381 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" podStartSLOduration=4.02731504 podStartE2EDuration="4.02731504s" podCreationTimestamp="2024-07-02 07:49:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:49:27.025090592 +0000 UTC m=+1.364582921" watchObservedRunningTime="2024-07-02 07:49:27.02731504 +0000 UTC m=+1.366807355" Jul 2 07:49:27.050533 kubelet[2083]: I0702 07:49:27.050496 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" podStartSLOduration=1.050432255 podStartE2EDuration="1.050432255s" podCreationTimestamp="2024-07-02 07:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:49:27.03848985 +0000 UTC m=+1.377982178" watchObservedRunningTime="2024-07-02 07:49:27.050432255 +0000 UTC m=+1.389924571" Jul 2 07:49:27.050783 kubelet[2083]: I0702 07:49:27.050634 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" podStartSLOduration=1.050606327 podStartE2EDuration="1.050606327s" podCreationTimestamp="2024-07-02 07:49:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:49:27.048668082 +0000 UTC m=+1.388160405" watchObservedRunningTime="2024-07-02 07:49:27.050606327 +0000 UTC m=+1.390098651" Jul 2 07:49:28.903424 sudo[1413]: pam_unix(sudo:session): session closed for user root Jul 2 07:49:28.946074 sshd[1410]: pam_unix(sshd:session): session closed for user core Jul 2 07:49:28.951159 systemd-logind[1211]: Session 5 logged out. Waiting for processes to exit. Jul 2 07:49:28.951462 systemd[1]: sshd@4-10.128.0.12:22-147.75.109.163:43108.service: Deactivated successfully. Jul 2 07:49:28.952617 systemd[1]: session-5.scope: Deactivated successfully. Jul 2 07:49:28.952888 systemd[1]: session-5.scope: Consumed 7.058s CPU time. Jul 2 07:49:28.953995 systemd-logind[1211]: Removed session 5. Jul 2 07:49:31.817578 update_engine[1212]: I0702 07:49:31.817513 1212 update_attempter.cc:509] Updating boot flags... Jul 2 07:49:40.311999 kubelet[2083]: I0702 07:49:40.311956 2083 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jul 2 07:49:40.312891 env[1224]: time="2024-07-02T07:49:40.312840362Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jul 2 07:49:40.313383 kubelet[2083]: I0702 07:49:40.313214 2083 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jul 2 07:49:40.341180 kubelet[2083]: I0702 07:49:40.341139 2083 topology_manager.go:215] "Topology Admit Handler" podUID="854b3f29-4cd6-4527-b345-4dab03b86eea" podNamespace="kube-system" podName="kube-proxy-7g7mc" Jul 2 07:49:40.350127 systemd[1]: Created slice kubepods-besteffort-pod854b3f29_4cd6_4527_b345_4dab03b86eea.slice. Jul 2 07:49:40.368027 kubelet[2083]: I0702 07:49:40.367990 2083 topology_manager.go:215] "Topology Admit Handler" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" podNamespace="kube-system" podName="cilium-psg5v" Jul 2 07:49:40.375671 systemd[1]: Created slice kubepods-burstable-pod222f0b80_bb4d_48ae_a0a0_384baa208d60.slice. Jul 2 07:49:40.440695 kubelet[2083]: I0702 07:49:40.440650 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-hostproc\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.440918 kubelet[2083]: I0702 07:49:40.440743 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-lib-modules\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.440918 kubelet[2083]: I0702 07:49:40.440804 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-etc-cni-netd\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.440918 kubelet[2083]: I0702 07:49:40.440838 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-xtables-lock\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.440918 kubelet[2083]: I0702 07:49:40.440893 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/854b3f29-4cd6-4527-b345-4dab03b86eea-xtables-lock\") pod \"kube-proxy-7g7mc\" (UID: \"854b3f29-4cd6-4527-b345-4dab03b86eea\") " pod="kube-system/kube-proxy-7g7mc" Jul 2 07:49:40.441197 kubelet[2083]: I0702 07:49:40.440950 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-cgroup\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441197 kubelet[2083]: I0702 07:49:40.440992 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/854b3f29-4cd6-4527-b345-4dab03b86eea-kube-proxy\") pod \"kube-proxy-7g7mc\" (UID: \"854b3f29-4cd6-4527-b345-4dab03b86eea\") " pod="kube-system/kube-proxy-7g7mc" Jul 2 07:49:40.441197 kubelet[2083]: I0702 07:49:40.441060 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-bpf-maps\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441197 kubelet[2083]: I0702 07:49:40.441118 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/222f0b80-bb4d-48ae-a0a0-384baa208d60-clustermesh-secrets\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441197 kubelet[2083]: I0702 07:49:40.441162 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474rj\" (UniqueName: \"kubernetes.io/projected/854b3f29-4cd6-4527-b345-4dab03b86eea-kube-api-access-474rj\") pod \"kube-proxy-7g7mc\" (UID: \"854b3f29-4cd6-4527-b345-4dab03b86eea\") " pod="kube-system/kube-proxy-7g7mc" Jul 2 07:49:40.441549 kubelet[2083]: I0702 07:49:40.441236 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cni-path\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441549 kubelet[2083]: I0702 07:49:40.441294 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-net\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441549 kubelet[2083]: I0702 07:49:40.441336 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hc7rh\" (UniqueName: \"kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-kube-api-access-hc7rh\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441549 kubelet[2083]: I0702 07:49:40.441396 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/854b3f29-4cd6-4527-b345-4dab03b86eea-lib-modules\") pod \"kube-proxy-7g7mc\" (UID: \"854b3f29-4cd6-4527-b345-4dab03b86eea\") " pod="kube-system/kube-proxy-7g7mc" Jul 2 07:49:40.441549 kubelet[2083]: I0702 07:49:40.441436 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-run\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441549 kubelet[2083]: I0702 07:49:40.441493 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-kernel\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441912 kubelet[2083]: I0702 07:49:40.441638 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-hubble-tls\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.441912 kubelet[2083]: I0702 07:49:40.441687 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-config-path\") pod \"cilium-psg5v\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " pod="kube-system/cilium-psg5v" Jul 2 07:49:40.536779 kubelet[2083]: I0702 07:49:40.536727 2083 topology_manager.go:215] "Topology Admit Handler" podUID="c091158d-3866-4a93-8622-ffc29817e3a2" podNamespace="kube-system" podName="cilium-operator-5cc964979-2ptgt" Jul 2 07:49:40.544879 systemd[1]: Created slice kubepods-besteffort-podc091158d_3866_4a93_8622_ffc29817e3a2.slice. Jul 2 07:49:40.600231 kubelet[2083]: E0702 07:49:40.600102 2083 projected.go:294] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Jul 2 07:49:40.600231 kubelet[2083]: E0702 07:49:40.600146 2083 projected.go:200] Error preparing data for projected volume kube-api-access-hc7rh for pod kube-system/cilium-psg5v: configmap "kube-root-ca.crt" not found Jul 2 07:49:40.600231 kubelet[2083]: E0702 07:49:40.600230 2083 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-kube-api-access-hc7rh podName:222f0b80-bb4d-48ae-a0a0-384baa208d60 nodeName:}" failed. No retries permitted until 2024-07-02 07:49:41.100201265 +0000 UTC m=+15.439693567 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hc7rh" (UniqueName: "kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-kube-api-access-hc7rh") pod "cilium-psg5v" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60") : configmap "kube-root-ca.crt" not found Jul 2 07:49:40.642896 kubelet[2083]: I0702 07:49:40.642839 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2g8f\" (UniqueName: \"kubernetes.io/projected/c091158d-3866-4a93-8622-ffc29817e3a2-kube-api-access-x2g8f\") pod \"cilium-operator-5cc964979-2ptgt\" (UID: \"c091158d-3866-4a93-8622-ffc29817e3a2\") " pod="kube-system/cilium-operator-5cc964979-2ptgt" Jul 2 07:49:40.643094 kubelet[2083]: I0702 07:49:40.642959 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c091158d-3866-4a93-8622-ffc29817e3a2-cilium-config-path\") pod \"cilium-operator-5cc964979-2ptgt\" (UID: \"c091158d-3866-4a93-8622-ffc29817e3a2\") " pod="kube-system/cilium-operator-5cc964979-2ptgt" Jul 2 07:49:40.664790 env[1224]: time="2024-07-02T07:49:40.663042462Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7g7mc,Uid:854b3f29-4cd6-4527-b345-4dab03b86eea,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:40.698826 env[1224]: time="2024-07-02T07:49:40.698374519Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:49:40.698826 env[1224]: time="2024-07-02T07:49:40.698423537Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:49:40.698826 env[1224]: time="2024-07-02T07:49:40.698440078Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:49:40.698826 env[1224]: time="2024-07-02T07:49:40.698608688Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/2ef9c7560201a01d579c78a0cd360d6e7fb6b9d8dc2a96bb06e607d50a8efe7b pid=2183 runtime=io.containerd.runc.v2 Jul 2 07:49:40.729854 systemd[1]: Started cri-containerd-2ef9c7560201a01d579c78a0cd360d6e7fb6b9d8dc2a96bb06e607d50a8efe7b.scope. Jul 2 07:49:40.795388 env[1224]: time="2024-07-02T07:49:40.795323312Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7g7mc,Uid:854b3f29-4cd6-4527-b345-4dab03b86eea,Namespace:kube-system,Attempt:0,} returns sandbox id \"2ef9c7560201a01d579c78a0cd360d6e7fb6b9d8dc2a96bb06e607d50a8efe7b\"" Jul 2 07:49:40.801658 env[1224]: time="2024-07-02T07:49:40.801585848Z" level=info msg="CreateContainer within sandbox \"2ef9c7560201a01d579c78a0cd360d6e7fb6b9d8dc2a96bb06e607d50a8efe7b\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jul 2 07:49:40.821700 env[1224]: time="2024-07-02T07:49:40.821657644Z" level=info msg="CreateContainer within sandbox \"2ef9c7560201a01d579c78a0cd360d6e7fb6b9d8dc2a96bb06e607d50a8efe7b\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"a8c487442ab892fe5678f41e81bc9f165944505bd23695d775f2dce672a7f2fe\"" Jul 2 07:49:40.822827 env[1224]: time="2024-07-02T07:49:40.822789151Z" level=info msg="StartContainer for \"a8c487442ab892fe5678f41e81bc9f165944505bd23695d775f2dce672a7f2fe\"" Jul 2 07:49:40.848137 systemd[1]: Started cri-containerd-a8c487442ab892fe5678f41e81bc9f165944505bd23695d775f2dce672a7f2fe.scope. Jul 2 07:49:40.856710 env[1224]: time="2024-07-02T07:49:40.856583479Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-2ptgt,Uid:c091158d-3866-4a93-8622-ffc29817e3a2,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:40.886259 env[1224]: time="2024-07-02T07:49:40.886171504Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:49:40.886519 env[1224]: time="2024-07-02T07:49:40.886479538Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:49:40.886679 env[1224]: time="2024-07-02T07:49:40.886643796Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:49:40.890179 env[1224]: time="2024-07-02T07:49:40.890131241Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8 pid=2248 runtime=io.containerd.runc.v2 Jul 2 07:49:40.914938 systemd[1]: Started cri-containerd-e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8.scope. Jul 2 07:49:40.929887 env[1224]: time="2024-07-02T07:49:40.929841385Z" level=info msg="StartContainer for \"a8c487442ab892fe5678f41e81bc9f165944505bd23695d775f2dce672a7f2fe\" returns successfully" Jul 2 07:49:41.017969 env[1224]: time="2024-07-02T07:49:41.017921488Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-2ptgt,Uid:c091158d-3866-4a93-8622-ffc29817e3a2,Namespace:kube-system,Attempt:0,} returns sandbox id \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\"" Jul 2 07:49:41.020734 env[1224]: time="2024-07-02T07:49:41.020693753Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jul 2 07:49:41.029311 kubelet[2083]: I0702 07:49:41.028758 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-7g7mc" podStartSLOduration=1.028690442 podStartE2EDuration="1.028690442s" podCreationTimestamp="2024-07-02 07:49:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:49:41.0284369 +0000 UTC m=+15.367929230" watchObservedRunningTime="2024-07-02 07:49:41.028690442 +0000 UTC m=+15.368182767" Jul 2 07:49:41.281363 env[1224]: time="2024-07-02T07:49:41.281298197Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-psg5v,Uid:222f0b80-bb4d-48ae-a0a0-384baa208d60,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:41.304790 env[1224]: time="2024-07-02T07:49:41.304302842Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:49:41.304790 env[1224]: time="2024-07-02T07:49:41.304386258Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:49:41.304790 env[1224]: time="2024-07-02T07:49:41.304408253Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:49:41.304790 env[1224]: time="2024-07-02T07:49:41.304631908Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753 pid=2383 runtime=io.containerd.runc.v2 Jul 2 07:49:41.327832 systemd[1]: Started cri-containerd-829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753.scope. Jul 2 07:49:41.374913 env[1224]: time="2024-07-02T07:49:41.374856484Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-psg5v,Uid:222f0b80-bb4d-48ae-a0a0-384baa208d60,Namespace:kube-system,Attempt:0,} returns sandbox id \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\"" Jul 2 07:49:41.577737 systemd[1]: run-containerd-runc-k8s.io-2ef9c7560201a01d579c78a0cd360d6e7fb6b9d8dc2a96bb06e607d50a8efe7b-runc.sC0taV.mount: Deactivated successfully. Jul 2 07:49:42.081401 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount838066867.mount: Deactivated successfully. Jul 2 07:49:42.953052 env[1224]: time="2024-07-02T07:49:42.952987179Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:42.955627 env[1224]: time="2024-07-02T07:49:42.955574796Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:42.958208 env[1224]: time="2024-07-02T07:49:42.958146488Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:42.959286 env[1224]: time="2024-07-02T07:49:42.959118563Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Jul 2 07:49:42.961716 env[1224]: time="2024-07-02T07:49:42.961683345Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jul 2 07:49:42.963511 env[1224]: time="2024-07-02T07:49:42.963467938Z" level=info msg="CreateContainer within sandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jul 2 07:49:42.983831 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount423846359.mount: Deactivated successfully. Jul 2 07:49:42.993754 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount672662416.mount: Deactivated successfully. Jul 2 07:49:42.998926 env[1224]: time="2024-07-02T07:49:42.994909538Z" level=info msg="CreateContainer within sandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\"" Jul 2 07:49:42.998926 env[1224]: time="2024-07-02T07:49:42.997004793Z" level=info msg="StartContainer for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\"" Jul 2 07:49:43.027103 systemd[1]: Started cri-containerd-c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb.scope. Jul 2 07:49:43.070899 env[1224]: time="2024-07-02T07:49:43.070849302Z" level=info msg="StartContainer for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" returns successfully" Jul 2 07:49:43.458857 systemd[1]: Started sshd@5-10.128.0.12:22-92.118.39.239:51610.service. Jul 2 07:49:44.143412 sshd[2492]: Failed password for root from 92.118.39.239 port 51610 ssh2 Jul 2 07:49:44.280155 sshd[2492]: Connection closed by authenticating user root 92.118.39.239 port 51610 [preauth] Jul 2 07:49:44.278177 systemd[1]: sshd@5-10.128.0.12:22-92.118.39.239:51610.service: Deactivated successfully. Jul 2 07:49:45.945011 kubelet[2083]: I0702 07:49:45.944973 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-5cc964979-2ptgt" podStartSLOduration=4.005299351 podStartE2EDuration="5.944918157s" podCreationTimestamp="2024-07-02 07:49:40 +0000 UTC" firstStartedPulling="2024-07-02 07:49:41.020123966 +0000 UTC m=+15.359616269" lastFinishedPulling="2024-07-02 07:49:42.959742752 +0000 UTC m=+17.299235075" observedRunningTime="2024-07-02 07:49:44.275594363 +0000 UTC m=+18.615086701" watchObservedRunningTime="2024-07-02 07:49:45.944918157 +0000 UTC m=+20.284410468" Jul 2 07:49:48.807198 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount395308878.mount: Deactivated successfully. Jul 2 07:49:52.194336 env[1224]: time="2024-07-02T07:49:52.194265909Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:52.196747 env[1224]: time="2024-07-02T07:49:52.196704716Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:52.199075 env[1224]: time="2024-07-02T07:49:52.199030478Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Jul 2 07:49:52.199918 env[1224]: time="2024-07-02T07:49:52.199872906Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Jul 2 07:49:52.202791 env[1224]: time="2024-07-02T07:49:52.202738992Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 2 07:49:52.221637 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount897020657.mount: Deactivated successfully. Jul 2 07:49:52.230789 env[1224]: time="2024-07-02T07:49:52.230726624Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\"" Jul 2 07:49:52.233101 env[1224]: time="2024-07-02T07:49:52.233024436Z" level=info msg="StartContainer for \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\"" Jul 2 07:49:52.270331 systemd[1]: Started cri-containerd-b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb.scope. Jul 2 07:49:52.312195 env[1224]: time="2024-07-02T07:49:52.312146369Z" level=info msg="StartContainer for \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\" returns successfully" Jul 2 07:49:52.324131 systemd[1]: cri-containerd-b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb.scope: Deactivated successfully. Jul 2 07:49:53.216233 systemd[1]: run-containerd-runc-k8s.io-b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb-runc.DAjgmH.mount: Deactivated successfully. Jul 2 07:49:53.216375 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb-rootfs.mount: Deactivated successfully. Jul 2 07:49:54.389204 env[1224]: time="2024-07-02T07:49:54.389126588Z" level=info msg="shim disconnected" id=b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb Jul 2 07:49:54.389204 env[1224]: time="2024-07-02T07:49:54.389206106Z" level=warning msg="cleaning up after shim disconnected" id=b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb namespace=k8s.io Jul 2 07:49:54.389204 env[1224]: time="2024-07-02T07:49:54.389222699Z" level=info msg="cleaning up dead shim" Jul 2 07:49:54.401825 env[1224]: time="2024-07-02T07:49:54.401750027Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:49:54Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2550 runtime=io.containerd.runc.v2\n" Jul 2 07:49:55.120655 env[1224]: time="2024-07-02T07:49:55.120558761Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 2 07:49:55.143480 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1530265520.mount: Deactivated successfully. Jul 2 07:49:55.153197 env[1224]: time="2024-07-02T07:49:55.153136883Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\"" Jul 2 07:49:55.160165 env[1224]: time="2024-07-02T07:49:55.160116780Z" level=info msg="StartContainer for \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\"" Jul 2 07:49:55.198894 systemd[1]: Started cri-containerd-a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82.scope. Jul 2 07:49:55.237510 env[1224]: time="2024-07-02T07:49:55.237459598Z" level=info msg="StartContainer for \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\" returns successfully" Jul 2 07:49:55.255609 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jul 2 07:49:55.256411 systemd[1]: Stopped systemd-sysctl.service. Jul 2 07:49:55.256628 systemd[1]: Stopping systemd-sysctl.service... Jul 2 07:49:55.259732 systemd[1]: Starting systemd-sysctl.service... Jul 2 07:49:55.271621 systemd[1]: cri-containerd-a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82.scope: Deactivated successfully. Jul 2 07:49:55.273531 systemd[1]: Finished systemd-sysctl.service. Jul 2 07:49:55.302734 env[1224]: time="2024-07-02T07:49:55.302556434Z" level=info msg="shim disconnected" id=a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82 Jul 2 07:49:55.302734 env[1224]: time="2024-07-02T07:49:55.302613854Z" level=warning msg="cleaning up after shim disconnected" id=a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82 namespace=k8s.io Jul 2 07:49:55.302734 env[1224]: time="2024-07-02T07:49:55.302631053Z" level=info msg="cleaning up dead shim" Jul 2 07:49:55.314816 env[1224]: time="2024-07-02T07:49:55.314757707Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:49:55Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2613 runtime=io.containerd.runc.v2\n" Jul 2 07:49:56.140833 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82-rootfs.mount: Deactivated successfully. Jul 2 07:49:56.144268 env[1224]: time="2024-07-02T07:49:56.141067000Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 2 07:49:56.193368 env[1224]: time="2024-07-02T07:49:56.193302225Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\"" Jul 2 07:49:56.194372 env[1224]: time="2024-07-02T07:49:56.194306530Z" level=info msg="StartContainer for \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\"" Jul 2 07:49:56.235887 systemd[1]: Started cri-containerd-fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39.scope. Jul 2 07:49:56.285465 env[1224]: time="2024-07-02T07:49:56.285408576Z" level=info msg="StartContainer for \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\" returns successfully" Jul 2 07:49:56.291998 systemd[1]: cri-containerd-fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39.scope: Deactivated successfully. Jul 2 07:49:56.324436 env[1224]: time="2024-07-02T07:49:56.324376283Z" level=info msg="shim disconnected" id=fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39 Jul 2 07:49:56.324758 env[1224]: time="2024-07-02T07:49:56.324464081Z" level=warning msg="cleaning up after shim disconnected" id=fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39 namespace=k8s.io Jul 2 07:49:56.324758 env[1224]: time="2024-07-02T07:49:56.324482130Z" level=info msg="cleaning up dead shim" Jul 2 07:49:56.336312 env[1224]: time="2024-07-02T07:49:56.336245804Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:49:56Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2671 runtime=io.containerd.runc.v2\n" Jul 2 07:49:57.131631 env[1224]: time="2024-07-02T07:49:57.131574701Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 2 07:49:57.140795 systemd[1]: run-containerd-runc-k8s.io-fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39-runc.u0jNe6.mount: Deactivated successfully. Jul 2 07:49:57.140949 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39-rootfs.mount: Deactivated successfully. Jul 2 07:49:57.158973 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4178825612.mount: Deactivated successfully. Jul 2 07:49:57.161796 env[1224]: time="2024-07-02T07:49:57.161212141Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\"" Jul 2 07:49:57.165794 env[1224]: time="2024-07-02T07:49:57.164116387Z" level=info msg="StartContainer for \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\"" Jul 2 07:49:57.199635 systemd[1]: Started cri-containerd-6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114.scope. Jul 2 07:49:57.241128 systemd[1]: cri-containerd-6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114.scope: Deactivated successfully. Jul 2 07:49:57.242017 env[1224]: time="2024-07-02T07:49:57.241681467Z" level=info msg="StartContainer for \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\" returns successfully" Jul 2 07:49:57.276492 env[1224]: time="2024-07-02T07:49:57.276423887Z" level=info msg="shim disconnected" id=6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114 Jul 2 07:49:57.276492 env[1224]: time="2024-07-02T07:49:57.276484316Z" level=warning msg="cleaning up after shim disconnected" id=6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114 namespace=k8s.io Jul 2 07:49:57.276904 env[1224]: time="2024-07-02T07:49:57.276499555Z" level=info msg="cleaning up dead shim" Jul 2 07:49:57.287782 env[1224]: time="2024-07-02T07:49:57.287723950Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:49:57Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2728 runtime=io.containerd.runc.v2\n" Jul 2 07:49:58.140465 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114-rootfs.mount: Deactivated successfully. Jul 2 07:49:58.145297 env[1224]: time="2024-07-02T07:49:58.145235422Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 2 07:49:58.173832 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1086745166.mount: Deactivated successfully. Jul 2 07:49:58.177171 env[1224]: time="2024-07-02T07:49:58.177111841Z" level=info msg="CreateContainer within sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\"" Jul 2 07:49:58.181432 env[1224]: time="2024-07-02T07:49:58.181385882Z" level=info msg="StartContainer for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\"" Jul 2 07:49:58.221665 systemd[1]: Started cri-containerd-1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e.scope. Jul 2 07:49:58.266745 env[1224]: time="2024-07-02T07:49:58.266688069Z" level=info msg="StartContainer for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" returns successfully" Jul 2 07:49:58.400702 kubelet[2083]: I0702 07:49:58.400557 2083 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Jul 2 07:49:58.431729 kubelet[2083]: I0702 07:49:58.431687 2083 topology_manager.go:215] "Topology Admit Handler" podUID="5c910845-809b-49a0-9d61-da1049247bc7" podNamespace="kube-system" podName="coredns-76f75df574-c282x" Jul 2 07:49:58.439878 systemd[1]: Created slice kubepods-burstable-pod5c910845_809b_49a0_9d61_da1049247bc7.slice. Jul 2 07:49:58.449511 kubelet[2083]: I0702 07:49:58.449484 2083 topology_manager.go:215] "Topology Admit Handler" podUID="38d069a9-64fa-4ca0-b2ea-061ae6132013" podNamespace="kube-system" podName="coredns-76f75df574-95qsg" Jul 2 07:49:58.459520 systemd[1]: Created slice kubepods-burstable-pod38d069a9_64fa_4ca0_b2ea_061ae6132013.slice. Jul 2 07:49:58.584040 kubelet[2083]: I0702 07:49:58.583994 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvbgl\" (UniqueName: \"kubernetes.io/projected/5c910845-809b-49a0-9d61-da1049247bc7-kube-api-access-bvbgl\") pod \"coredns-76f75df574-c282x\" (UID: \"5c910845-809b-49a0-9d61-da1049247bc7\") " pod="kube-system/coredns-76f75df574-c282x" Jul 2 07:49:58.584269 kubelet[2083]: I0702 07:49:58.584246 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5c910845-809b-49a0-9d61-da1049247bc7-config-volume\") pod \"coredns-76f75df574-c282x\" (UID: \"5c910845-809b-49a0-9d61-da1049247bc7\") " pod="kube-system/coredns-76f75df574-c282x" Jul 2 07:49:58.584366 kubelet[2083]: I0702 07:49:58.584351 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38d069a9-64fa-4ca0-b2ea-061ae6132013-config-volume\") pod \"coredns-76f75df574-95qsg\" (UID: \"38d069a9-64fa-4ca0-b2ea-061ae6132013\") " pod="kube-system/coredns-76f75df574-95qsg" Jul 2 07:49:58.584483 kubelet[2083]: I0702 07:49:58.584451 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kl2r\" (UniqueName: \"kubernetes.io/projected/38d069a9-64fa-4ca0-b2ea-061ae6132013-kube-api-access-9kl2r\") pod \"coredns-76f75df574-95qsg\" (UID: \"38d069a9-64fa-4ca0-b2ea-061ae6132013\") " pod="kube-system/coredns-76f75df574-95qsg" Jul 2 07:49:58.747824 env[1224]: time="2024-07-02T07:49:58.747199884Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-c282x,Uid:5c910845-809b-49a0-9d61-da1049247bc7,Namespace:kube-system,Attempt:0,}" Jul 2 07:49:58.765504 env[1224]: time="2024-07-02T07:49:58.765448540Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-95qsg,Uid:38d069a9-64fa-4ca0-b2ea-061ae6132013,Namespace:kube-system,Attempt:0,}" Jul 2 07:50:00.582555 systemd-networkd[1022]: cilium_host: Link UP Jul 2 07:50:00.585853 systemd-networkd[1022]: cilium_net: Link UP Jul 2 07:50:00.590792 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Jul 2 07:50:00.591957 systemd-networkd[1022]: cilium_net: Gained carrier Jul 2 07:50:00.599807 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Jul 2 07:50:00.599986 systemd-networkd[1022]: cilium_host: Gained carrier Jul 2 07:50:00.747308 systemd-networkd[1022]: cilium_vxlan: Link UP Jul 2 07:50:00.747325 systemd-networkd[1022]: cilium_vxlan: Gained carrier Jul 2 07:50:00.913328 systemd-networkd[1022]: cilium_net: Gained IPv6LL Jul 2 07:50:01.018790 kernel: NET: Registered PF_ALG protocol family Jul 2 07:50:01.418375 systemd-networkd[1022]: cilium_host: Gained IPv6LL Jul 2 07:50:01.865370 systemd-networkd[1022]: cilium_vxlan: Gained IPv6LL Jul 2 07:50:01.876620 systemd-networkd[1022]: lxc_health: Link UP Jul 2 07:50:01.921806 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Jul 2 07:50:01.922361 systemd-networkd[1022]: lxc_health: Gained carrier Jul 2 07:50:02.301191 systemd-networkd[1022]: lxc95e389b83314: Link UP Jul 2 07:50:02.310786 kernel: eth0: renamed from tmp512de Jul 2 07:50:02.329610 systemd-networkd[1022]: lxc87e2f26c4623: Link UP Jul 2 07:50:02.345795 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc95e389b83314: link becomes ready Jul 2 07:50:02.347510 systemd-networkd[1022]: lxc95e389b83314: Gained carrier Jul 2 07:50:02.354822 kernel: eth0: renamed from tmp248a9 Jul 2 07:50:02.374839 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc87e2f26c4623: link becomes ready Jul 2 07:50:02.374457 systemd-networkd[1022]: lxc87e2f26c4623: Gained carrier Jul 2 07:50:03.145611 systemd-networkd[1022]: lxc_health: Gained IPv6LL Jul 2 07:50:03.330151 kubelet[2083]: I0702 07:50:03.330088 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-psg5v" podStartSLOduration=12.507451607 podStartE2EDuration="23.330031266s" podCreationTimestamp="2024-07-02 07:49:40 +0000 UTC" firstStartedPulling="2024-07-02 07:49:41.377752294 +0000 UTC m=+15.717244608" lastFinishedPulling="2024-07-02 07:49:52.200331956 +0000 UTC m=+26.539824267" observedRunningTime="2024-07-02 07:49:59.181706067 +0000 UTC m=+33.521198409" watchObservedRunningTime="2024-07-02 07:50:03.330031266 +0000 UTC m=+37.669523590" Jul 2 07:50:03.401511 systemd-networkd[1022]: lxc87e2f26c4623: Gained IPv6LL Jul 2 07:50:04.361593 systemd-networkd[1022]: lxc95e389b83314: Gained IPv6LL Jul 2 07:50:07.351132 env[1224]: time="2024-07-02T07:50:07.351054150Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:50:07.351882 env[1224]: time="2024-07-02T07:50:07.351840910Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:50:07.352076 env[1224]: time="2024-07-02T07:50:07.351867272Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:50:07.352470 env[1224]: time="2024-07-02T07:50:07.352393073Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/512de6df2de033e96a1e30737594da4b953a6885e3b8b4554f0966b8dfb5fa1c pid=3262 runtime=io.containerd.runc.v2 Jul 2 07:50:07.368012 env[1224]: time="2024-07-02T07:50:07.367879159Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:50:07.368269 env[1224]: time="2024-07-02T07:50:07.368210026Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:50:07.368483 env[1224]: time="2024-07-02T07:50:07.368430619Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:50:07.368968 env[1224]: time="2024-07-02T07:50:07.368907975Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/248a90d323b6f1d254f1bba643efb1496790013ee960483211f4061960c5e70b pid=3279 runtime=io.containerd.runc.v2 Jul 2 07:50:07.393605 systemd[1]: Started cri-containerd-248a90d323b6f1d254f1bba643efb1496790013ee960483211f4061960c5e70b.scope. Jul 2 07:50:07.436987 systemd[1]: Started cri-containerd-512de6df2de033e96a1e30737594da4b953a6885e3b8b4554f0966b8dfb5fa1c.scope. Jul 2 07:50:07.508874 env[1224]: time="2024-07-02T07:50:07.508818902Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-c282x,Uid:5c910845-809b-49a0-9d61-da1049247bc7,Namespace:kube-system,Attempt:0,} returns sandbox id \"512de6df2de033e96a1e30737594da4b953a6885e3b8b4554f0966b8dfb5fa1c\"" Jul 2 07:50:07.513136 env[1224]: time="2024-07-02T07:50:07.513091518Z" level=info msg="CreateContainer within sandbox \"512de6df2de033e96a1e30737594da4b953a6885e3b8b4554f0966b8dfb5fa1c\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 2 07:50:07.535391 env[1224]: time="2024-07-02T07:50:07.535340673Z" level=info msg="CreateContainer within sandbox \"512de6df2de033e96a1e30737594da4b953a6885e3b8b4554f0966b8dfb5fa1c\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"eba13d3de8a4e3cad7999d9260e8bcddf0683f0b412b45dd4f87d7125406baed\"" Jul 2 07:50:07.536380 env[1224]: time="2024-07-02T07:50:07.536321945Z" level=info msg="StartContainer for \"eba13d3de8a4e3cad7999d9260e8bcddf0683f0b412b45dd4f87d7125406baed\"" Jul 2 07:50:07.592974 env[1224]: time="2024-07-02T07:50:07.592916357Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-95qsg,Uid:38d069a9-64fa-4ca0-b2ea-061ae6132013,Namespace:kube-system,Attempt:0,} returns sandbox id \"248a90d323b6f1d254f1bba643efb1496790013ee960483211f4061960c5e70b\"" Jul 2 07:50:07.599221 env[1224]: time="2024-07-02T07:50:07.599153133Z" level=info msg="CreateContainer within sandbox \"248a90d323b6f1d254f1bba643efb1496790013ee960483211f4061960c5e70b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jul 2 07:50:07.614916 systemd[1]: Started cri-containerd-eba13d3de8a4e3cad7999d9260e8bcddf0683f0b412b45dd4f87d7125406baed.scope. Jul 2 07:50:07.630264 env[1224]: time="2024-07-02T07:50:07.630204081Z" level=info msg="CreateContainer within sandbox \"248a90d323b6f1d254f1bba643efb1496790013ee960483211f4061960c5e70b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"eaf67983bf9031c15c010d3d58657de66fce82b5dc3eb5fae7484ae90e18e28c\"" Jul 2 07:50:07.631560 env[1224]: time="2024-07-02T07:50:07.631510824Z" level=info msg="StartContainer for \"eaf67983bf9031c15c010d3d58657de66fce82b5dc3eb5fae7484ae90e18e28c\"" Jul 2 07:50:07.678914 systemd[1]: Started cri-containerd-eaf67983bf9031c15c010d3d58657de66fce82b5dc3eb5fae7484ae90e18e28c.scope. Jul 2 07:50:07.699720 env[1224]: time="2024-07-02T07:50:07.699658217Z" level=info msg="StartContainer for \"eba13d3de8a4e3cad7999d9260e8bcddf0683f0b412b45dd4f87d7125406baed\" returns successfully" Jul 2 07:50:07.766246 env[1224]: time="2024-07-02T07:50:07.766158094Z" level=info msg="StartContainer for \"eaf67983bf9031c15c010d3d58657de66fce82b5dc3eb5fae7484ae90e18e28c\" returns successfully" Jul 2 07:50:08.190257 kubelet[2083]: I0702 07:50:08.190216 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-c282x" podStartSLOduration=28.190166497 podStartE2EDuration="28.190166497s" podCreationTimestamp="2024-07-02 07:49:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:50:08.188321796 +0000 UTC m=+42.527814120" watchObservedRunningTime="2024-07-02 07:50:08.190166497 +0000 UTC m=+42.529658821" Jul 2 07:50:23.495465 systemd[1]: Started sshd@6-10.128.0.12:22-147.75.109.163:33874.service. Jul 2 07:50:23.783673 sshd[3426]: Accepted publickey for core from 147.75.109.163 port 33874 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:23.785748 sshd[3426]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:23.793672 systemd[1]: Started session-6.scope. Jul 2 07:50:23.795003 systemd-logind[1211]: New session 6 of user core. Jul 2 07:50:24.087965 sshd[3426]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:24.092600 systemd[1]: sshd@6-10.128.0.12:22-147.75.109.163:33874.service: Deactivated successfully. Jul 2 07:50:24.093835 systemd[1]: session-6.scope: Deactivated successfully. Jul 2 07:50:24.095046 systemd-logind[1211]: Session 6 logged out. Waiting for processes to exit. Jul 2 07:50:24.096468 systemd-logind[1211]: Removed session 6. Jul 2 07:50:29.134712 systemd[1]: Started sshd@7-10.128.0.12:22-147.75.109.163:33876.service. Jul 2 07:50:29.421744 sshd[3441]: Accepted publickey for core from 147.75.109.163 port 33876 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:29.423821 sshd[3441]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:29.430732 systemd[1]: Started session-7.scope. Jul 2 07:50:29.431548 systemd-logind[1211]: New session 7 of user core. Jul 2 07:50:29.705002 sshd[3441]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:29.709691 systemd[1]: sshd@7-10.128.0.12:22-147.75.109.163:33876.service: Deactivated successfully. Jul 2 07:50:29.710911 systemd[1]: session-7.scope: Deactivated successfully. Jul 2 07:50:29.712187 systemd-logind[1211]: Session 7 logged out. Waiting for processes to exit. Jul 2 07:50:29.713571 systemd-logind[1211]: Removed session 7. Jul 2 07:50:34.751758 systemd[1]: Started sshd@8-10.128.0.12:22-147.75.109.163:52876.service. Jul 2 07:50:35.038337 sshd[3454]: Accepted publickey for core from 147.75.109.163 port 52876 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:35.040734 sshd[3454]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:35.048031 systemd[1]: Started session-8.scope. Jul 2 07:50:35.048854 systemd-logind[1211]: New session 8 of user core. Jul 2 07:50:35.334955 sshd[3454]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:35.339825 systemd[1]: sshd@8-10.128.0.12:22-147.75.109.163:52876.service: Deactivated successfully. Jul 2 07:50:35.340982 systemd[1]: session-8.scope: Deactivated successfully. Jul 2 07:50:35.341907 systemd-logind[1211]: Session 8 logged out. Waiting for processes to exit. Jul 2 07:50:35.343281 systemd-logind[1211]: Removed session 8. Jul 2 07:50:40.383726 systemd[1]: Started sshd@9-10.128.0.12:22-147.75.109.163:52878.service. Jul 2 07:50:40.676981 sshd[3467]: Accepted publickey for core from 147.75.109.163 port 52878 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:40.679530 sshd[3467]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:40.686842 systemd-logind[1211]: New session 9 of user core. Jul 2 07:50:40.687296 systemd[1]: Started session-9.scope. Jul 2 07:50:40.975865 sshd[3467]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:40.980864 systemd[1]: sshd@9-10.128.0.12:22-147.75.109.163:52878.service: Deactivated successfully. Jul 2 07:50:40.982013 systemd[1]: session-9.scope: Deactivated successfully. Jul 2 07:50:40.982987 systemd-logind[1211]: Session 9 logged out. Waiting for processes to exit. Jul 2 07:50:40.984245 systemd-logind[1211]: Removed session 9. Jul 2 07:50:41.022914 systemd[1]: Started sshd@10-10.128.0.12:22-147.75.109.163:52884.service. Jul 2 07:50:41.316017 sshd[3479]: Accepted publickey for core from 147.75.109.163 port 52884 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:41.318238 sshd[3479]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:41.325179 systemd[1]: Started session-10.scope. Jul 2 07:50:41.326072 systemd-logind[1211]: New session 10 of user core. Jul 2 07:50:41.646578 sshd[3479]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:41.651482 systemd[1]: sshd@10-10.128.0.12:22-147.75.109.163:52884.service: Deactivated successfully. Jul 2 07:50:41.652725 systemd[1]: session-10.scope: Deactivated successfully. Jul 2 07:50:41.653737 systemd-logind[1211]: Session 10 logged out. Waiting for processes to exit. Jul 2 07:50:41.655213 systemd-logind[1211]: Removed session 10. Jul 2 07:50:41.692811 systemd[1]: Started sshd@11-10.128.0.12:22-147.75.109.163:52898.service. Jul 2 07:50:41.982978 sshd[3490]: Accepted publickey for core from 147.75.109.163 port 52898 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:41.985581 sshd[3490]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:41.993439 systemd[1]: Started session-11.scope. Jul 2 07:50:41.994317 systemd-logind[1211]: New session 11 of user core. Jul 2 07:50:42.271608 sshd[3490]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:42.276030 systemd[1]: sshd@11-10.128.0.12:22-147.75.109.163:52898.service: Deactivated successfully. Jul 2 07:50:42.277111 systemd[1]: session-11.scope: Deactivated successfully. Jul 2 07:50:42.278066 systemd-logind[1211]: Session 11 logged out. Waiting for processes to exit. Jul 2 07:50:42.279452 systemd-logind[1211]: Removed session 11. Jul 2 07:50:47.318333 systemd[1]: Started sshd@12-10.128.0.12:22-147.75.109.163:47138.service. Jul 2 07:50:47.605962 sshd[3503]: Accepted publickey for core from 147.75.109.163 port 47138 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:47.607992 sshd[3503]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:47.615703 systemd[1]: Started session-12.scope. Jul 2 07:50:47.616687 systemd-logind[1211]: New session 12 of user core. Jul 2 07:50:47.894116 sshd[3503]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:47.898472 systemd[1]: sshd@12-10.128.0.12:22-147.75.109.163:47138.service: Deactivated successfully. Jul 2 07:50:47.899690 systemd[1]: session-12.scope: Deactivated successfully. Jul 2 07:50:47.901098 systemd-logind[1211]: Session 12 logged out. Waiting for processes to exit. Jul 2 07:50:47.902459 systemd-logind[1211]: Removed session 12. Jul 2 07:50:52.940680 systemd[1]: Started sshd@13-10.128.0.12:22-147.75.109.163:45096.service. Jul 2 07:50:53.226516 sshd[3515]: Accepted publickey for core from 147.75.109.163 port 45096 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:53.228319 sshd[3515]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:53.235429 systemd[1]: Started session-13.scope. Jul 2 07:50:53.236299 systemd-logind[1211]: New session 13 of user core. Jul 2 07:50:53.523141 sshd[3515]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:53.528259 systemd[1]: sshd@13-10.128.0.12:22-147.75.109.163:45096.service: Deactivated successfully. Jul 2 07:50:53.529459 systemd[1]: session-13.scope: Deactivated successfully. Jul 2 07:50:53.530437 systemd-logind[1211]: Session 13 logged out. Waiting for processes to exit. Jul 2 07:50:53.531880 systemd-logind[1211]: Removed session 13. Jul 2 07:50:53.571050 systemd[1]: Started sshd@14-10.128.0.12:22-147.75.109.163:45110.service. Jul 2 07:50:53.862365 sshd[3527]: Accepted publickey for core from 147.75.109.163 port 45110 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:53.865192 sshd[3527]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:53.872497 systemd[1]: Started session-14.scope. Jul 2 07:50:53.873428 systemd-logind[1211]: New session 14 of user core. Jul 2 07:50:54.222189 sshd[3527]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:54.227228 systemd[1]: sshd@14-10.128.0.12:22-147.75.109.163:45110.service: Deactivated successfully. Jul 2 07:50:54.228416 systemd[1]: session-14.scope: Deactivated successfully. Jul 2 07:50:54.229373 systemd-logind[1211]: Session 14 logged out. Waiting for processes to exit. Jul 2 07:50:54.230801 systemd-logind[1211]: Removed session 14. Jul 2 07:50:54.268701 systemd[1]: Started sshd@15-10.128.0.12:22-147.75.109.163:45118.service. Jul 2 07:50:54.558710 sshd[3537]: Accepted publickey for core from 147.75.109.163 port 45118 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:54.560958 sshd[3537]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:54.568584 systemd[1]: Started session-15.scope. Jul 2 07:50:54.569450 systemd-logind[1211]: New session 15 of user core. Jul 2 07:50:56.271273 sshd[3537]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:56.277098 systemd[1]: sshd@15-10.128.0.12:22-147.75.109.163:45118.service: Deactivated successfully. Jul 2 07:50:56.278308 systemd[1]: session-15.scope: Deactivated successfully. Jul 2 07:50:56.279553 systemd-logind[1211]: Session 15 logged out. Waiting for processes to exit. Jul 2 07:50:56.282349 systemd-logind[1211]: Removed session 15. Jul 2 07:50:56.312878 systemd[1]: Started sshd@16-10.128.0.12:22-147.75.109.163:45122.service. Jul 2 07:50:56.598440 sshd[3554]: Accepted publickey for core from 147.75.109.163 port 45122 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:56.600018 sshd[3554]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:56.606849 systemd-logind[1211]: New session 16 of user core. Jul 2 07:50:56.607119 systemd[1]: Started session-16.scope. Jul 2 07:50:57.031229 sshd[3554]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:57.035929 systemd[1]: sshd@16-10.128.0.12:22-147.75.109.163:45122.service: Deactivated successfully. Jul 2 07:50:57.037151 systemd[1]: session-16.scope: Deactivated successfully. Jul 2 07:50:57.038281 systemd-logind[1211]: Session 16 logged out. Waiting for processes to exit. Jul 2 07:50:57.039865 systemd-logind[1211]: Removed session 16. Jul 2 07:50:57.078246 systemd[1]: Started sshd@17-10.128.0.12:22-147.75.109.163:45124.service. Jul 2 07:50:57.366620 sshd[3564]: Accepted publickey for core from 147.75.109.163 port 45124 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:50:57.369215 sshd[3564]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:50:57.376955 systemd[1]: Started session-17.scope. Jul 2 07:50:57.378125 systemd-logind[1211]: New session 17 of user core. Jul 2 07:50:57.648735 sshd[3564]: pam_unix(sshd:session): session closed for user core Jul 2 07:50:57.653424 systemd-logind[1211]: Session 17 logged out. Waiting for processes to exit. Jul 2 07:50:57.653702 systemd[1]: sshd@17-10.128.0.12:22-147.75.109.163:45124.service: Deactivated successfully. Jul 2 07:50:57.654898 systemd[1]: session-17.scope: Deactivated successfully. Jul 2 07:50:57.656113 systemd-logind[1211]: Removed session 17. Jul 2 07:51:02.696635 systemd[1]: Started sshd@18-10.128.0.12:22-147.75.109.163:49138.service. Jul 2 07:51:02.985235 sshd[3579]: Accepted publickey for core from 147.75.109.163 port 49138 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:02.987743 sshd[3579]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:02.995366 systemd[1]: Started session-18.scope. Jul 2 07:51:02.995992 systemd-logind[1211]: New session 18 of user core. Jul 2 07:51:03.268613 sshd[3579]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:03.272880 systemd-logind[1211]: Session 18 logged out. Waiting for processes to exit. Jul 2 07:51:03.273329 systemd[1]: sshd@18-10.128.0.12:22-147.75.109.163:49138.service: Deactivated successfully. Jul 2 07:51:03.274490 systemd[1]: session-18.scope: Deactivated successfully. Jul 2 07:51:03.275850 systemd-logind[1211]: Removed session 18. Jul 2 07:51:08.317422 systemd[1]: Started sshd@19-10.128.0.12:22-147.75.109.163:49152.service. Jul 2 07:51:08.606636 sshd[3591]: Accepted publickey for core from 147.75.109.163 port 49152 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:08.608431 sshd[3591]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:08.615285 systemd[1]: Started session-19.scope. Jul 2 07:51:08.616174 systemd-logind[1211]: New session 19 of user core. Jul 2 07:51:08.892724 sshd[3591]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:08.897363 systemd[1]: sshd@19-10.128.0.12:22-147.75.109.163:49152.service: Deactivated successfully. Jul 2 07:51:08.898543 systemd[1]: session-19.scope: Deactivated successfully. Jul 2 07:51:08.899616 systemd-logind[1211]: Session 19 logged out. Waiting for processes to exit. Jul 2 07:51:08.901247 systemd-logind[1211]: Removed session 19. Jul 2 07:51:09.817632 update_engine[1212]: I0702 07:51:09.817566 1212 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Jul 2 07:51:09.817632 update_engine[1212]: I0702 07:51:09.817616 1212 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Jul 2 07:51:09.818719 update_engine[1212]: I0702 07:51:09.818675 1212 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Jul 2 07:51:09.819376 update_engine[1212]: I0702 07:51:09.819331 1212 omaha_request_params.cc:62] Current group set to lts Jul 2 07:51:09.819584 update_engine[1212]: I0702 07:51:09.819561 1212 update_attempter.cc:499] Already updated boot flags. Skipping. Jul 2 07:51:09.819584 update_engine[1212]: I0702 07:51:09.819581 1212 update_attempter.cc:643] Scheduling an action processor start. Jul 2 07:51:09.819724 update_engine[1212]: I0702 07:51:09.819603 1212 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Jul 2 07:51:09.819724 update_engine[1212]: I0702 07:51:09.819650 1212 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Jul 2 07:51:09.819870 update_engine[1212]: I0702 07:51:09.819734 1212 omaha_request_action.cc:270] Posting an Omaha request to disabled Jul 2 07:51:09.819870 update_engine[1212]: I0702 07:51:09.819743 1212 omaha_request_action.cc:271] Request: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: Jul 2 07:51:09.819870 update_engine[1212]: I0702 07:51:09.819751 1212 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jul 2 07:51:09.820928 locksmithd[1256]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Jul 2 07:51:09.821861 update_engine[1212]: I0702 07:51:09.821478 1212 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jul 2 07:51:09.821861 update_engine[1212]: I0702 07:51:09.821723 1212 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jul 2 07:51:09.853892 update_engine[1212]: E0702 07:51:09.853843 1212 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jul 2 07:51:09.854052 update_engine[1212]: I0702 07:51:09.854019 1212 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Jul 2 07:51:13.939557 systemd[1]: Started sshd@20-10.128.0.12:22-147.75.109.163:41144.service. Jul 2 07:51:14.226677 sshd[3605]: Accepted publickey for core from 147.75.109.163 port 41144 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:14.228460 sshd[3605]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:14.234193 systemd-logind[1211]: New session 20 of user core. Jul 2 07:51:14.235440 systemd[1]: Started session-20.scope. Jul 2 07:51:14.511252 sshd[3605]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:14.516294 systemd-logind[1211]: Session 20 logged out. Waiting for processes to exit. Jul 2 07:51:14.516583 systemd[1]: sshd@20-10.128.0.12:22-147.75.109.163:41144.service: Deactivated successfully. Jul 2 07:51:14.517818 systemd[1]: session-20.scope: Deactivated successfully. Jul 2 07:51:14.519103 systemd-logind[1211]: Removed session 20. Jul 2 07:51:14.558299 systemd[1]: Started sshd@21-10.128.0.12:22-147.75.109.163:41152.service. Jul 2 07:51:14.847992 sshd[3617]: Accepted publickey for core from 147.75.109.163 port 41152 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:14.850955 sshd[3617]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:14.861799 systemd[1]: Started session-21.scope. Jul 2 07:51:14.862463 systemd-logind[1211]: New session 21 of user core. Jul 2 07:51:16.627789 kubelet[2083]: I0702 07:51:16.627728 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-95qsg" podStartSLOduration=96.627646964 podStartE2EDuration="1m36.627646964s" podCreationTimestamp="2024-07-02 07:49:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:50:08.220887006 +0000 UTC m=+42.560379333" watchObservedRunningTime="2024-07-02 07:51:16.627646964 +0000 UTC m=+110.967139288" Jul 2 07:51:16.644455 env[1224]: time="2024-07-02T07:51:16.644402543Z" level=info msg="StopContainer for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" with timeout 30 (s)" Jul 2 07:51:16.645031 env[1224]: time="2024-07-02T07:51:16.644948276Z" level=info msg="Stop container \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" with signal terminated" Jul 2 07:51:16.671940 systemd[1]: run-containerd-runc-k8s.io-1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e-runc.vdSfA5.mount: Deactivated successfully. Jul 2 07:51:16.676756 systemd[1]: cri-containerd-c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb.scope: Deactivated successfully. Jul 2 07:51:16.704019 env[1224]: time="2024-07-02T07:51:16.703952441Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jul 2 07:51:16.713101 env[1224]: time="2024-07-02T07:51:16.713056329Z" level=info msg="StopContainer for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" with timeout 2 (s)" Jul 2 07:51:16.713631 env[1224]: time="2024-07-02T07:51:16.713582419Z" level=info msg="Stop container \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" with signal terminated" Jul 2 07:51:16.721013 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb-rootfs.mount: Deactivated successfully. Jul 2 07:51:16.728173 systemd-networkd[1022]: lxc_health: Link DOWN Jul 2 07:51:16.728185 systemd-networkd[1022]: lxc_health: Lost carrier Jul 2 07:51:16.750283 env[1224]: time="2024-07-02T07:51:16.750221138Z" level=info msg="shim disconnected" id=c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb Jul 2 07:51:16.750555 env[1224]: time="2024-07-02T07:51:16.750294874Z" level=warning msg="cleaning up after shim disconnected" id=c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb namespace=k8s.io Jul 2 07:51:16.750555 env[1224]: time="2024-07-02T07:51:16.750313781Z" level=info msg="cleaning up dead shim" Jul 2 07:51:16.751558 systemd[1]: cri-containerd-1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e.scope: Deactivated successfully. Jul 2 07:51:16.751902 systemd[1]: cri-containerd-1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e.scope: Consumed 9.342s CPU time. Jul 2 07:51:16.766167 env[1224]: time="2024-07-02T07:51:16.766120520Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:16Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3673 runtime=io.containerd.runc.v2\n" Jul 2 07:51:16.768583 env[1224]: time="2024-07-02T07:51:16.768537333Z" level=info msg="StopContainer for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" returns successfully" Jul 2 07:51:16.769963 env[1224]: time="2024-07-02T07:51:16.769912618Z" level=info msg="StopPodSandbox for \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\"" Jul 2 07:51:16.777065 env[1224]: time="2024-07-02T07:51:16.769991633Z" level=info msg="Container to stop \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:16.773663 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8-shm.mount: Deactivated successfully. Jul 2 07:51:16.793391 systemd[1]: cri-containerd-e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8.scope: Deactivated successfully. Jul 2 07:51:16.812240 env[1224]: time="2024-07-02T07:51:16.812168733Z" level=info msg="shim disconnected" id=1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e Jul 2 07:51:16.812673 env[1224]: time="2024-07-02T07:51:16.812627458Z" level=warning msg="cleaning up after shim disconnected" id=1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e namespace=k8s.io Jul 2 07:51:16.812860 env[1224]: time="2024-07-02T07:51:16.812833520Z" level=info msg="cleaning up dead shim" Jul 2 07:51:16.831452 env[1224]: time="2024-07-02T07:51:16.831413771Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:16Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3711 runtime=io.containerd.runc.v2\n" Jul 2 07:51:16.834971 env[1224]: time="2024-07-02T07:51:16.834861559Z" level=info msg="StopContainer for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" returns successfully" Jul 2 07:51:16.836142 env[1224]: time="2024-07-02T07:51:16.836107670Z" level=info msg="StopPodSandbox for \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\"" Jul 2 07:51:16.836570 env[1224]: time="2024-07-02T07:51:16.836537451Z" level=info msg="Container to stop \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:16.836808 env[1224]: time="2024-07-02T07:51:16.836734629Z" level=info msg="Container to stop \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:16.837241 env[1224]: time="2024-07-02T07:51:16.836826868Z" level=info msg="Container to stop \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:16.837241 env[1224]: time="2024-07-02T07:51:16.836862633Z" level=info msg="Container to stop \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:16.837241 env[1224]: time="2024-07-02T07:51:16.836896247Z" level=info msg="Container to stop \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:16.840903 env[1224]: time="2024-07-02T07:51:16.840725239Z" level=info msg="shim disconnected" id=e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8 Jul 2 07:51:16.840903 env[1224]: time="2024-07-02T07:51:16.840803451Z" level=warning msg="cleaning up after shim disconnected" id=e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8 namespace=k8s.io Jul 2 07:51:16.840903 env[1224]: time="2024-07-02T07:51:16.840820212Z" level=info msg="cleaning up dead shim" Jul 2 07:51:16.864700 systemd[1]: cri-containerd-829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753.scope: Deactivated successfully. Jul 2 07:51:16.878592 env[1224]: time="2024-07-02T07:51:16.878398717Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:16Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3733 runtime=io.containerd.runc.v2\n" Jul 2 07:51:16.879310 env[1224]: time="2024-07-02T07:51:16.879267769Z" level=info msg="TearDown network for sandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" successfully" Jul 2 07:51:16.879504 env[1224]: time="2024-07-02T07:51:16.879472784Z" level=info msg="StopPodSandbox for \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" returns successfully" Jul 2 07:51:16.910505 env[1224]: time="2024-07-02T07:51:16.910433704Z" level=info msg="shim disconnected" id=829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753 Jul 2 07:51:16.910505 env[1224]: time="2024-07-02T07:51:16.910500183Z" level=warning msg="cleaning up after shim disconnected" id=829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753 namespace=k8s.io Jul 2 07:51:16.910505 env[1224]: time="2024-07-02T07:51:16.910515224Z" level=info msg="cleaning up dead shim" Jul 2 07:51:16.922805 env[1224]: time="2024-07-02T07:51:16.922747920Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:16Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3765 runtime=io.containerd.runc.v2\n" Jul 2 07:51:16.923324 env[1224]: time="2024-07-02T07:51:16.923283541Z" level=info msg="TearDown network for sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" successfully" Jul 2 07:51:16.923503 env[1224]: time="2024-07-02T07:51:16.923456032Z" level=info msg="StopPodSandbox for \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" returns successfully" Jul 2 07:51:16.978232 kubelet[2083]: I0702 07:51:16.978188 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c091158d-3866-4a93-8622-ffc29817e3a2-cilium-config-path\") pod \"c091158d-3866-4a93-8622-ffc29817e3a2\" (UID: \"c091158d-3866-4a93-8622-ffc29817e3a2\") " Jul 2 07:51:16.978457 kubelet[2083]: I0702 07:51:16.978319 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2g8f\" (UniqueName: \"kubernetes.io/projected/c091158d-3866-4a93-8622-ffc29817e3a2-kube-api-access-x2g8f\") pod \"c091158d-3866-4a93-8622-ffc29817e3a2\" (UID: \"c091158d-3866-4a93-8622-ffc29817e3a2\") " Jul 2 07:51:16.985098 kubelet[2083]: I0702 07:51:16.985053 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c091158d-3866-4a93-8622-ffc29817e3a2-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "c091158d-3866-4a93-8622-ffc29817e3a2" (UID: "c091158d-3866-4a93-8622-ffc29817e3a2"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 2 07:51:16.986695 kubelet[2083]: I0702 07:51:16.986657 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c091158d-3866-4a93-8622-ffc29817e3a2-kube-api-access-x2g8f" (OuterVolumeSpecName: "kube-api-access-x2g8f") pod "c091158d-3866-4a93-8622-ffc29817e3a2" (UID: "c091158d-3866-4a93-8622-ffc29817e3a2"). InnerVolumeSpecName "kube-api-access-x2g8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 2 07:51:17.079428 kubelet[2083]: I0702 07:51:17.079383 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-lib-modules\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.079641 kubelet[2083]: I0702 07:51:17.079448 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-hubble-tls\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.079641 kubelet[2083]: I0702 07:51:17.079481 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-kernel\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.079641 kubelet[2083]: I0702 07:51:17.079531 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/222f0b80-bb4d-48ae-a0a0-384baa208d60-clustermesh-secrets\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.079641 kubelet[2083]: I0702 07:51:17.079564 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-config-path\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.079641 kubelet[2083]: I0702 07:51:17.079597 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hc7rh\" (UniqueName: \"kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-kube-api-access-hc7rh\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.079641 kubelet[2083]: I0702 07:51:17.079625 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-run\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080022 kubelet[2083]: I0702 07:51:17.079652 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cni-path\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080022 kubelet[2083]: I0702 07:51:17.079682 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-net\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080022 kubelet[2083]: I0702 07:51:17.079715 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-bpf-maps\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080022 kubelet[2083]: I0702 07:51:17.079744 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-xtables-lock\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080022 kubelet[2083]: I0702 07:51:17.079787 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-hostproc\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080022 kubelet[2083]: I0702 07:51:17.079818 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-etc-cni-netd\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080371 kubelet[2083]: I0702 07:51:17.079850 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-cgroup\") pod \"222f0b80-bb4d-48ae-a0a0-384baa208d60\" (UID: \"222f0b80-bb4d-48ae-a0a0-384baa208d60\") " Jul 2 07:51:17.080371 kubelet[2083]: I0702 07:51:17.079912 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/c091158d-3866-4a93-8622-ffc29817e3a2-cilium-config-path\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.080371 kubelet[2083]: I0702 07:51:17.079937 2083 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-x2g8f\" (UniqueName: \"kubernetes.io/projected/c091158d-3866-4a93-8622-ffc29817e3a2-kube-api-access-x2g8f\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.080371 kubelet[2083]: I0702 07:51:17.079984 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.080673 kubelet[2083]: I0702 07:51:17.080635 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.080893 kubelet[2083]: I0702 07:51:17.080859 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cni-path" (OuterVolumeSpecName: "cni-path") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.086378 kubelet[2083]: I0702 07:51:17.081041 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.086566 kubelet[2083]: I0702 07:51:17.081080 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.086931 kubelet[2083]: I0702 07:51:17.081104 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.087093 kubelet[2083]: I0702 07:51:17.081122 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-hostproc" (OuterVolumeSpecName: "hostproc") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.087310 kubelet[2083]: I0702 07:51:17.081141 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.087471 kubelet[2083]: I0702 07:51:17.085172 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.087610 kubelet[2083]: I0702 07:51:17.085828 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:17.087738 kubelet[2083]: I0702 07:51:17.086326 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 2 07:51:17.087927 kubelet[2083]: I0702 07:51:17.086872 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 2 07:51:17.089268 kubelet[2083]: I0702 07:51:17.089206 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-kube-api-access-hc7rh" (OuterVolumeSpecName: "kube-api-access-hc7rh") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "kube-api-access-hc7rh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 2 07:51:17.091264 kubelet[2083]: I0702 07:51:17.091212 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/222f0b80-bb4d-48ae-a0a0-384baa208d60-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "222f0b80-bb4d-48ae-a0a0-384baa208d60" (UID: "222f0b80-bb4d-48ae-a0a0-384baa208d60"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.180962 2083 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/222f0b80-bb4d-48ae-a0a0-384baa208d60-clustermesh-secrets\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.181013 2083 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-hubble-tls\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.181036 2083 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-kernel\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.181056 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-config-path\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.181077 2083 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-hc7rh\" (UniqueName: \"kubernetes.io/projected/222f0b80-bb4d-48ae-a0a0-384baa208d60-kube-api-access-hc7rh\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.181096 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-run\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.181337 kubelet[2083]: I0702 07:51:17.181116 2083 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-bpf-maps\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.182017 kubelet[2083]: I0702 07:51:17.181280 2083 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-xtables-lock\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.182017 kubelet[2083]: I0702 07:51:17.181306 2083 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cni-path\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.182914 kubelet[2083]: I0702 07:51:17.182876 2083 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-host-proc-sys-net\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.183098 kubelet[2083]: I0702 07:51:17.183081 2083 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-hostproc\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.183246 kubelet[2083]: I0702 07:51:17.183230 2083 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-etc-cni-netd\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.183435 kubelet[2083]: I0702 07:51:17.183419 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-cilium-cgroup\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.183621 kubelet[2083]: I0702 07:51:17.183605 2083 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/222f0b80-bb4d-48ae-a0a0-384baa208d60-lib-modules\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:17.348491 kubelet[2083]: I0702 07:51:17.348455 2083 scope.go:117] "RemoveContainer" containerID="1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e" Jul 2 07:51:17.350246 env[1224]: time="2024-07-02T07:51:17.350186860Z" level=info msg="RemoveContainer for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\"" Jul 2 07:51:17.358235 systemd[1]: Removed slice kubepods-burstable-pod222f0b80_bb4d_48ae_a0a0_384baa208d60.slice. Jul 2 07:51:17.358433 systemd[1]: kubepods-burstable-pod222f0b80_bb4d_48ae_a0a0_384baa208d60.slice: Consumed 9.484s CPU time. Jul 2 07:51:17.360477 env[1224]: time="2024-07-02T07:51:17.360436310Z" level=info msg="RemoveContainer for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" returns successfully" Jul 2 07:51:17.361411 kubelet[2083]: I0702 07:51:17.361383 2083 scope.go:117] "RemoveContainer" containerID="6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114" Jul 2 07:51:17.366625 env[1224]: time="2024-07-02T07:51:17.366573732Z" level=info msg="RemoveContainer for \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\"" Jul 2 07:51:17.373418 env[1224]: time="2024-07-02T07:51:17.373379878Z" level=info msg="RemoveContainer for \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\" returns successfully" Jul 2 07:51:17.376299 kubelet[2083]: I0702 07:51:17.376274 2083 scope.go:117] "RemoveContainer" containerID="fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39" Jul 2 07:51:17.377297 systemd[1]: Removed slice kubepods-besteffort-podc091158d_3866_4a93_8622_ffc29817e3a2.slice. Jul 2 07:51:17.381694 env[1224]: time="2024-07-02T07:51:17.381651287Z" level=info msg="RemoveContainer for \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\"" Jul 2 07:51:17.387951 env[1224]: time="2024-07-02T07:51:17.387831717Z" level=info msg="RemoveContainer for \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\" returns successfully" Jul 2 07:51:17.388111 kubelet[2083]: I0702 07:51:17.388081 2083 scope.go:117] "RemoveContainer" containerID="a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82" Jul 2 07:51:17.392019 env[1224]: time="2024-07-02T07:51:17.391985854Z" level=info msg="RemoveContainer for \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\"" Jul 2 07:51:17.396368 env[1224]: time="2024-07-02T07:51:17.396325980Z" level=info msg="RemoveContainer for \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\" returns successfully" Jul 2 07:51:17.396728 kubelet[2083]: I0702 07:51:17.396652 2083 scope.go:117] "RemoveContainer" containerID="b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb" Jul 2 07:51:17.398493 env[1224]: time="2024-07-02T07:51:17.398171758Z" level=info msg="RemoveContainer for \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\"" Jul 2 07:51:17.401905 env[1224]: time="2024-07-02T07:51:17.401870907Z" level=info msg="RemoveContainer for \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\" returns successfully" Jul 2 07:51:17.402260 kubelet[2083]: I0702 07:51:17.402210 2083 scope.go:117] "RemoveContainer" containerID="1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e" Jul 2 07:51:17.402613 env[1224]: time="2024-07-02T07:51:17.402517281Z" level=error msg="ContainerStatus for \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\": not found" Jul 2 07:51:17.402796 kubelet[2083]: E0702 07:51:17.402752 2083 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\": not found" containerID="1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e" Jul 2 07:51:17.402917 kubelet[2083]: I0702 07:51:17.402899 2083 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e"} err="failed to get container status \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\": rpc error: code = NotFound desc = an error occurred when try to find container \"1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e\": not found" Jul 2 07:51:17.402982 kubelet[2083]: I0702 07:51:17.402922 2083 scope.go:117] "RemoveContainer" containerID="6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114" Jul 2 07:51:17.403214 env[1224]: time="2024-07-02T07:51:17.403143894Z" level=error msg="ContainerStatus for \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\": not found" Jul 2 07:51:17.403396 kubelet[2083]: E0702 07:51:17.403375 2083 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\": not found" containerID="6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114" Jul 2 07:51:17.403514 kubelet[2083]: I0702 07:51:17.403420 2083 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114"} err="failed to get container status \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\": rpc error: code = NotFound desc = an error occurred when try to find container \"6aaace62ee0aebdb6e659b044006ac2bb59fc1f80891ac22d5dca2a06fce7114\": not found" Jul 2 07:51:17.403514 kubelet[2083]: I0702 07:51:17.403437 2083 scope.go:117] "RemoveContainer" containerID="fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39" Jul 2 07:51:17.403783 env[1224]: time="2024-07-02T07:51:17.403695478Z" level=error msg="ContainerStatus for \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\": not found" Jul 2 07:51:17.403927 kubelet[2083]: E0702 07:51:17.403908 2083 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\": not found" containerID="fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39" Jul 2 07:51:17.404027 kubelet[2083]: I0702 07:51:17.403947 2083 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39"} err="failed to get container status \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\": rpc error: code = NotFound desc = an error occurred when try to find container \"fefbf04825f8ccaccf529934a7b802fdbe5d83f1373c1cb7789006adaf1c9f39\": not found" Jul 2 07:51:17.404027 kubelet[2083]: I0702 07:51:17.403964 2083 scope.go:117] "RemoveContainer" containerID="a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82" Jul 2 07:51:17.404319 env[1224]: time="2024-07-02T07:51:17.404187696Z" level=error msg="ContainerStatus for \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\": not found" Jul 2 07:51:17.404623 kubelet[2083]: E0702 07:51:17.404601 2083 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\": not found" containerID="a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82" Jul 2 07:51:17.404820 kubelet[2083]: I0702 07:51:17.404797 2083 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82"} err="failed to get container status \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\": rpc error: code = NotFound desc = an error occurred when try to find container \"a8e982ccb6da16dce9968113ff12f31ac211ee8885b86d73c00713e75a2d3a82\": not found" Jul 2 07:51:17.404820 kubelet[2083]: I0702 07:51:17.404823 2083 scope.go:117] "RemoveContainer" containerID="b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb" Jul 2 07:51:17.405223 env[1224]: time="2024-07-02T07:51:17.405136994Z" level=error msg="ContainerStatus for \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\": not found" Jul 2 07:51:17.405427 kubelet[2083]: E0702 07:51:17.405408 2083 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\": not found" containerID="b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb" Jul 2 07:51:17.405533 kubelet[2083]: I0702 07:51:17.405467 2083 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb"} err="failed to get container status \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\": rpc error: code = NotFound desc = an error occurred when try to find container \"b966b3ba09e44e8c8fc5763ea9923de888346c4120823f2819708601c68246bb\": not found" Jul 2 07:51:17.405533 kubelet[2083]: I0702 07:51:17.405494 2083 scope.go:117] "RemoveContainer" containerID="c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb" Jul 2 07:51:17.408780 env[1224]: time="2024-07-02T07:51:17.408732216Z" level=info msg="RemoveContainer for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\"" Jul 2 07:51:17.413787 env[1224]: time="2024-07-02T07:51:17.413719994Z" level=info msg="RemoveContainer for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" returns successfully" Jul 2 07:51:17.414089 kubelet[2083]: I0702 07:51:17.414061 2083 scope.go:117] "RemoveContainer" containerID="c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb" Jul 2 07:51:17.414777 env[1224]: time="2024-07-02T07:51:17.414696497Z" level=error msg="ContainerStatus for \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\": not found" Jul 2 07:51:17.415118 kubelet[2083]: E0702 07:51:17.415099 2083 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\": not found" containerID="c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb" Jul 2 07:51:17.415330 kubelet[2083]: I0702 07:51:17.415283 2083 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb"} err="failed to get container status \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\": rpc error: code = NotFound desc = an error occurred when try to find container \"c8ff42095054582436931e43e5ee18974586ff5270b571201eea1792e429effb\": not found" Jul 2 07:51:17.660873 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1fa0734014cb320879b2d9703e6aa897f3f4ce0aac0541b962f6c3a5b2d8d27e-rootfs.mount: Deactivated successfully. Jul 2 07:51:17.661030 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753-rootfs.mount: Deactivated successfully. Jul 2 07:51:17.661145 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753-shm.mount: Deactivated successfully. Jul 2 07:51:17.661263 systemd[1]: var-lib-kubelet-pods-222f0b80\x2dbb4d\x2d48ae\x2da0a0\x2d384baa208d60-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dhc7rh.mount: Deactivated successfully. Jul 2 07:51:17.661384 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8-rootfs.mount: Deactivated successfully. Jul 2 07:51:17.661507 systemd[1]: var-lib-kubelet-pods-c091158d\x2d3866\x2d4a93\x2d8622\x2dffc29817e3a2-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dx2g8f.mount: Deactivated successfully. Jul 2 07:51:17.661632 systemd[1]: var-lib-kubelet-pods-222f0b80\x2dbb4d\x2d48ae\x2da0a0\x2d384baa208d60-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jul 2 07:51:17.661742 systemd[1]: var-lib-kubelet-pods-222f0b80\x2dbb4d\x2d48ae\x2da0a0\x2d384baa208d60-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jul 2 07:51:17.931570 kubelet[2083]: I0702 07:51:17.930667 2083 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" path="/var/lib/kubelet/pods/222f0b80-bb4d-48ae-a0a0-384baa208d60/volumes" Jul 2 07:51:17.932271 kubelet[2083]: I0702 07:51:17.931845 2083 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="c091158d-3866-4a93-8622-ffc29817e3a2" path="/var/lib/kubelet/pods/c091158d-3866-4a93-8622-ffc29817e3a2/volumes" Jul 2 07:51:18.627183 sshd[3617]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:18.632344 systemd[1]: sshd@21-10.128.0.12:22-147.75.109.163:41152.service: Deactivated successfully. Jul 2 07:51:18.633435 systemd[1]: session-21.scope: Deactivated successfully. Jul 2 07:51:18.633655 systemd[1]: session-21.scope: Consumed 1.018s CPU time. Jul 2 07:51:18.634446 systemd-logind[1211]: Session 21 logged out. Waiting for processes to exit. Jul 2 07:51:18.636173 systemd-logind[1211]: Removed session 21. Jul 2 07:51:18.674942 systemd[1]: Started sshd@22-10.128.0.12:22-147.75.109.163:41166.service. Jul 2 07:51:18.973003 sshd[3784]: Accepted publickey for core from 147.75.109.163 port 41166 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:18.974706 sshd[3784]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:18.982466 systemd[1]: Started session-22.scope. Jul 2 07:51:18.983411 systemd-logind[1211]: New session 22 of user core. Jul 2 07:51:19.817467 update_engine[1212]: I0702 07:51:19.816893 1212 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jul 2 07:51:19.817467 update_engine[1212]: I0702 07:51:19.817199 1212 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jul 2 07:51:19.817467 update_engine[1212]: I0702 07:51:19.817412 1212 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jul 2 07:51:19.826179 update_engine[1212]: E0702 07:51:19.826023 1212 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jul 2 07:51:19.826179 update_engine[1212]: I0702 07:51:19.826148 1212 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Jul 2 07:51:20.184103 kubelet[2083]: I0702 07:51:20.183954 2083 topology_manager.go:215] "Topology Admit Handler" podUID="ea97d56a-8fd9-4ddf-81b9-70e038a695ee" podNamespace="kube-system" podName="cilium-t7hqf" Jul 2 07:51:20.184818 kubelet[2083]: E0702 07:51:20.184789 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" containerName="mount-cgroup" Jul 2 07:51:20.185033 kubelet[2083]: E0702 07:51:20.185014 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" containerName="apply-sysctl-overwrites" Jul 2 07:51:20.185195 kubelet[2083]: E0702 07:51:20.185177 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" containerName="mount-bpf-fs" Jul 2 07:51:20.185364 kubelet[2083]: E0702 07:51:20.185345 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" containerName="cilium-agent" Jul 2 07:51:20.185546 kubelet[2083]: E0702 07:51:20.185526 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c091158d-3866-4a93-8622-ffc29817e3a2" containerName="cilium-operator" Jul 2 07:51:20.185707 kubelet[2083]: E0702 07:51:20.185690 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" containerName="clean-cilium-state" Jul 2 07:51:20.185914 kubelet[2083]: I0702 07:51:20.185884 2083 memory_manager.go:354] "RemoveStaleState removing state" podUID="c091158d-3866-4a93-8622-ffc29817e3a2" containerName="cilium-operator" Jul 2 07:51:20.186072 kubelet[2083]: I0702 07:51:20.186053 2083 memory_manager.go:354] "RemoveStaleState removing state" podUID="222f0b80-bb4d-48ae-a0a0-384baa208d60" containerName="cilium-agent" Jul 2 07:51:20.200887 systemd[1]: Created slice kubepods-burstable-podea97d56a_8fd9_4ddf_81b9_70e038a695ee.slice. Jul 2 07:51:20.204186 sshd[3784]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:20.208437 systemd[1]: sshd@22-10.128.0.12:22-147.75.109.163:41166.service: Deactivated successfully. Jul 2 07:51:20.209673 systemd[1]: session-22.scope: Deactivated successfully. Jul 2 07:51:20.210013 systemd[1]: session-22.scope: Consumed 1.009s CPU time. Jul 2 07:51:20.214400 systemd-logind[1211]: Session 22 logged out. Waiting for processes to exit. Jul 2 07:51:20.217520 systemd-logind[1211]: Removed session 22. Jul 2 07:51:20.253902 systemd[1]: Started sshd@23-10.128.0.12:22-147.75.109.163:41178.service. Jul 2 07:51:20.310555 kubelet[2083]: I0702 07:51:20.310513 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-lib-modules\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310555 kubelet[2083]: I0702 07:51:20.310575 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hostproc\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310937 kubelet[2083]: I0702 07:51:20.310609 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-config-path\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310937 kubelet[2083]: I0702 07:51:20.310639 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn6fg\" (UniqueName: \"kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-kube-api-access-sn6fg\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310937 kubelet[2083]: I0702 07:51:20.310670 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-run\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310937 kubelet[2083]: I0702 07:51:20.310703 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-cgroup\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310937 kubelet[2083]: I0702 07:51:20.310737 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-xtables-lock\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.310937 kubelet[2083]: I0702 07:51:20.310781 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-clustermesh-secrets\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311263 kubelet[2083]: I0702 07:51:20.310818 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-bpf-maps\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311263 kubelet[2083]: I0702 07:51:20.310847 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cni-path\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311263 kubelet[2083]: I0702 07:51:20.310875 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hubble-tls\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311263 kubelet[2083]: I0702 07:51:20.310910 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-etc-cni-netd\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311263 kubelet[2083]: I0702 07:51:20.310943 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-ipsec-secrets\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311263 kubelet[2083]: I0702 07:51:20.310981 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-net\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.311609 kubelet[2083]: I0702 07:51:20.311019 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-kernel\") pod \"cilium-t7hqf\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " pod="kube-system/cilium-t7hqf" Jul 2 07:51:20.519575 env[1224]: time="2024-07-02T07:51:20.519522974Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-t7hqf,Uid:ea97d56a-8fd9-4ddf-81b9-70e038a695ee,Namespace:kube-system,Attempt:0,}" Jul 2 07:51:20.544639 env[1224]: time="2024-07-02T07:51:20.543602517Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:51:20.544639 env[1224]: time="2024-07-02T07:51:20.543700138Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:51:20.544639 env[1224]: time="2024-07-02T07:51:20.543743518Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:51:20.544639 env[1224]: time="2024-07-02T07:51:20.543966152Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91 pid=3809 runtime=io.containerd.runc.v2 Jul 2 07:51:20.563893 sshd[3796]: Accepted publickey for core from 147.75.109.163 port 41178 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:20.564954 sshd[3796]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:20.571714 systemd-logind[1211]: New session 23 of user core. Jul 2 07:51:20.572746 systemd[1]: Started session-23.scope. Jul 2 07:51:20.579681 systemd[1]: Started cri-containerd-e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91.scope. Jul 2 07:51:20.624777 env[1224]: time="2024-07-02T07:51:20.624724797Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-t7hqf,Uid:ea97d56a-8fd9-4ddf-81b9-70e038a695ee,Namespace:kube-system,Attempt:0,} returns sandbox id \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\"" Jul 2 07:51:20.629920 env[1224]: time="2024-07-02T07:51:20.629876907Z" level=info msg="CreateContainer within sandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 2 07:51:20.644535 env[1224]: time="2024-07-02T07:51:20.644498197Z" level=info msg="CreateContainer within sandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\"" Jul 2 07:51:20.645355 env[1224]: time="2024-07-02T07:51:20.645320675Z" level=info msg="StartContainer for \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\"" Jul 2 07:51:20.666220 systemd[1]: Started cri-containerd-3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719.scope. Jul 2 07:51:20.684384 systemd[1]: cri-containerd-3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719.scope: Deactivated successfully. Jul 2 07:51:20.705646 env[1224]: time="2024-07-02T07:51:20.705559892Z" level=info msg="shim disconnected" id=3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719 Jul 2 07:51:20.705646 env[1224]: time="2024-07-02T07:51:20.705629264Z" level=warning msg="cleaning up after shim disconnected" id=3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719 namespace=k8s.io Jul 2 07:51:20.705646 env[1224]: time="2024-07-02T07:51:20.705645633Z" level=info msg="cleaning up dead shim" Jul 2 07:51:20.718245 env[1224]: time="2024-07-02T07:51:20.718176523Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:20Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3870 runtime=io.containerd.runc.v2\ntime=\"2024-07-02T07:51:20Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Jul 2 07:51:20.718629 env[1224]: time="2024-07-02T07:51:20.718506273Z" level=error msg="copy shim log" error="read /proc/self/fd/39: file already closed" Jul 2 07:51:20.718927 env[1224]: time="2024-07-02T07:51:20.718868004Z" level=error msg="Failed to pipe stdout of container \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\"" error="reading from a closed fifo" Jul 2 07:51:20.719431 env[1224]: time="2024-07-02T07:51:20.719090527Z" level=error msg="Failed to pipe stderr of container \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\"" error="reading from a closed fifo" Jul 2 07:51:20.721099 env[1224]: time="2024-07-02T07:51:20.721025118Z" level=error msg="StartContainer for \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Jul 2 07:51:20.721359 kubelet[2083]: E0702 07:51:20.721328 2083 remote_runtime.go:343] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719" Jul 2 07:51:20.721550 kubelet[2083]: E0702 07:51:20.721526 2083 kuberuntime_manager.go:1262] init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Jul 2 07:51:20.721550 kubelet[2083]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Jul 2 07:51:20.721550 kubelet[2083]: rm /hostbin/cilium-mount Jul 2 07:51:20.721746 kubelet[2083]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-sn6fg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-t7hqf_kube-system(ea97d56a-8fd9-4ddf-81b9-70e038a695ee): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Jul 2 07:51:20.721746 kubelet[2083]: E0702 07:51:20.721594 2083 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-t7hqf" podUID="ea97d56a-8fd9-4ddf-81b9-70e038a695ee" Jul 2 07:51:20.867106 sshd[3796]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:20.874128 systemd[1]: sshd@23-10.128.0.12:22-147.75.109.163:41178.service: Deactivated successfully. Jul 2 07:51:20.875390 systemd[1]: session-23.scope: Deactivated successfully. Jul 2 07:51:20.876661 systemd-logind[1211]: Session 23 logged out. Waiting for processes to exit. Jul 2 07:51:20.878843 systemd-logind[1211]: Removed session 23. Jul 2 07:51:20.913678 systemd[1]: Started sshd@24-10.128.0.12:22-147.75.109.163:41188.service. Jul 2 07:51:21.024573 kubelet[2083]: E0702 07:51:21.024534 2083 kubelet.go:2892] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 2 07:51:21.203361 sshd[3891]: Accepted publickey for core from 147.75.109.163 port 41188 ssh2: RSA SHA256:GSxC+U3gD/L2tgNRotlYHTLXvYsmaWMokGyA5lBCl2s Jul 2 07:51:21.206043 sshd[3891]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jul 2 07:51:21.213510 systemd[1]: Started session-24.scope. Jul 2 07:51:21.214572 systemd-logind[1211]: New session 24 of user core. Jul 2 07:51:21.372598 env[1224]: time="2024-07-02T07:51:21.372545681Z" level=info msg="StopPodSandbox for \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\"" Jul 2 07:51:21.372945 env[1224]: time="2024-07-02T07:51:21.372900536Z" level=info msg="Container to stop \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jul 2 07:51:21.396108 systemd[1]: cri-containerd-e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91.scope: Deactivated successfully. Jul 2 07:51:21.434134 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91-shm.mount: Deactivated successfully. Jul 2 07:51:21.461424 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91-rootfs.mount: Deactivated successfully. Jul 2 07:51:21.465945 env[1224]: time="2024-07-02T07:51:21.465884294Z" level=info msg="shim disconnected" id=e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91 Jul 2 07:51:21.466504 env[1224]: time="2024-07-02T07:51:21.466471247Z" level=warning msg="cleaning up after shim disconnected" id=e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91 namespace=k8s.io Jul 2 07:51:21.466678 env[1224]: time="2024-07-02T07:51:21.466653181Z" level=info msg="cleaning up dead shim" Jul 2 07:51:21.489265 env[1224]: time="2024-07-02T07:51:21.485821243Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:21Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3917 runtime=io.containerd.runc.v2\n" Jul 2 07:51:21.489265 env[1224]: time="2024-07-02T07:51:21.486245427Z" level=info msg="TearDown network for sandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" successfully" Jul 2 07:51:21.489265 env[1224]: time="2024-07-02T07:51:21.486276639Z" level=info msg="StopPodSandbox for \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" returns successfully" Jul 2 07:51:21.620775 kubelet[2083]: I0702 07:51:21.620715 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-cgroup\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.621494 kubelet[2083]: I0702 07:51:21.621471 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-xtables-lock\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.621692 kubelet[2083]: I0702 07:51:21.621675 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hubble-tls\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.621883 kubelet[2083]: I0702 07:51:21.621861 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-etc-cni-netd\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.622076 kubelet[2083]: I0702 07:51:21.622057 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-lib-modules\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.622253 kubelet[2083]: I0702 07:51:21.622239 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-clustermesh-secrets\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.622405 kubelet[2083]: I0702 07:51:21.622392 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-ipsec-secrets\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.622545 kubelet[2083]: I0702 07:51:21.622530 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-net\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.622775 kubelet[2083]: I0702 07:51:21.622730 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-kernel\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.622947 kubelet[2083]: I0702 07:51:21.622930 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-bpf-maps\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.623115 kubelet[2083]: I0702 07:51:21.623097 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cni-path\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.623331 kubelet[2083]: I0702 07:51:21.623300 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn6fg\" (UniqueName: \"kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-kube-api-access-sn6fg\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.623708 kubelet[2083]: I0702 07:51:21.623691 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-config-path\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.624137 kubelet[2083]: I0702 07:51:21.624120 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hostproc\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.624345 kubelet[2083]: I0702 07:51:21.624318 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.625679 kubelet[2083]: I0702 07:51:21.624337 2083 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-run\") pod \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\" (UID: \"ea97d56a-8fd9-4ddf-81b9-70e038a695ee\") " Jul 2 07:51:21.625803 kubelet[2083]: I0702 07:51:21.625742 2083 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-etc-cni-netd\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.625803 kubelet[2083]: I0702 07:51:21.624026 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.625803 kubelet[2083]: I0702 07:51:21.624379 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.625803 kubelet[2083]: I0702 07:51:21.623970 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.625803 kubelet[2083]: I0702 07:51:21.624405 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hostproc" (OuterVolumeSpecName: "hostproc") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.626097 kubelet[2083]: I0702 07:51:21.624485 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.626377 kubelet[2083]: I0702 07:51:21.626350 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.633678 kubelet[2083]: I0702 07:51:21.626515 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.634926 kubelet[2083]: I0702 07:51:21.626541 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.637102 kubelet[2083]: I0702 07:51:21.626560 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cni-path" (OuterVolumeSpecName: "cni-path") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jul 2 07:51:21.637102 kubelet[2083]: I0702 07:51:21.633627 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jul 2 07:51:21.641012 systemd[1]: var-lib-kubelet-pods-ea97d56a\x2d8fd9\x2d4ddf\x2d81b9\x2d70e038a695ee-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jul 2 07:51:21.646546 systemd[1]: var-lib-kubelet-pods-ea97d56a\x2d8fd9\x2d4ddf\x2d81b9\x2d70e038a695ee-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dsn6fg.mount: Deactivated successfully. Jul 2 07:51:21.652788 kubelet[2083]: I0702 07:51:21.650658 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 2 07:51:21.659067 systemd[1]: var-lib-kubelet-pods-ea97d56a\x2d8fd9\x2d4ddf\x2d81b9\x2d70e038a695ee-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jul 2 07:51:21.659210 systemd[1]: var-lib-kubelet-pods-ea97d56a\x2d8fd9\x2d4ddf\x2d81b9\x2d70e038a695ee-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Jul 2 07:51:21.661965 kubelet[2083]: I0702 07:51:21.661927 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 2 07:51:21.662604 kubelet[2083]: I0702 07:51:21.662572 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-kube-api-access-sn6fg" (OuterVolumeSpecName: "kube-api-access-sn6fg") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "kube-api-access-sn6fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jul 2 07:51:21.664996 kubelet[2083]: I0702 07:51:21.664960 2083 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "ea97d56a-8fd9-4ddf-81b9-70e038a695ee" (UID: "ea97d56a-8fd9-4ddf-81b9-70e038a695ee"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jul 2 07:51:21.726499 kubelet[2083]: I0702 07:51:21.726377 2083 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-kernel\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.726740 kubelet[2083]: I0702 07:51:21.726714 2083 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-bpf-maps\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.726912 kubelet[2083]: I0702 07:51:21.726893 2083 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cni-path\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727037 kubelet[2083]: I0702 07:51:21.727021 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-ipsec-secrets\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727261 kubelet[2083]: I0702 07:51:21.727244 2083 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-host-proc-sys-net\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727392 kubelet[2083]: I0702 07:51:21.727376 2083 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-sn6fg\" (UniqueName: \"kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-kube-api-access-sn6fg\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727526 kubelet[2083]: I0702 07:51:21.727496 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-config-path\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727648 kubelet[2083]: I0702 07:51:21.727633 2083 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hostproc\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727783 kubelet[2083]: I0702 07:51:21.727754 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-run\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.727909 kubelet[2083]: I0702 07:51:21.727893 2083 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-xtables-lock\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.728026 kubelet[2083]: I0702 07:51:21.728010 2083 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-hubble-tls\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.728138 kubelet[2083]: I0702 07:51:21.728121 2083 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-cilium-cgroup\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.728262 kubelet[2083]: I0702 07:51:21.728246 2083 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-clustermesh-secrets\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.728380 kubelet[2083]: I0702 07:51:21.728365 2083 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea97d56a-8fd9-4ddf-81b9-70e038a695ee-lib-modules\") on node \"ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal\" DevicePath \"\"" Jul 2 07:51:21.935801 systemd[1]: Removed slice kubepods-burstable-podea97d56a_8fd9_4ddf_81b9_70e038a695ee.slice. Jul 2 07:51:22.375591 kubelet[2083]: I0702 07:51:22.375558 2083 scope.go:117] "RemoveContainer" containerID="3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719" Jul 2 07:51:22.381285 env[1224]: time="2024-07-02T07:51:22.381238754Z" level=info msg="RemoveContainer for \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\"" Jul 2 07:51:22.387171 env[1224]: time="2024-07-02T07:51:22.387113689Z" level=info msg="RemoveContainer for \"3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719\" returns successfully" Jul 2 07:51:22.418213 kubelet[2083]: I0702 07:51:22.418166 2083 topology_manager.go:215] "Topology Admit Handler" podUID="2e8a2a43-9177-4d3a-9cca-bdf9d395bce9" podNamespace="kube-system" podName="cilium-qs849" Jul 2 07:51:22.418566 kubelet[2083]: E0702 07:51:22.418541 2083 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="ea97d56a-8fd9-4ddf-81b9-70e038a695ee" containerName="mount-cgroup" Jul 2 07:51:22.418797 kubelet[2083]: I0702 07:51:22.418751 2083 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea97d56a-8fd9-4ddf-81b9-70e038a695ee" containerName="mount-cgroup" Jul 2 07:51:22.430642 systemd[1]: Created slice kubepods-burstable-pod2e8a2a43_9177_4d3a_9cca_bdf9d395bce9.slice. Jul 2 07:51:22.535957 kubelet[2083]: I0702 07:51:22.535915 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-cilium-ipsec-secrets\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.536500 kubelet[2083]: I0702 07:51:22.536418 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-host-proc-sys-kernel\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.536848 kubelet[2083]: I0702 07:51:22.536825 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-cilium-run\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.537121 kubelet[2083]: I0702 07:51:22.537102 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-etc-cni-netd\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.537728 kubelet[2083]: I0702 07:51:22.537655 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-clustermesh-secrets\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.537986 kubelet[2083]: I0702 07:51:22.537942 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-cilium-config-path\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.538196 kubelet[2083]: I0702 07:51:22.538166 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-bpf-maps\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.538512 kubelet[2083]: I0702 07:51:22.538489 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-hostproc\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.538738 kubelet[2083]: I0702 07:51:22.538700 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-cilium-cgroup\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.538982 kubelet[2083]: I0702 07:51:22.538958 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-cni-path\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.539209 kubelet[2083]: I0702 07:51:22.539171 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-lib-modules\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.539421 kubelet[2083]: I0702 07:51:22.539387 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-host-proc-sys-net\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.539618 kubelet[2083]: I0702 07:51:22.539589 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-hubble-tls\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.539833 kubelet[2083]: I0702 07:51:22.539813 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-xtables-lock\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.540055 kubelet[2083]: I0702 07:51:22.540019 2083 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj6qz\" (UniqueName: \"kubernetes.io/projected/2e8a2a43-9177-4d3a-9cca-bdf9d395bce9-kube-api-access-kj6qz\") pod \"cilium-qs849\" (UID: \"2e8a2a43-9177-4d3a-9cca-bdf9d395bce9\") " pod="kube-system/cilium-qs849" Jul 2 07:51:22.740840 env[1224]: time="2024-07-02T07:51:22.740668443Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qs849,Uid:2e8a2a43-9177-4d3a-9cca-bdf9d395bce9,Namespace:kube-system,Attempt:0,}" Jul 2 07:51:22.760249 env[1224]: time="2024-07-02T07:51:22.760161087Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jul 2 07:51:22.760249 env[1224]: time="2024-07-02T07:51:22.760212643Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jul 2 07:51:22.760583 env[1224]: time="2024-07-02T07:51:22.760230407Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jul 2 07:51:22.760959 env[1224]: time="2024-07-02T07:51:22.760889942Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430 pid=3945 runtime=io.containerd.runc.v2 Jul 2 07:51:22.780152 systemd[1]: Started cri-containerd-ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430.scope. Jul 2 07:51:22.812155 env[1224]: time="2024-07-02T07:51:22.812102551Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-qs849,Uid:2e8a2a43-9177-4d3a-9cca-bdf9d395bce9,Namespace:kube-system,Attempt:0,} returns sandbox id \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\"" Jul 2 07:51:22.818140 env[1224]: time="2024-07-02T07:51:22.818067568Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jul 2 07:51:22.834229 env[1224]: time="2024-07-02T07:51:22.834189898Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4\"" Jul 2 07:51:22.835238 env[1224]: time="2024-07-02T07:51:22.835200424Z" level=info msg="StartContainer for \"adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4\"" Jul 2 07:51:22.859104 systemd[1]: Started cri-containerd-adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4.scope. Jul 2 07:51:22.905484 env[1224]: time="2024-07-02T07:51:22.905427950Z" level=info msg="StartContainer for \"adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4\" returns successfully" Jul 2 07:51:22.918753 systemd[1]: cri-containerd-adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4.scope: Deactivated successfully. Jul 2 07:51:22.954899 env[1224]: time="2024-07-02T07:51:22.954842343Z" level=info msg="shim disconnected" id=adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4 Jul 2 07:51:22.954899 env[1224]: time="2024-07-02T07:51:22.954901683Z" level=warning msg="cleaning up after shim disconnected" id=adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4 namespace=k8s.io Jul 2 07:51:22.955311 env[1224]: time="2024-07-02T07:51:22.954916682Z" level=info msg="cleaning up dead shim" Jul 2 07:51:22.966231 env[1224]: time="2024-07-02T07:51:22.966194696Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:22Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4032 runtime=io.containerd.runc.v2\n" Jul 2 07:51:23.385979 env[1224]: time="2024-07-02T07:51:23.385872323Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jul 2 07:51:23.409572 env[1224]: time="2024-07-02T07:51:23.409488245Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f\"" Jul 2 07:51:23.410265 env[1224]: time="2024-07-02T07:51:23.410217543Z" level=info msg="StartContainer for \"10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f\"" Jul 2 07:51:23.453896 systemd[1]: Started cri-containerd-10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f.scope. Jul 2 07:51:23.463481 systemd[1]: run-containerd-runc-k8s.io-10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f-runc.kGVc5j.mount: Deactivated successfully. Jul 2 07:51:23.497812 env[1224]: time="2024-07-02T07:51:23.497389448Z" level=info msg="StartContainer for \"10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f\" returns successfully" Jul 2 07:51:23.506643 systemd[1]: cri-containerd-10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f.scope: Deactivated successfully. Jul 2 07:51:23.539586 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f-rootfs.mount: Deactivated successfully. Jul 2 07:51:23.542232 env[1224]: time="2024-07-02T07:51:23.542165782Z" level=info msg="shim disconnected" id=10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f Jul 2 07:51:23.542417 env[1224]: time="2024-07-02T07:51:23.542333048Z" level=warning msg="cleaning up after shim disconnected" id=10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f namespace=k8s.io Jul 2 07:51:23.542417 env[1224]: time="2024-07-02T07:51:23.542360934Z" level=info msg="cleaning up dead shim" Jul 2 07:51:23.553255 env[1224]: time="2024-07-02T07:51:23.553206624Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:23Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4096 runtime=io.containerd.runc.v2\ntime=\"2024-07-02T07:51:23Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" Jul 2 07:51:23.815852 kubelet[2083]: W0702 07:51:23.815794 2083 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea97d56a_8fd9_4ddf_81b9_70e038a695ee.slice/cri-containerd-3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719.scope WatchSource:0}: container "3f93169f20ab0f7609cf8ad1bd3caf8b8e9da7ae5236d73ada7502fcd56ff719" in namespace "k8s.io": not found Jul 2 07:51:23.931443 kubelet[2083]: I0702 07:51:23.931402 2083 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="ea97d56a-8fd9-4ddf-81b9-70e038a695ee" path="/var/lib/kubelet/pods/ea97d56a-8fd9-4ddf-81b9-70e038a695ee/volumes" Jul 2 07:51:24.389911 env[1224]: time="2024-07-02T07:51:24.389839395Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jul 2 07:51:24.420516 env[1224]: time="2024-07-02T07:51:24.420446218Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a\"" Jul 2 07:51:24.421190 env[1224]: time="2024-07-02T07:51:24.421146590Z" level=info msg="StartContainer for \"e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a\"" Jul 2 07:51:24.433816 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount695481932.mount: Deactivated successfully. Jul 2 07:51:24.465938 systemd[1]: Started cri-containerd-e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a.scope. Jul 2 07:51:24.475263 systemd[1]: run-containerd-runc-k8s.io-e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a-runc.sn77Mg.mount: Deactivated successfully. Jul 2 07:51:24.516808 env[1224]: time="2024-07-02T07:51:24.515727855Z" level=info msg="StartContainer for \"e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a\" returns successfully" Jul 2 07:51:24.522930 systemd[1]: cri-containerd-e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a.scope: Deactivated successfully. Jul 2 07:51:24.555742 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a-rootfs.mount: Deactivated successfully. Jul 2 07:51:24.562304 env[1224]: time="2024-07-02T07:51:24.562242018Z" level=info msg="shim disconnected" id=e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a Jul 2 07:51:24.562613 env[1224]: time="2024-07-02T07:51:24.562307869Z" level=warning msg="cleaning up after shim disconnected" id=e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a namespace=k8s.io Jul 2 07:51:24.562613 env[1224]: time="2024-07-02T07:51:24.562324692Z" level=info msg="cleaning up dead shim" Jul 2 07:51:24.573235 env[1224]: time="2024-07-02T07:51:24.573177460Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:24Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4158 runtime=io.containerd.runc.v2\n" Jul 2 07:51:25.395804 env[1224]: time="2024-07-02T07:51:25.395725234Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jul 2 07:51:25.417823 env[1224]: time="2024-07-02T07:51:25.415938257Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544\"" Jul 2 07:51:25.418393 env[1224]: time="2024-07-02T07:51:25.418339409Z" level=info msg="StartContainer for \"2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544\"" Jul 2 07:51:25.459684 systemd[1]: run-containerd-runc-k8s.io-2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544-runc.eqHdUs.mount: Deactivated successfully. Jul 2 07:51:25.466713 systemd[1]: Started cri-containerd-2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544.scope. Jul 2 07:51:25.508483 systemd[1]: cri-containerd-2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544.scope: Deactivated successfully. Jul 2 07:51:25.510306 env[1224]: time="2024-07-02T07:51:25.510198879Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e8a2a43_9177_4d3a_9cca_bdf9d395bce9.slice/cri-containerd-2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544.scope/memory.events\": no such file or directory" Jul 2 07:51:25.512846 env[1224]: time="2024-07-02T07:51:25.512742556Z" level=info msg="StartContainer for \"2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544\" returns successfully" Jul 2 07:51:25.547240 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544-rootfs.mount: Deactivated successfully. Jul 2 07:51:25.554271 env[1224]: time="2024-07-02T07:51:25.554210950Z" level=info msg="shim disconnected" id=2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544 Jul 2 07:51:25.554604 env[1224]: time="2024-07-02T07:51:25.554275236Z" level=warning msg="cleaning up after shim disconnected" id=2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544 namespace=k8s.io Jul 2 07:51:25.554604 env[1224]: time="2024-07-02T07:51:25.554294467Z" level=info msg="cleaning up dead shim" Jul 2 07:51:25.566656 env[1224]: time="2024-07-02T07:51:25.566605016Z" level=warning msg="cleanup warnings time=\"2024-07-02T07:51:25Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4212 runtime=io.containerd.runc.v2\n" Jul 2 07:51:25.827617 env[1224]: time="2024-07-02T07:51:25.827538569Z" level=info msg="StopPodSandbox for \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\"" Jul 2 07:51:25.827890 env[1224]: time="2024-07-02T07:51:25.827679159Z" level=info msg="TearDown network for sandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" successfully" Jul 2 07:51:25.827890 env[1224]: time="2024-07-02T07:51:25.827731370Z" level=info msg="StopPodSandbox for \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" returns successfully" Jul 2 07:51:25.828524 env[1224]: time="2024-07-02T07:51:25.828478467Z" level=info msg="RemovePodSandbox for \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\"" Jul 2 07:51:25.828694 env[1224]: time="2024-07-02T07:51:25.828528299Z" level=info msg="Forcibly stopping sandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\"" Jul 2 07:51:25.828694 env[1224]: time="2024-07-02T07:51:25.828659141Z" level=info msg="TearDown network for sandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" successfully" Jul 2 07:51:25.833224 env[1224]: time="2024-07-02T07:51:25.833178609Z" level=info msg="RemovePodSandbox \"e9b0e1c4b7d74a1861322407d826dcbc5b8f41bdaddd459eb65fab951d84e3c8\" returns successfully" Jul 2 07:51:25.833816 env[1224]: time="2024-07-02T07:51:25.833722710Z" level=info msg="StopPodSandbox for \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\"" Jul 2 07:51:25.833949 env[1224]: time="2024-07-02T07:51:25.833873186Z" level=info msg="TearDown network for sandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" successfully" Jul 2 07:51:25.833949 env[1224]: time="2024-07-02T07:51:25.833926364Z" level=info msg="StopPodSandbox for \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" returns successfully" Jul 2 07:51:25.834470 env[1224]: time="2024-07-02T07:51:25.834422026Z" level=info msg="RemovePodSandbox for \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\"" Jul 2 07:51:25.834606 env[1224]: time="2024-07-02T07:51:25.834459299Z" level=info msg="Forcibly stopping sandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\"" Jul 2 07:51:25.834606 env[1224]: time="2024-07-02T07:51:25.834565203Z" level=info msg="TearDown network for sandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" successfully" Jul 2 07:51:25.838296 env[1224]: time="2024-07-02T07:51:25.838243850Z" level=info msg="RemovePodSandbox \"e9c91b53426d03524b310686aa64c6bd365e59e5671bb0300c17c8b74ff55a91\" returns successfully" Jul 2 07:51:25.838879 env[1224]: time="2024-07-02T07:51:25.838837253Z" level=info msg="StopPodSandbox for \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\"" Jul 2 07:51:25.839168 env[1224]: time="2024-07-02T07:51:25.839077254Z" level=info msg="TearDown network for sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" successfully" Jul 2 07:51:25.839168 env[1224]: time="2024-07-02T07:51:25.839146993Z" level=info msg="StopPodSandbox for \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" returns successfully" Jul 2 07:51:25.839743 env[1224]: time="2024-07-02T07:51:25.839708294Z" level=info msg="RemovePodSandbox for \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\"" Jul 2 07:51:25.839887 env[1224]: time="2024-07-02T07:51:25.839748066Z" level=info msg="Forcibly stopping sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\"" Jul 2 07:51:25.839957 env[1224]: time="2024-07-02T07:51:25.839881133Z" level=info msg="TearDown network for sandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" successfully" Jul 2 07:51:25.846083 env[1224]: time="2024-07-02T07:51:25.846039377Z" level=info msg="RemovePodSandbox \"829d3d88c6d4f66606138153e4fd86b21c5f86fad1f4aae24ed34a57d5515753\" returns successfully" Jul 2 07:51:26.025845 kubelet[2083]: E0702 07:51:26.025810 2083 kubelet.go:2892] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jul 2 07:51:26.400505 env[1224]: time="2024-07-02T07:51:26.400447076Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jul 2 07:51:26.422735 env[1224]: time="2024-07-02T07:51:26.422683842Z" level=info msg="CreateContainer within sandbox \"ee8d5a57f73bce606f0bdb2305f0faa7f5072a8c791e8ec5a555cfd04bbbe430\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb\"" Jul 2 07:51:26.430659 env[1224]: time="2024-07-02T07:51:26.427745316Z" level=info msg="StartContainer for \"25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb\"" Jul 2 07:51:26.464260 systemd[1]: run-containerd-runc-k8s.io-25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb-runc.5mbCZ5.mount: Deactivated successfully. Jul 2 07:51:26.470426 systemd[1]: Started cri-containerd-25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb.scope. Jul 2 07:51:26.519878 env[1224]: time="2024-07-02T07:51:26.519816565Z" level=info msg="StartContainer for \"25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb\" returns successfully" Jul 2 07:51:26.928174 kubelet[2083]: W0702 07:51:26.928112 2083 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e8a2a43_9177_4d3a_9cca_bdf9d395bce9.slice/cri-containerd-adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4.scope WatchSource:0}: task adda37c5aed92f664d3aed4e89002e76b218a5fe0b710f8a8a050646394343c4 not found: not found Jul 2 07:51:26.996827 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Jul 2 07:51:27.417926 kubelet[2083]: I0702 07:51:27.417888 2083 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-qs849" podStartSLOduration=5.417832887 podStartE2EDuration="5.417832887s" podCreationTimestamp="2024-07-02 07:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 07:51:27.417604796 +0000 UTC m=+121.757097121" watchObservedRunningTime="2024-07-02 07:51:27.417832887 +0000 UTC m=+121.757325219" Jul 2 07:51:27.433651 systemd[1]: run-containerd-runc-k8s.io-25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb-runc.0fc01f.mount: Deactivated successfully. Jul 2 07:51:27.732308 systemd[1]: run-containerd-runc-k8s.io-25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb-runc.4cwmJT.mount: Deactivated successfully. Jul 2 07:51:28.208152 kubelet[2083]: I0702 07:51:28.208101 2083 setters.go:568] "Node became not ready" node="ci-3510-3-5-8df16e6abe40ae309741.c.flatcar-212911.internal" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-07-02T07:51:28Z","lastTransitionTime":"2024-07-02T07:51:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jul 2 07:51:28.927587 kubelet[2083]: E0702 07:51:28.927534 2083 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-76f75df574-95qsg" podUID="38d069a9-64fa-4ca0-b2ea-061ae6132013" Jul 2 07:51:29.820918 update_engine[1212]: I0702 07:51:29.820863 1212 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jul 2 07:51:29.821467 update_engine[1212]: I0702 07:51:29.821180 1212 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jul 2 07:51:29.821467 update_engine[1212]: I0702 07:51:29.821408 1212 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jul 2 07:51:29.829665 update_engine[1212]: E0702 07:51:29.829624 1212 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jul 2 07:51:29.829833 update_engine[1212]: I0702 07:51:29.829785 1212 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Jul 2 07:51:29.930095 kubelet[2083]: E0702 07:51:29.930031 2083 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-76f75df574-c282x" podUID="5c910845-809b-49a0-9d61-da1049247bc7" Jul 2 07:51:29.982785 systemd-networkd[1022]: lxc_health: Link UP Jul 2 07:51:29.991866 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Jul 2 07:51:29.994247 systemd-networkd[1022]: lxc_health: Gained carrier Jul 2 07:51:30.045398 kubelet[2083]: W0702 07:51:30.045344 2083 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e8a2a43_9177_4d3a_9cca_bdf9d395bce9.slice/cri-containerd-10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f.scope WatchSource:0}: task 10a5ba9dbab05b11a68a5dddcd61e5ce2c0b80956b9dbb6e94376198c878089f not found: not found Jul 2 07:51:30.927176 kubelet[2083]: E0702 07:51:30.927135 2083 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-76f75df574-95qsg" podUID="38d069a9-64fa-4ca0-b2ea-061ae6132013" Jul 2 07:51:31.209678 systemd-networkd[1022]: lxc_health: Gained IPv6LL Jul 2 07:51:32.274153 systemd[1]: run-containerd-runc-k8s.io-25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb-runc.41m991.mount: Deactivated successfully. Jul 2 07:51:33.159023 kubelet[2083]: W0702 07:51:33.158962 2083 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e8a2a43_9177_4d3a_9cca_bdf9d395bce9.slice/cri-containerd-e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a.scope WatchSource:0}: task e5b2271795342e9a31a66c5d6a4d9a17102cb519f49cd5dced4b2b7193c9f14a not found: not found Jul 2 07:51:34.505710 systemd[1]: run-containerd-runc-k8s.io-25900f272ff2a636309d3d15d4e4cf1fae552ab86af12d545b4a081a2121a4eb-runc.Yu7xWu.mount: Deactivated successfully. Jul 2 07:51:36.271809 kubelet[2083]: W0702 07:51:36.271731 2083 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e8a2a43_9177_4d3a_9cca_bdf9d395bce9.slice/cri-containerd-2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544.scope WatchSource:0}: task 2b78125fd6b79e148515ee2d14fa63a664bcc6105a69c00e4af052a269c7d544 not found: not found Jul 2 07:51:36.856842 sshd[3891]: pam_unix(sshd:session): session closed for user core Jul 2 07:51:36.861154 systemd[1]: sshd@24-10.128.0.12:22-147.75.109.163:41188.service: Deactivated successfully. Jul 2 07:51:36.862325 systemd[1]: session-24.scope: Deactivated successfully. Jul 2 07:51:36.863978 systemd-logind[1211]: Session 24 logged out. Waiting for processes to exit. Jul 2 07:51:36.865335 systemd-logind[1211]: Removed session 24. Jul 2 07:51:39.820201 update_engine[1212]: I0702 07:51:39.820132 1212 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jul 2 07:51:39.820802 update_engine[1212]: I0702 07:51:39.820474 1212 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jul 2 07:51:39.820802 update_engine[1212]: I0702 07:51:39.820745 1212 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jul 2 07:51:39.882611 update_engine[1212]: E0702 07:51:39.882552 1212 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jul 2 07:51:39.882841 update_engine[1212]: I0702 07:51:39.882694 1212 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Jul 2 07:51:39.882841 update_engine[1212]: I0702 07:51:39.882709 1212 omaha_request_action.cc:621] Omaha request response: Jul 2 07:51:39.882841 update_engine[1212]: E0702 07:51:39.882828 1212 omaha_request_action.cc:640] Omaha request network transfer failed. Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.882853 1212 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.882860 1212 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.882866 1212 update_attempter.cc:306] Processing Done. Jul 2 07:51:39.883028 update_engine[1212]: E0702 07:51:39.882885 1212 update_attempter.cc:619] Update failed. Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.882894 1212 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.882899 1212 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.882908 1212 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Jul 2 07:51:39.883028 update_engine[1212]: I0702 07:51:39.883005 1212 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Jul 2 07:51:39.883438 update_engine[1212]: I0702 07:51:39.883035 1212 omaha_request_action.cc:270] Posting an Omaha request to disabled Jul 2 07:51:39.883438 update_engine[1212]: I0702 07:51:39.883044 1212 omaha_request_action.cc:271] Request: Jul 2 07:51:39.883438 update_engine[1212]: Jul 2 07:51:39.883438 update_engine[1212]: Jul 2 07:51:39.883438 update_engine[1212]: Jul 2 07:51:39.883438 update_engine[1212]: Jul 2 07:51:39.883438 update_engine[1212]: Jul 2 07:51:39.883438 update_engine[1212]: Jul 2 07:51:39.883438 update_engine[1212]: I0702 07:51:39.883057 1212 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jul 2 07:51:39.883438 update_engine[1212]: I0702 07:51:39.883322 1212 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jul 2 07:51:39.883936 update_engine[1212]: I0702 07:51:39.883541 1212 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jul 2 07:51:39.884073 locksmithd[1256]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Jul 2 07:51:39.896164 update_engine[1212]: E0702 07:51:39.896122 1212 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jul 2 07:51:39.896291 update_engine[1212]: I0702 07:51:39.896258 1212 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Jul 2 07:51:39.896291 update_engine[1212]: I0702 07:51:39.896271 1212 omaha_request_action.cc:621] Omaha request response: Jul 2 07:51:39.896291 update_engine[1212]: I0702 07:51:39.896281 1212 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Jul 2 07:51:39.896291 update_engine[1212]: I0702 07:51:39.896286 1212 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Jul 2 07:51:39.896499 update_engine[1212]: I0702 07:51:39.896293 1212 update_attempter.cc:306] Processing Done. Jul 2 07:51:39.896499 update_engine[1212]: I0702 07:51:39.896302 1212 update_attempter.cc:310] Error event sent. Jul 2 07:51:39.896499 update_engine[1212]: I0702 07:51:39.896312 1212 update_check_scheduler.cc:74] Next update check in 48m11s Jul 2 07:51:39.896805 locksmithd[1256]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0