Feb 9 08:58:34.041689 kernel: Linux version 5.15.148-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Thu Feb 8 21:14:17 -00 2024 Feb 9 08:58:34.050758 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ae7db544026ede4699ee2036449b75950d3fb7929b25a6731d0ad396f1aa37c9 Feb 9 08:58:34.050808 kernel: BIOS-provided physical RAM map: Feb 9 08:58:34.050821 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Feb 9 08:58:34.050832 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Feb 9 08:58:34.050844 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Feb 9 08:58:34.050859 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000007ffdafff] usable Feb 9 08:58:34.050872 kernel: BIOS-e820: [mem 0x000000007ffdb000-0x000000007fffffff] reserved Feb 9 08:58:34.050887 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Feb 9 08:58:34.050900 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Feb 9 08:58:34.050912 kernel: NX (Execute Disable) protection: active Feb 9 08:58:34.050924 kernel: SMBIOS 2.8 present. Feb 9 08:58:34.050937 kernel: DMI: DigitalOcean Droplet/Droplet, BIOS 20171212 12/12/2017 Feb 9 08:58:34.050949 kernel: Hypervisor detected: KVM Feb 9 08:58:34.050965 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Feb 9 08:58:34.050982 kernel: kvm-clock: cpu 0, msr 40faa001, primary cpu clock Feb 9 08:58:34.050995 kernel: kvm-clock: using sched offset of 3526661219 cycles Feb 9 08:58:34.051010 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Feb 9 08:58:34.051024 kernel: tsc: Detected 2494.140 MHz processor Feb 9 08:58:34.051038 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Feb 9 08:58:34.051052 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Feb 9 08:58:34.051066 kernel: last_pfn = 0x7ffdb max_arch_pfn = 0x400000000 Feb 9 08:58:34.051088 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Feb 9 08:58:34.051105 kernel: ACPI: Early table checksum verification disabled Feb 9 08:58:34.051120 kernel: ACPI: RSDP 0x00000000000F5950 000014 (v00 BOCHS ) Feb 9 08:58:34.051133 kernel: ACPI: RSDT 0x000000007FFE1986 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051147 kernel: ACPI: FACP 0x000000007FFE176A 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051160 kernel: ACPI: DSDT 0x000000007FFE0040 00172A (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051172 kernel: ACPI: FACS 0x000000007FFE0000 000040 Feb 9 08:58:34.051186 kernel: ACPI: APIC 0x000000007FFE17DE 000080 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051200 kernel: ACPI: HPET 0x000000007FFE185E 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051231 kernel: ACPI: SRAT 0x000000007FFE1896 0000C8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051247 kernel: ACPI: WAET 0x000000007FFE195E 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Feb 9 08:58:34.051261 kernel: ACPI: Reserving FACP table memory at [mem 0x7ffe176a-0x7ffe17dd] Feb 9 08:58:34.051275 kernel: ACPI: Reserving DSDT table memory at [mem 0x7ffe0040-0x7ffe1769] Feb 9 08:58:34.051288 kernel: ACPI: Reserving FACS table memory at [mem 0x7ffe0000-0x7ffe003f] Feb 9 08:58:34.051302 kernel: ACPI: Reserving APIC table memory at [mem 0x7ffe17de-0x7ffe185d] Feb 9 08:58:34.051316 kernel: ACPI: Reserving HPET table memory at [mem 0x7ffe185e-0x7ffe1895] Feb 9 08:58:34.051329 kernel: ACPI: Reserving SRAT table memory at [mem 0x7ffe1896-0x7ffe195d] Feb 9 08:58:34.051343 kernel: ACPI: Reserving WAET table memory at [mem 0x7ffe195e-0x7ffe1985] Feb 9 08:58:34.051366 kernel: SRAT: PXM 0 -> APIC 0x00 -> Node 0 Feb 9 08:58:34.051380 kernel: SRAT: PXM 0 -> APIC 0x01 -> Node 0 Feb 9 08:58:34.051395 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x0009ffff] Feb 9 08:58:34.051410 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00100000-0x7fffffff] Feb 9 08:58:34.051424 kernel: NUMA: Node 0 [mem 0x00000000-0x0009ffff] + [mem 0x00100000-0x7ffdafff] -> [mem 0x00000000-0x7ffdafff] Feb 9 08:58:34.051438 kernel: NODE_DATA(0) allocated [mem 0x7ffd5000-0x7ffdafff] Feb 9 08:58:34.051456 kernel: Zone ranges: Feb 9 08:58:34.051471 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Feb 9 08:58:34.051485 kernel: DMA32 [mem 0x0000000001000000-0x000000007ffdafff] Feb 9 08:58:34.051500 kernel: Normal empty Feb 9 08:58:34.051514 kernel: Movable zone start for each node Feb 9 08:58:34.051528 kernel: Early memory node ranges Feb 9 08:58:34.051543 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Feb 9 08:58:34.051557 kernel: node 0: [mem 0x0000000000100000-0x000000007ffdafff] Feb 9 08:58:34.051572 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000007ffdafff] Feb 9 08:58:34.051589 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Feb 9 08:58:34.051608 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Feb 9 08:58:34.051623 kernel: On node 0, zone DMA32: 37 pages in unavailable ranges Feb 9 08:58:34.051638 kernel: ACPI: PM-Timer IO Port: 0x608 Feb 9 08:58:34.051653 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Feb 9 08:58:34.051667 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Feb 9 08:58:34.051682 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Feb 9 08:58:34.051697 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Feb 9 08:58:34.051712 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Feb 9 08:58:34.055567 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Feb 9 08:58:34.055590 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Feb 9 08:58:34.055605 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Feb 9 08:58:34.055620 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Feb 9 08:58:34.055635 kernel: TSC deadline timer available Feb 9 08:58:34.055655 kernel: smpboot: Allowing 2 CPUs, 0 hotplug CPUs Feb 9 08:58:34.055671 kernel: [mem 0x80000000-0xfeffbfff] available for PCI devices Feb 9 08:58:34.055685 kernel: Booting paravirtualized kernel on KVM Feb 9 08:58:34.055701 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Feb 9 08:58:34.055810 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:2 nr_node_ids:1 Feb 9 08:58:34.055826 kernel: percpu: Embedded 55 pages/cpu s185624 r8192 d31464 u1048576 Feb 9 08:58:34.055841 kernel: pcpu-alloc: s185624 r8192 d31464 u1048576 alloc=1*2097152 Feb 9 08:58:34.055856 kernel: pcpu-alloc: [0] 0 1 Feb 9 08:58:34.055871 kernel: kvm-guest: stealtime: cpu 0, msr 7dc1c0c0 Feb 9 08:58:34.055886 kernel: kvm-guest: PV spinlocks disabled, no host support Feb 9 08:58:34.055901 kernel: Built 1 zonelists, mobility grouping on. Total pages: 515803 Feb 9 08:58:34.055916 kernel: Policy zone: DMA32 Feb 9 08:58:34.055932 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ae7db544026ede4699ee2036449b75950d3fb7929b25a6731d0ad396f1aa37c9 Feb 9 08:58:34.055952 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Feb 9 08:58:34.055967 kernel: Dentry cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Feb 9 08:58:34.055982 kernel: Inode-cache hash table entries: 131072 (order: 8, 1048576 bytes, linear) Feb 9 08:58:34.055997 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Feb 9 08:58:34.056009 kernel: Memory: 1975332K/2096612K available (12294K kernel code, 2275K rwdata, 13700K rodata, 45496K init, 4048K bss, 121020K reserved, 0K cma-reserved) Feb 9 08:58:34.056022 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Feb 9 08:58:34.056035 kernel: Kernel/User page tables isolation: enabled Feb 9 08:58:34.056048 kernel: ftrace: allocating 34475 entries in 135 pages Feb 9 08:58:34.056065 kernel: ftrace: allocated 135 pages with 4 groups Feb 9 08:58:34.056077 kernel: rcu: Hierarchical RCU implementation. Feb 9 08:58:34.056090 kernel: rcu: RCU event tracing is enabled. Feb 9 08:58:34.056104 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Feb 9 08:58:34.056116 kernel: Rude variant of Tasks RCU enabled. Feb 9 08:58:34.056129 kernel: Tracing variant of Tasks RCU enabled. Feb 9 08:58:34.056144 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Feb 9 08:58:34.056160 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Feb 9 08:58:34.056174 kernel: NR_IRQS: 33024, nr_irqs: 440, preallocated irqs: 16 Feb 9 08:58:34.056194 kernel: random: crng init done Feb 9 08:58:34.056208 kernel: Console: colour VGA+ 80x25 Feb 9 08:58:34.056223 kernel: printk: console [tty0] enabled Feb 9 08:58:34.056238 kernel: printk: console [ttyS0] enabled Feb 9 08:58:34.056253 kernel: ACPI: Core revision 20210730 Feb 9 08:58:34.056267 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Feb 9 08:58:34.056282 kernel: APIC: Switch to symmetric I/O mode setup Feb 9 08:58:34.056297 kernel: x2apic enabled Feb 9 08:58:34.056312 kernel: Switched APIC routing to physical x2apic. Feb 9 08:58:34.056327 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Feb 9 08:58:34.056346 kernel: clocksource: tsc-early: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Feb 9 08:58:34.056361 kernel: Calibrating delay loop (skipped) preset value.. 4988.28 BogoMIPS (lpj=2494140) Feb 9 08:58:34.056381 kernel: Last level iTLB entries: 4KB 0, 2MB 0, 4MB 0 Feb 9 08:58:34.056395 kernel: Last level dTLB entries: 4KB 0, 2MB 0, 4MB 0, 1GB 0 Feb 9 08:58:34.056411 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Feb 9 08:58:34.056425 kernel: Spectre V2 : Mitigation: Retpolines Feb 9 08:58:34.056441 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Feb 9 08:58:34.056458 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Feb 9 08:58:34.056477 kernel: Spectre V2 : Enabling Restricted Speculation for firmware calls Feb 9 08:58:34.056505 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Feb 9 08:58:34.056521 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Feb 9 08:58:34.056540 kernel: MDS: Mitigation: Clear CPU buffers Feb 9 08:58:34.056555 kernel: MMIO Stale Data: Vulnerable: Clear CPU buffers attempted, no microcode Feb 9 08:58:34.056572 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Feb 9 08:58:34.056587 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Feb 9 08:58:34.056603 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Feb 9 08:58:34.056619 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Feb 9 08:58:34.056635 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Feb 9 08:58:34.056653 kernel: Freeing SMP alternatives memory: 32K Feb 9 08:58:34.056667 kernel: pid_max: default: 32768 minimum: 301 Feb 9 08:58:34.056683 kernel: LSM: Security Framework initializing Feb 9 08:58:34.056698 kernel: SELinux: Initializing. Feb 9 08:58:34.056714 kernel: Mount-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 9 08:58:34.056743 kernel: Mountpoint-cache hash table entries: 4096 (order: 3, 32768 bytes, linear) Feb 9 08:58:34.056759 kernel: smpboot: CPU0: Intel DO-Regular (family: 0x6, model: 0x3f, stepping: 0x2) Feb 9 08:58:34.056779 kernel: Performance Events: unsupported p6 CPU model 63 no PMU driver, software events only. Feb 9 08:58:34.056794 kernel: signal: max sigframe size: 1776 Feb 9 08:58:34.056810 kernel: rcu: Hierarchical SRCU implementation. Feb 9 08:58:34.056827 kernel: NMI watchdog: Perf NMI watchdog permanently disabled Feb 9 08:58:34.056842 kernel: smp: Bringing up secondary CPUs ... Feb 9 08:58:34.056858 kernel: x86: Booting SMP configuration: Feb 9 08:58:34.056874 kernel: .... node #0, CPUs: #1 Feb 9 08:58:34.056890 kernel: kvm-clock: cpu 1, msr 40faa041, secondary cpu clock Feb 9 08:58:34.056903 kernel: kvm-guest: stealtime: cpu 1, msr 7dd1c0c0 Feb 9 08:58:34.056923 kernel: smp: Brought up 1 node, 2 CPUs Feb 9 08:58:34.056939 kernel: smpboot: Max logical packages: 1 Feb 9 08:58:34.056955 kernel: smpboot: Total of 2 processors activated (9976.56 BogoMIPS) Feb 9 08:58:34.056971 kernel: devtmpfs: initialized Feb 9 08:58:34.056986 kernel: x86/mm: Memory block size: 128MB Feb 9 08:58:34.057002 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Feb 9 08:58:34.057018 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Feb 9 08:58:34.057035 kernel: pinctrl core: initialized pinctrl subsystem Feb 9 08:58:34.057051 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Feb 9 08:58:34.057070 kernel: audit: initializing netlink subsys (disabled) Feb 9 08:58:34.057086 kernel: audit: type=2000 audit(1707469113.593:1): state=initialized audit_enabled=0 res=1 Feb 9 08:58:34.057101 kernel: thermal_sys: Registered thermal governor 'step_wise' Feb 9 08:58:34.057117 kernel: thermal_sys: Registered thermal governor 'user_space' Feb 9 08:58:34.057133 kernel: cpuidle: using governor menu Feb 9 08:58:34.057149 kernel: ACPI: bus type PCI registered Feb 9 08:58:34.057164 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Feb 9 08:58:34.057179 kernel: dca service started, version 1.12.1 Feb 9 08:58:34.057195 kernel: PCI: Using configuration type 1 for base access Feb 9 08:58:34.057214 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Feb 9 08:58:34.057238 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Feb 9 08:58:34.057253 kernel: ACPI: Added _OSI(Module Device) Feb 9 08:58:34.057270 kernel: ACPI: Added _OSI(Processor Device) Feb 9 08:58:34.057285 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Feb 9 08:58:34.057301 kernel: ACPI: Added _OSI(Processor Aggregator Device) Feb 9 08:58:34.057317 kernel: ACPI: Added _OSI(Linux-Dell-Video) Feb 9 08:58:34.057333 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Feb 9 08:58:34.057348 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Feb 9 08:58:34.057367 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Feb 9 08:58:34.057383 kernel: ACPI: Interpreter enabled Feb 9 08:58:34.057399 kernel: ACPI: PM: (supports S0 S5) Feb 9 08:58:34.057414 kernel: ACPI: Using IOAPIC for interrupt routing Feb 9 08:58:34.057430 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Feb 9 08:58:34.057445 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Feb 9 08:58:34.057461 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Feb 9 08:58:34.057750 kernel: acpi PNP0A03:00: _OSC: OS supports [ASPM ClockPM Segments MSI HPX-Type3] Feb 9 08:58:34.057909 kernel: acpi PNP0A03:00: fail to add MMCONFIG information, can't access extended PCI configuration space under this bridge. Feb 9 08:58:34.057928 kernel: acpiphp: Slot [3] registered Feb 9 08:58:34.057940 kernel: acpiphp: Slot [4] registered Feb 9 08:58:34.057953 kernel: acpiphp: Slot [5] registered Feb 9 08:58:34.057965 kernel: acpiphp: Slot [6] registered Feb 9 08:58:34.057977 kernel: acpiphp: Slot [7] registered Feb 9 08:58:34.057988 kernel: acpiphp: Slot [8] registered Feb 9 08:58:34.058002 kernel: acpiphp: Slot [9] registered Feb 9 08:58:34.058014 kernel: acpiphp: Slot [10] registered Feb 9 08:58:34.058033 kernel: acpiphp: Slot [11] registered Feb 9 08:58:34.058048 kernel: acpiphp: Slot [12] registered Feb 9 08:58:34.058062 kernel: acpiphp: Slot [13] registered Feb 9 08:58:34.058077 kernel: acpiphp: Slot [14] registered Feb 9 08:58:34.058091 kernel: acpiphp: Slot [15] registered Feb 9 08:58:34.058106 kernel: acpiphp: Slot [16] registered Feb 9 08:58:34.058122 kernel: acpiphp: Slot [17] registered Feb 9 08:58:34.058136 kernel: acpiphp: Slot [18] registered Feb 9 08:58:34.058151 kernel: acpiphp: Slot [19] registered Feb 9 08:58:34.058170 kernel: acpiphp: Slot [20] registered Feb 9 08:58:34.058184 kernel: acpiphp: Slot [21] registered Feb 9 08:58:34.058198 kernel: acpiphp: Slot [22] registered Feb 9 08:58:34.058214 kernel: acpiphp: Slot [23] registered Feb 9 08:58:34.058231 kernel: acpiphp: Slot [24] registered Feb 9 08:58:34.058245 kernel: acpiphp: Slot [25] registered Feb 9 08:58:34.058259 kernel: acpiphp: Slot [26] registered Feb 9 08:58:34.058275 kernel: acpiphp: Slot [27] registered Feb 9 08:58:34.058288 kernel: acpiphp: Slot [28] registered Feb 9 08:58:34.058302 kernel: acpiphp: Slot [29] registered Feb 9 08:58:34.058320 kernel: acpiphp: Slot [30] registered Feb 9 08:58:34.058335 kernel: acpiphp: Slot [31] registered Feb 9 08:58:34.058352 kernel: PCI host bridge to bus 0000:00 Feb 9 08:58:34.058580 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Feb 9 08:58:34.058771 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Feb 9 08:58:34.058922 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Feb 9 08:58:34.059040 kernel: pci_bus 0000:00: root bus resource [mem 0x80000000-0xfebfffff window] Feb 9 08:58:34.059170 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x17fffffff window] Feb 9 08:58:34.059396 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Feb 9 08:58:34.059583 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Feb 9 08:58:34.065518 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Feb 9 08:58:34.066645 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Feb 9 08:58:34.066817 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc1e0-0xc1ef] Feb 9 08:58:34.066967 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Feb 9 08:58:34.067105 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Feb 9 08:58:34.067258 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Feb 9 08:58:34.067405 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Feb 9 08:58:34.067564 kernel: pci 0000:00:01.2: [8086:7020] type 00 class 0x0c0300 Feb 9 08:58:34.067711 kernel: pci 0000:00:01.2: reg 0x20: [io 0xc180-0xc19f] Feb 9 08:58:34.071281 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Feb 9 08:58:34.071450 kernel: pci 0000:00:01.3: quirk: [io 0x0600-0x063f] claimed by PIIX4 ACPI Feb 9 08:58:34.071581 kernel: pci 0000:00:01.3: quirk: [io 0x0700-0x070f] claimed by PIIX4 SMB Feb 9 08:58:34.071774 kernel: pci 0000:00:02.0: [1af4:1050] type 00 class 0x030000 Feb 9 08:58:34.071911 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xfe000000-0xfe7fffff pref] Feb 9 08:58:34.072053 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xfe800000-0xfe803fff 64bit pref] Feb 9 08:58:34.072183 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfebf0000-0xfebf0fff] Feb 9 08:58:34.072311 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xfebe0000-0xfebeffff pref] Feb 9 08:58:34.072446 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Feb 9 08:58:34.072593 kernel: pci 0000:00:03.0: [1af4:1000] type 00 class 0x020000 Feb 9 08:58:34.072754 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc1a0-0xc1bf] Feb 9 08:58:34.072888 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebf1000-0xfebf1fff] Feb 9 08:58:34.073017 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe804000-0xfe807fff 64bit pref] Feb 9 08:58:34.073159 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Feb 9 08:58:34.073298 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc1c0-0xc1df] Feb 9 08:58:34.073428 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebf2000-0xfebf2fff] Feb 9 08:58:34.073556 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe808000-0xfe80bfff 64bit pref] Feb 9 08:58:34.073699 kernel: pci 0000:00:05.0: [1af4:1004] type 00 class 0x010000 Feb 9 08:58:34.073845 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc100-0xc13f] Feb 9 08:58:34.073974 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xfebf3000-0xfebf3fff] Feb 9 08:58:34.074107 kernel: pci 0000:00:05.0: reg 0x20: [mem 0xfe80c000-0xfe80ffff 64bit pref] Feb 9 08:58:34.074261 kernel: pci 0000:00:06.0: [1af4:1001] type 00 class 0x010000 Feb 9 08:58:34.074399 kernel: pci 0000:00:06.0: reg 0x10: [io 0xc000-0xc07f] Feb 9 08:58:34.074531 kernel: pci 0000:00:06.0: reg 0x14: [mem 0xfebf4000-0xfebf4fff] Feb 9 08:58:34.074662 kernel: pci 0000:00:06.0: reg 0x20: [mem 0xfe810000-0xfe813fff 64bit pref] Feb 9 08:58:34.075907 kernel: pci 0000:00:07.0: [1af4:1001] type 00 class 0x010000 Feb 9 08:58:34.076060 kernel: pci 0000:00:07.0: reg 0x10: [io 0xc080-0xc0ff] Feb 9 08:58:34.076196 kernel: pci 0000:00:07.0: reg 0x14: [mem 0xfebf5000-0xfebf5fff] Feb 9 08:58:34.076335 kernel: pci 0000:00:07.0: reg 0x20: [mem 0xfe814000-0xfe817fff 64bit pref] Feb 9 08:58:34.076501 kernel: pci 0000:00:08.0: [1af4:1002] type 00 class 0x00ff00 Feb 9 08:58:34.076634 kernel: pci 0000:00:08.0: reg 0x10: [io 0xc140-0xc17f] Feb 9 08:58:34.077879 kernel: pci 0000:00:08.0: reg 0x20: [mem 0xfe818000-0xfe81bfff 64bit pref] Feb 9 08:58:34.077908 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Feb 9 08:58:34.077924 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Feb 9 08:58:34.077940 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Feb 9 08:58:34.077961 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Feb 9 08:58:34.077976 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Feb 9 08:58:34.077992 kernel: iommu: Default domain type: Translated Feb 9 08:58:34.078007 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Feb 9 08:58:34.078153 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Feb 9 08:58:34.078291 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Feb 9 08:58:34.078425 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Feb 9 08:58:34.078445 kernel: vgaarb: loaded Feb 9 08:58:34.078461 kernel: pps_core: LinuxPPS API ver. 1 registered Feb 9 08:58:34.078482 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Feb 9 08:58:34.078498 kernel: PTP clock support registered Feb 9 08:58:34.078513 kernel: PCI: Using ACPI for IRQ routing Feb 9 08:58:34.078528 kernel: PCI: pci_cache_line_size set to 64 bytes Feb 9 08:58:34.078543 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Feb 9 08:58:34.078558 kernel: e820: reserve RAM buffer [mem 0x7ffdb000-0x7fffffff] Feb 9 08:58:34.078573 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Feb 9 08:58:34.078588 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Feb 9 08:58:34.078603 kernel: clocksource: Switched to clocksource kvm-clock Feb 9 08:58:34.078622 kernel: VFS: Disk quotas dquot_6.6.0 Feb 9 08:58:34.078638 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Feb 9 08:58:34.078654 kernel: pnp: PnP ACPI init Feb 9 08:58:34.078669 kernel: pnp: PnP ACPI: found 4 devices Feb 9 08:58:34.078684 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Feb 9 08:58:34.078700 kernel: NET: Registered PF_INET protocol family Feb 9 08:58:34.078715 kernel: IP idents hash table entries: 32768 (order: 6, 262144 bytes, linear) Feb 9 08:58:34.078742 kernel: tcp_listen_portaddr_hash hash table entries: 1024 (order: 2, 16384 bytes, linear) Feb 9 08:58:34.078761 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Feb 9 08:58:34.078776 kernel: TCP established hash table entries: 16384 (order: 5, 131072 bytes, linear) Feb 9 08:58:34.078791 kernel: TCP bind hash table entries: 16384 (order: 6, 262144 bytes, linear) Feb 9 08:58:34.078807 kernel: TCP: Hash tables configured (established 16384 bind 16384) Feb 9 08:58:34.078822 kernel: UDP hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 9 08:58:34.078837 kernel: UDP-Lite hash table entries: 1024 (order: 3, 32768 bytes, linear) Feb 9 08:58:34.078852 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Feb 9 08:58:34.078868 kernel: NET: Registered PF_XDP protocol family Feb 9 08:58:34.079004 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Feb 9 08:58:34.079132 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Feb 9 08:58:34.079267 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Feb 9 08:58:34.079384 kernel: pci_bus 0000:00: resource 7 [mem 0x80000000-0xfebfffff window] Feb 9 08:58:34.079504 kernel: pci_bus 0000:00: resource 8 [mem 0x100000000-0x17fffffff window] Feb 9 08:58:34.079644 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Feb 9 08:58:34.081866 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Feb 9 08:58:34.082023 kernel: pci 0000:00:01.0: Activating ISA DMA hang workarounds Feb 9 08:58:34.082045 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 11 Feb 9 08:58:34.082191 kernel: pci 0000:00:01.2: quirk_usb_early_handoff+0x0/0x730 took 33913 usecs Feb 9 08:58:34.082210 kernel: PCI: CLS 0 bytes, default 64 Feb 9 08:58:34.082225 kernel: RAPL PMU: API unit is 2^-32 Joules, 0 fixed counters, 10737418240 ms ovfl timer Feb 9 08:58:34.082241 kernel: clocksource: tsc: mask: 0xffffffffffffffff max_cycles: 0x23f39a1d859, max_idle_ns: 440795326830 ns Feb 9 08:58:34.082257 kernel: Initialise system trusted keyrings Feb 9 08:58:34.082272 kernel: workingset: timestamp_bits=39 max_order=19 bucket_order=0 Feb 9 08:58:34.082287 kernel: Key type asymmetric registered Feb 9 08:58:34.082302 kernel: Asymmetric key parser 'x509' registered Feb 9 08:58:34.082317 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Feb 9 08:58:34.082337 kernel: io scheduler mq-deadline registered Feb 9 08:58:34.082353 kernel: io scheduler kyber registered Feb 9 08:58:34.082368 kernel: io scheduler bfq registered Feb 9 08:58:34.082384 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Feb 9 08:58:34.082400 kernel: ACPI: \_SB_.LNKB: Enabled at IRQ 10 Feb 9 08:58:34.082415 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Feb 9 08:58:34.082431 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Feb 9 08:58:34.082446 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Feb 9 08:58:34.082462 kernel: 00:00: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Feb 9 08:58:34.082480 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Feb 9 08:58:34.082495 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Feb 9 08:58:34.082511 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Feb 9 08:58:34.082677 kernel: rtc_cmos 00:03: RTC can wake from S4 Feb 9 08:58:34.082699 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Feb 9 08:58:34.083878 kernel: rtc_cmos 00:03: registered as rtc0 Feb 9 08:58:34.084014 kernel: rtc_cmos 00:03: setting system clock to 2024-02-09T08:58:33 UTC (1707469113) Feb 9 08:58:34.084139 kernel: rtc_cmos 00:03: alarms up to one day, y3k, 242 bytes nvram Feb 9 08:58:34.084163 kernel: intel_pstate: CPU model not supported Feb 9 08:58:34.084179 kernel: NET: Registered PF_INET6 protocol family Feb 9 08:58:34.084195 kernel: Segment Routing with IPv6 Feb 9 08:58:34.084210 kernel: In-situ OAM (IOAM) with IPv6 Feb 9 08:58:34.084226 kernel: NET: Registered PF_PACKET protocol family Feb 9 08:58:34.084241 kernel: Key type dns_resolver registered Feb 9 08:58:34.084256 kernel: IPI shorthand broadcast: enabled Feb 9 08:58:34.084271 kernel: sched_clock: Marking stable (705271553, 80436010)->(901173721, -115466158) Feb 9 08:58:34.084286 kernel: registered taskstats version 1 Feb 9 08:58:34.084305 kernel: Loading compiled-in X.509 certificates Feb 9 08:58:34.084321 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.148-flatcar: e9d857ae0e8100c174221878afd1046acbb054a6' Feb 9 08:58:34.084336 kernel: Key type .fscrypt registered Feb 9 08:58:34.084351 kernel: Key type fscrypt-provisioning registered Feb 9 08:58:34.084367 kernel: ima: No TPM chip found, activating TPM-bypass! Feb 9 08:58:34.084382 kernel: ima: Allocated hash algorithm: sha1 Feb 9 08:58:34.084397 kernel: ima: No architecture policies found Feb 9 08:58:34.084412 kernel: Freeing unused kernel image (initmem) memory: 45496K Feb 9 08:58:34.084430 kernel: Write protecting the kernel read-only data: 28672k Feb 9 08:58:34.084446 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Feb 9 08:58:34.084460 kernel: Freeing unused kernel image (rodata/data gap) memory: 636K Feb 9 08:58:34.084476 kernel: Run /init as init process Feb 9 08:58:34.084491 kernel: with arguments: Feb 9 08:58:34.084507 kernel: /init Feb 9 08:58:34.084548 kernel: with environment: Feb 9 08:58:34.084566 kernel: HOME=/ Feb 9 08:58:34.084582 kernel: TERM=linux Feb 9 08:58:34.084601 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Feb 9 08:58:34.084622 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Feb 9 08:58:34.084642 systemd[1]: Detected virtualization kvm. Feb 9 08:58:34.084656 systemd[1]: Detected architecture x86-64. Feb 9 08:58:34.084669 systemd[1]: Running in initrd. Feb 9 08:58:34.084681 systemd[1]: No hostname configured, using default hostname. Feb 9 08:58:34.084696 systemd[1]: Hostname set to . Feb 9 08:58:34.084713 systemd[1]: Initializing machine ID from VM UUID. Feb 9 08:58:34.084742 systemd[1]: Queued start job for default target initrd.target. Feb 9 08:58:34.084759 systemd[1]: Started systemd-ask-password-console.path. Feb 9 08:58:34.084775 systemd[1]: Reached target cryptsetup.target. Feb 9 08:58:34.084788 systemd[1]: Reached target paths.target. Feb 9 08:58:34.084800 systemd[1]: Reached target slices.target. Feb 9 08:58:34.084814 systemd[1]: Reached target swap.target. Feb 9 08:58:34.084828 systemd[1]: Reached target timers.target. Feb 9 08:58:34.084847 systemd[1]: Listening on iscsid.socket. Feb 9 08:58:34.084862 systemd[1]: Listening on iscsiuio.socket. Feb 9 08:58:34.084876 systemd[1]: Listening on systemd-journald-audit.socket. Feb 9 08:58:34.084893 systemd[1]: Listening on systemd-journald-dev-log.socket. Feb 9 08:58:34.084910 systemd[1]: Listening on systemd-journald.socket. Feb 9 08:58:34.084927 systemd[1]: Listening on systemd-networkd.socket. Feb 9 08:58:34.084943 systemd[1]: Listening on systemd-udevd-control.socket. Feb 9 08:58:34.084959 systemd[1]: Listening on systemd-udevd-kernel.socket. Feb 9 08:58:34.084974 systemd[1]: Reached target sockets.target. Feb 9 08:58:34.084995 systemd[1]: Starting kmod-static-nodes.service... Feb 9 08:58:34.085013 systemd[1]: Finished network-cleanup.service. Feb 9 08:58:34.085033 systemd[1]: Starting systemd-fsck-usr.service... Feb 9 08:58:34.085049 systemd[1]: Starting systemd-journald.service... Feb 9 08:58:34.085065 systemd[1]: Starting systemd-modules-load.service... Feb 9 08:58:34.085083 systemd[1]: Starting systemd-resolved.service... Feb 9 08:58:34.085099 systemd[1]: Starting systemd-vconsole-setup.service... Feb 9 08:58:34.085115 systemd[1]: Finished kmod-static-nodes.service. Feb 9 08:58:34.085132 systemd[1]: Finished systemd-fsck-usr.service. Feb 9 08:58:34.085150 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Feb 9 08:58:34.085167 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Feb 9 08:58:34.085183 kernel: audit: type=1130 audit(1707469114.057:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.085197 systemd[1]: Finished systemd-vconsole-setup.service. Feb 9 08:58:34.085214 kernel: audit: type=1130 audit(1707469114.061:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.085234 systemd[1]: Starting dracut-cmdline-ask.service... Feb 9 08:58:34.085252 systemd[1]: Started systemd-resolved.service. Feb 9 08:58:34.085269 kernel: audit: type=1130 audit(1707469114.067:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.085286 systemd[1]: Reached target nss-lookup.target. Feb 9 08:58:34.085308 systemd-journald[184]: Journal started Feb 9 08:58:34.085396 systemd-journald[184]: Runtime Journal (/run/log/journal/0f2d3f05f4034df2965e592d36475234) is 4.9M, max 39.5M, 34.5M free. Feb 9 08:58:34.057000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.061000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.067000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.041821 systemd-modules-load[185]: Inserted module 'overlay' Feb 9 08:58:34.090310 systemd[1]: Started systemd-journald.service. Feb 9 08:58:34.048475 systemd-resolved[186]: Positive Trust Anchors: Feb 9 08:58:34.101800 kernel: audit: type=1130 audit(1707469114.089:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.089000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.048489 systemd-resolved[186]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 9 08:58:34.048549 systemd-resolved[186]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Feb 9 08:58:34.116016 kernel: audit: type=1130 audit(1707469114.101:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.116054 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Feb 9 08:58:34.101000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.055041 systemd-resolved[186]: Defaulting to hostname 'linux'. Feb 9 08:58:34.102024 systemd[1]: Finished dracut-cmdline-ask.service. Feb 9 08:58:34.104025 systemd[1]: Starting dracut-cmdline.service... Feb 9 08:58:34.120473 dracut-cmdline[201]: dracut-dracut-053 Feb 9 08:58:34.123765 kernel: Bridge firewalling registered Feb 9 08:58:34.123360 systemd-modules-load[185]: Inserted module 'br_netfilter' Feb 9 08:58:34.124371 dracut-cmdline[201]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200n8 console=tty0 flatcar.first_boot=detected flatcar.oem.id=digitalocean verity.usrhash=ae7db544026ede4699ee2036449b75950d3fb7929b25a6731d0ad396f1aa37c9 Feb 9 08:58:34.157753 kernel: SCSI subsystem initialized Feb 9 08:58:34.176758 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Feb 9 08:58:34.179934 kernel: device-mapper: uevent: version 1.0.3 Feb 9 08:58:34.180014 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Feb 9 08:58:34.187856 systemd-modules-load[185]: Inserted module 'dm_multipath' Feb 9 08:58:34.189551 systemd[1]: Finished systemd-modules-load.service. Feb 9 08:58:34.203514 kernel: audit: type=1130 audit(1707469114.189:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.189000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.192264 systemd[1]: Starting systemd-sysctl.service... Feb 9 08:58:34.212653 systemd[1]: Finished systemd-sysctl.service. Feb 9 08:58:34.217749 kernel: audit: type=1130 audit(1707469114.212:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.212000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.240757 kernel: Loading iSCSI transport class v2.0-870. Feb 9 08:58:34.258746 kernel: iscsi: registered transport (tcp) Feb 9 08:58:34.291773 kernel: iscsi: registered transport (qla4xxx) Feb 9 08:58:34.291875 kernel: QLogic iSCSI HBA Driver Feb 9 08:58:34.344236 systemd[1]: Finished dracut-cmdline.service. Feb 9 08:58:34.343000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.345971 systemd[1]: Starting dracut-pre-udev.service... Feb 9 08:58:34.349768 kernel: audit: type=1130 audit(1707469114.343:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.403782 kernel: raid6: avx2x4 gen() 15046 MB/s Feb 9 08:58:34.420781 kernel: raid6: avx2x4 xor() 8153 MB/s Feb 9 08:58:34.437807 kernel: raid6: avx2x2 gen() 12701 MB/s Feb 9 08:58:34.454834 kernel: raid6: avx2x2 xor() 13721 MB/s Feb 9 08:58:34.471806 kernel: raid6: avx2x1 gen() 11119 MB/s Feb 9 08:58:34.488776 kernel: raid6: avx2x1 xor() 14431 MB/s Feb 9 08:58:34.505789 kernel: raid6: sse2x4 gen() 11116 MB/s Feb 9 08:58:34.522786 kernel: raid6: sse2x4 xor() 4858 MB/s Feb 9 08:58:34.539784 kernel: raid6: sse2x2 gen() 10371 MB/s Feb 9 08:58:34.556779 kernel: raid6: sse2x2 xor() 7853 MB/s Feb 9 08:58:34.573775 kernel: raid6: sse2x1 gen() 10911 MB/s Feb 9 08:58:34.590914 kernel: raid6: sse2x1 xor() 6024 MB/s Feb 9 08:58:34.591005 kernel: raid6: using algorithm avx2x4 gen() 15046 MB/s Feb 9 08:58:34.591041 kernel: raid6: .... xor() 8153 MB/s, rmw enabled Feb 9 08:58:34.592008 kernel: raid6: using avx2x2 recovery algorithm Feb 9 08:58:34.604752 kernel: xor: automatically using best checksumming function avx Feb 9 08:58:34.714764 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Feb 9 08:58:34.728761 kernel: audit: type=1130 audit(1707469114.725:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.725000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.728000 audit: BPF prog-id=7 op=LOAD Feb 9 08:58:34.726121 systemd[1]: Finished dracut-pre-udev.service. Feb 9 08:58:34.728000 audit: BPF prog-id=8 op=LOAD Feb 9 08:58:34.729475 systemd[1]: Starting systemd-udevd.service... Feb 9 08:58:34.748968 systemd-udevd[383]: Using default interface naming scheme 'v252'. Feb 9 08:58:34.757000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.757512 systemd[1]: Started systemd-udevd.service. Feb 9 08:58:34.759164 systemd[1]: Starting dracut-pre-trigger.service... Feb 9 08:58:34.779768 dracut-pre-trigger[388]: rd.md=0: removing MD RAID activation Feb 9 08:58:34.823550 systemd[1]: Finished dracut-pre-trigger.service. Feb 9 08:58:34.823000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.825255 systemd[1]: Starting systemd-udev-trigger.service... Feb 9 08:58:34.886000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:34.886687 systemd[1]: Finished systemd-udev-trigger.service. Feb 9 08:58:34.957185 kernel: virtio_blk virtio4: [vda] 125829120 512-byte logical blocks (64.4 GB/60.0 GiB) Feb 9 08:58:34.965753 kernel: scsi host0: Virtio SCSI HBA Feb 9 08:58:34.968743 kernel: cryptd: max_cpu_qlen set to 1000 Feb 9 08:58:34.982074 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Feb 9 08:58:34.982144 kernel: GPT:9289727 != 125829119 Feb 9 08:58:34.982163 kernel: GPT:Alternate GPT header not at the end of the disk. Feb 9 08:58:34.982885 kernel: GPT:9289727 != 125829119 Feb 9 08:58:34.984167 kernel: GPT: Use GNU Parted to correct GPT errors. Feb 9 08:58:34.984202 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 9 08:58:34.992764 kernel: virtio_blk virtio5: [vdb] 1000 512-byte logical blocks (512 kB/500 KiB) Feb 9 08:58:35.004396 kernel: ACPI: bus type USB registered Feb 9 08:58:35.004452 kernel: usbcore: registered new interface driver usbfs Feb 9 08:58:35.005272 kernel: usbcore: registered new interface driver hub Feb 9 08:58:35.008922 kernel: usbcore: registered new device driver usb Feb 9 08:58:35.027964 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Feb 9 08:58:35.034292 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Feb 9 08:58:35.035403 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Feb 9 08:58:35.038742 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (437) Feb 9 08:58:35.040142 systemd[1]: Starting disk-uuid.service... Feb 9 08:58:35.046894 kernel: libata version 3.00 loaded. Feb 9 08:58:35.047638 disk-uuid[455]: Primary Header is updated. Feb 9 08:58:35.047638 disk-uuid[455]: Secondary Entries is updated. Feb 9 08:58:35.047638 disk-uuid[455]: Secondary Header is updated. Feb 9 08:58:35.055740 kernel: ehci_hcd: USB 2.0 'Enhanced' Host Controller (EHCI) Driver Feb 9 08:58:35.058741 kernel: ehci-pci: EHCI PCI platform driver Feb 9 08:58:35.062920 kernel: ata_piix 0000:00:01.1: version 2.13 Feb 9 08:58:35.067317 kernel: AVX2 version of gcm_enc/dec engaged. Feb 9 08:58:35.066864 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Feb 9 08:58:35.071757 kernel: AES CTR mode by8 optimization enabled Feb 9 08:58:35.078087 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Feb 9 08:58:35.140761 kernel: scsi host1: ata_piix Feb 9 08:58:35.141010 kernel: scsi host2: ata_piix Feb 9 08:58:35.141214 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc1e0 irq 14 Feb 9 08:58:35.141236 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc1e8 irq 15 Feb 9 08:58:35.141260 kernel: uhci_hcd: USB Universal Host Controller Interface driver Feb 9 08:58:35.148369 kernel: uhci_hcd 0000:00:01.2: UHCI Host Controller Feb 9 08:58:35.148610 kernel: uhci_hcd 0000:00:01.2: new USB bus registered, assigned bus number 1 Feb 9 08:58:35.148745 kernel: uhci_hcd 0000:00:01.2: detected 2 ports Feb 9 08:58:35.151745 kernel: uhci_hcd 0000:00:01.2: irq 11, io base 0x0000c180 Feb 9 08:58:35.151994 kernel: hub 1-0:1.0: USB hub found Feb 9 08:58:35.153037 kernel: hub 1-0:1.0: 2 ports detected Feb 9 08:58:36.059880 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Feb 9 08:58:36.062753 disk-uuid[456]: The operation has completed successfully. Feb 9 08:58:36.101061 systemd[1]: disk-uuid.service: Deactivated successfully. Feb 9 08:58:36.100000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.100000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.101174 systemd[1]: Finished disk-uuid.service. Feb 9 08:58:36.102647 systemd[1]: Starting verity-setup.service... Feb 9 08:58:36.122752 kernel: device-mapper: verity: sha256 using implementation "sha256-avx2" Feb 9 08:58:36.181254 systemd[1]: Found device dev-mapper-usr.device. Feb 9 08:58:36.182642 systemd[1]: Finished verity-setup.service. Feb 9 08:58:36.182000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.185035 systemd[1]: Mounting sysusr-usr.mount... Feb 9 08:58:36.274784 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Feb 9 08:58:36.274395 systemd[1]: Mounted sysusr-usr.mount. Feb 9 08:58:36.275088 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Feb 9 08:58:36.276274 systemd[1]: Starting ignition-setup.service... Feb 9 08:58:36.279485 systemd[1]: Starting parse-ip-for-networkd.service... Feb 9 08:58:36.295892 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 9 08:58:36.295971 kernel: BTRFS info (device vda6): using free space tree Feb 9 08:58:36.295996 kernel: BTRFS info (device vda6): has skinny extents Feb 9 08:58:36.315946 systemd[1]: mnt-oem.mount: Deactivated successfully. Feb 9 08:58:36.325938 systemd[1]: Finished ignition-setup.service. Feb 9 08:58:36.325000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.327549 systemd[1]: Starting ignition-fetch-offline.service... Feb 9 08:58:36.450000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.451000 audit: BPF prog-id=9 op=LOAD Feb 9 08:58:36.450780 systemd[1]: Finished parse-ip-for-networkd.service. Feb 9 08:58:36.453213 systemd[1]: Starting systemd-networkd.service... Feb 9 08:58:36.478559 ignition[608]: Ignition 2.14.0 Feb 9 08:58:36.478570 ignition[608]: Stage: fetch-offline Feb 9 08:58:36.478645 ignition[608]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:36.478672 ignition[608]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:36.483644 ignition[608]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:36.483814 ignition[608]: parsed url from cmdline: "" Feb 9 08:58:36.483819 ignition[608]: no config URL provided Feb 9 08:58:36.483825 ignition[608]: reading system config file "/usr/lib/ignition/user.ign" Feb 9 08:58:36.485000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.485317 systemd[1]: Finished ignition-fetch-offline.service. Feb 9 08:58:36.483835 ignition[608]: no config at "/usr/lib/ignition/user.ign" Feb 9 08:58:36.483841 ignition[608]: failed to fetch config: resource requires networking Feb 9 08:58:36.483957 ignition[608]: Ignition finished successfully Feb 9 08:58:36.492685 systemd-networkd[688]: lo: Link UP Feb 9 08:58:36.494000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.492701 systemd-networkd[688]: lo: Gained carrier Feb 9 08:58:36.493558 systemd-networkd[688]: Enumeration completed Feb 9 08:58:36.494154 systemd-networkd[688]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Feb 9 08:58:36.494270 systemd[1]: Started systemd-networkd.service. Feb 9 08:58:36.495274 systemd[1]: Reached target network.target. Feb 9 08:58:36.495911 systemd-networkd[688]: eth0: Configuring with /usr/lib/systemd/network/yy-digitalocean.network. Feb 9 08:58:36.496843 systemd-networkd[688]: eth1: Link UP Feb 9 08:58:36.496847 systemd-networkd[688]: eth1: Gained carrier Feb 9 08:58:36.497836 systemd[1]: Starting ignition-fetch.service... Feb 9 08:58:36.499711 systemd[1]: Starting iscsiuio.service... Feb 9 08:58:36.511630 systemd-networkd[688]: eth0: Link UP Feb 9 08:58:36.511637 systemd-networkd[688]: eth0: Gained carrier Feb 9 08:58:36.521990 ignition[690]: Ignition 2.14.0 Feb 9 08:58:36.522003 ignition[690]: Stage: fetch Feb 9 08:58:36.522149 ignition[690]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:36.522175 ignition[690]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:36.526002 systemd[1]: Started iscsiuio.service. Feb 9 08:58:36.525000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.524555 ignition[690]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:36.538000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.526885 systemd-networkd[688]: eth1: DHCPv4 address 10.124.0.3/20 acquired from 169.254.169.253 Feb 9 08:58:36.524685 ignition[690]: parsed url from cmdline: "" Feb 9 08:58:36.527817 systemd[1]: Starting iscsid.service... Feb 9 08:58:36.552875 iscsid[698]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Feb 9 08:58:36.552875 iscsid[698]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log Feb 9 08:58:36.552875 iscsid[698]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Feb 9 08:58:36.552875 iscsid[698]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Feb 9 08:58:36.552875 iscsid[698]: If using hardware iscsi like qla4xxx this message can be ignored. Feb 9 08:58:36.552875 iscsid[698]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Feb 9 08:58:36.552875 iscsid[698]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Feb 9 08:58:36.524689 ignition[690]: no config URL provided Feb 9 08:58:36.530902 systemd-networkd[688]: eth0: DHCPv4 address 64.23.148.25/20, gateway 64.23.144.1 acquired from 169.254.169.253 Feb 9 08:58:36.524694 ignition[690]: reading system config file "/usr/lib/ignition/user.ign" Feb 9 08:58:36.537582 systemd[1]: Started iscsid.service. Feb 9 08:58:36.524703 ignition[690]: no config at "/usr/lib/ignition/user.ign" Feb 9 08:58:36.540870 systemd[1]: Starting dracut-initqueue.service... Feb 9 08:58:36.524745 ignition[690]: GET http://169.254.169.254/metadata/v1/user-data: attempt #1 Feb 9 08:58:36.569879 systemd[1]: Finished dracut-initqueue.service. Feb 9 08:58:36.570000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.571024 systemd[1]: Reached target remote-fs-pre.target. Feb 9 08:58:36.571881 systemd[1]: Reached target remote-cryptsetup.target. Feb 9 08:58:36.572682 systemd[1]: Reached target remote-fs.target. Feb 9 08:58:36.572874 ignition[690]: GET result: OK Feb 9 08:58:36.576812 systemd[1]: Starting dracut-pre-mount.service... Feb 9 08:58:36.573102 ignition[690]: parsing config with SHA512: 883647f0949a1219aadd0a59c12ad90d396eb85275c69cec59b045cf052cc23a694fad9a90e44bd39ef9bb8bdaab4b89adbef762949595bde2ab03ad0209999a Feb 9 08:58:36.598000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.598505 systemd[1]: Finished dracut-pre-mount.service. Feb 9 08:58:36.648007 unknown[690]: fetched base config from "system" Feb 9 08:58:36.648691 unknown[690]: fetched base config from "system" Feb 9 08:58:36.649203 unknown[690]: fetched user config from "digitalocean" Feb 9 08:58:36.650438 ignition[690]: fetch: fetch complete Feb 9 08:58:36.650994 ignition[690]: fetch: fetch passed Feb 9 08:58:36.651552 ignition[690]: Ignition finished successfully Feb 9 08:58:36.654286 systemd[1]: Finished ignition-fetch.service. Feb 9 08:58:36.654000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.656240 systemd[1]: Starting ignition-kargs.service... Feb 9 08:58:36.675044 ignition[713]: Ignition 2.14.0 Feb 9 08:58:36.675785 ignition[713]: Stage: kargs Feb 9 08:58:36.676008 ignition[713]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:36.676035 ignition[713]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:36.678984 ignition[713]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:36.683705 ignition[713]: kargs: kargs passed Feb 9 08:58:36.683829 ignition[713]: Ignition finished successfully Feb 9 08:58:36.685696 systemd[1]: Finished ignition-kargs.service. Feb 9 08:58:36.685000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.687768 systemd[1]: Starting ignition-disks.service... Feb 9 08:58:36.700069 ignition[719]: Ignition 2.14.0 Feb 9 08:58:36.700083 ignition[719]: Stage: disks Feb 9 08:58:36.700250 ignition[719]: reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:36.700278 ignition[719]: parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:36.703097 ignition[719]: no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:36.706171 ignition[719]: disks: disks passed Feb 9 08:58:36.706256 ignition[719]: Ignition finished successfully Feb 9 08:58:36.708152 systemd[1]: Finished ignition-disks.service. Feb 9 08:58:36.707000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.708687 systemd[1]: Reached target initrd-root-device.target. Feb 9 08:58:36.709067 systemd[1]: Reached target local-fs-pre.target. Feb 9 08:58:36.709880 systemd[1]: Reached target local-fs.target. Feb 9 08:58:36.710635 systemd[1]: Reached target sysinit.target. Feb 9 08:58:36.711441 systemd[1]: Reached target basic.target. Feb 9 08:58:36.713223 systemd[1]: Starting systemd-fsck-root.service... Feb 9 08:58:36.732525 systemd-fsck[727]: ROOT: clean, 602/553520 files, 56014/553472 blocks Feb 9 08:58:36.736503 systemd[1]: Finished systemd-fsck-root.service. Feb 9 08:58:36.736000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.738640 systemd[1]: Mounting sysroot.mount... Feb 9 08:58:36.752764 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Feb 9 08:58:36.753546 systemd[1]: Mounted sysroot.mount. Feb 9 08:58:36.754815 systemd[1]: Reached target initrd-root-fs.target. Feb 9 08:58:36.757830 systemd[1]: Mounting sysroot-usr.mount... Feb 9 08:58:36.760366 systemd[1]: Starting flatcar-digitalocean-network.service... Feb 9 08:58:36.763976 systemd[1]: Starting flatcar-metadata-hostname.service... Feb 9 08:58:36.765271 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Feb 9 08:58:36.765944 systemd[1]: Reached target ignition-diskful.target. Feb 9 08:58:36.773529 systemd[1]: Mounted sysroot-usr.mount. Feb 9 08:58:36.778146 systemd[1]: Starting initrd-setup-root.service... Feb 9 08:58:36.791623 initrd-setup-root[739]: cut: /sysroot/etc/passwd: No such file or directory Feb 9 08:58:36.806263 initrd-setup-root[747]: cut: /sysroot/etc/group: No such file or directory Feb 9 08:58:36.819412 initrd-setup-root[757]: cut: /sysroot/etc/shadow: No such file or directory Feb 9 08:58:36.827394 initrd-setup-root[765]: cut: /sysroot/etc/gshadow: No such file or directory Feb 9 08:58:36.921578 systemd[1]: Finished initrd-setup-root.service. Feb 9 08:58:36.921000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.923618 systemd[1]: Starting ignition-mount.service... Feb 9 08:58:36.927199 systemd[1]: Starting sysroot-boot.service... Feb 9 08:58:36.938594 coreos-metadata[734]: Feb 09 08:58:36.938 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 08:58:36.943684 bash[784]: umount: /sysroot/usr/share/oem: not mounted. Feb 9 08:58:36.965317 coreos-metadata[734]: Feb 09 08:58:36.965 INFO Fetch successful Feb 9 08:58:36.969713 ignition[786]: INFO : Ignition 2.14.0 Feb 9 08:58:36.970598 ignition[786]: INFO : Stage: mount Feb 9 08:58:36.971449 ignition[786]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:36.972285 ignition[786]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:36.976639 ignition[786]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:36.978016 coreos-metadata[734]: Feb 09 08:58:36.977 INFO wrote hostname ci-3510.3.2-5-6e53ffd492 to /sysroot/etc/hostname Feb 9 08:58:36.979020 systemd[1]: Finished flatcar-metadata-hostname.service. Feb 9 08:58:36.978000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.985317 ignition[786]: INFO : mount: mount passed Feb 9 08:58:36.985991 ignition[786]: INFO : Ignition finished successfully Feb 9 08:58:36.986628 coreos-metadata[733]: Feb 09 08:58:36.986 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 08:58:36.989789 systemd[1]: Finished ignition-mount.service. Feb 9 08:58:36.989000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:36.993062 systemd[1]: Finished sysroot-boot.service. Feb 9 08:58:36.992000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:37.000089 coreos-metadata[733]: Feb 09 08:58:37.000 INFO Fetch successful Feb 9 08:58:37.007662 systemd[1]: flatcar-digitalocean-network.service: Deactivated successfully. Feb 9 08:58:37.007815 systemd[1]: Finished flatcar-digitalocean-network.service. Feb 9 08:58:37.008000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-digitalocean-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:37.008000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-digitalocean-network comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:37.201951 systemd[1]: Mounting sysroot-usr-share-oem.mount... Feb 9 08:58:37.210751 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (793) Feb 9 08:58:37.213330 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Feb 9 08:58:37.213411 kernel: BTRFS info (device vda6): using free space tree Feb 9 08:58:37.213429 kernel: BTRFS info (device vda6): has skinny extents Feb 9 08:58:37.219510 systemd[1]: Mounted sysroot-usr-share-oem.mount. Feb 9 08:58:37.221929 systemd[1]: Starting ignition-files.service... Feb 9 08:58:37.250489 ignition[813]: INFO : Ignition 2.14.0 Feb 9 08:58:37.250489 ignition[813]: INFO : Stage: files Feb 9 08:58:37.251807 ignition[813]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:37.251807 ignition[813]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:37.253536 ignition[813]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:37.258217 ignition[813]: DEBUG : files: compiled without relabeling support, skipping Feb 9 08:58:37.258928 ignition[813]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Feb 9 08:58:37.258928 ignition[813]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Feb 9 08:58:37.263630 ignition[813]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Feb 9 08:58:37.264540 ignition[813]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Feb 9 08:58:37.265244 ignition[813]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Feb 9 08:58:37.264863 unknown[813]: wrote ssh authorized keys file for user: core Feb 9 08:58:37.267078 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Feb 9 08:58:37.267078 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Feb 9 08:58:37.294254 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Feb 9 08:58:37.355816 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Feb 9 08:58:37.356715 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/crictl-v1.26.0-linux-amd64.tar.gz" Feb 9 08:58:37.357769 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.26.0/crictl-v1.26.0-linux-amd64.tar.gz: attempt #1 Feb 9 08:58:37.619013 systemd-networkd[688]: eth0: Gained IPv6LL Feb 9 08:58:37.856173 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Feb 9 08:58:37.969729 ignition[813]: DEBUG : files: createFilesystemsFiles: createFiles: op(4): file matches expected sum of: a3a2c02a90b008686c20babaf272e703924db2a3e2a0d4e2a7c81d994cbc68c47458a4a354ecc243af095b390815c7f203348b9749351ae817bd52a522300449 Feb 9 08:58:37.970854 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/crictl-v1.26.0-linux-amd64.tar.gz" Feb 9 08:58:37.970854 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/cni-plugins-linux-amd64-v1.1.1.tgz" Feb 9 08:58:37.970854 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/containernetworking/plugins/releases/download/v1.1.1/cni-plugins-linux-amd64-v1.1.1.tgz: attempt #1 Feb 9 08:58:38.130998 systemd-networkd[688]: eth1: Gained IPv6LL Feb 9 08:58:38.398457 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Feb 9 08:58:38.673313 ignition[813]: DEBUG : files: createFilesystemsFiles: createFiles: op(5): file matches expected sum of: 4d0ed0abb5951b9cf83cba938ef84bdc5b681f4ac869da8143974f6a53a3ff30c666389fa462b9d14d30af09bf03f6cdf77598c572f8fb3ea00cecdda467a48d Feb 9 08:58:38.673313 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/cni-plugins-linux-amd64-v1.1.1.tgz" Feb 9 08:58:38.675973 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/etc/docker/daemon.json" Feb 9 08:58:38.675973 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/etc/docker/daemon.json" Feb 9 08:58:38.675973 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/bin/kubectl" Feb 9 08:58:38.675973 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://dl.k8s.io/release/v1.26.5/bin/linux/amd64/kubectl: attempt #1 Feb 9 08:58:38.738774 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Feb 9 08:58:39.106221 ignition[813]: DEBUG : files: createFilesystemsFiles: createFiles: op(7): file matches expected sum of: 97840854134909d75a1a2563628cc4ba632067369ce7fc8a8a1e90a387d32dd7bfd73f4f5b5a82ef842088e7470692951eb7fc869c5f297dd740f855672ee628 Feb 9 08:58:39.107440 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/bin/kubectl" Feb 9 08:58:39.107440 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/opt/bin/kubelet" Feb 9 08:58:39.107440 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET https://dl.k8s.io/release/v1.26.5/bin/linux/amd64/kubelet: attempt #1 Feb 9 08:58:39.153715 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET result: OK Feb 9 08:58:40.023112 ignition[813]: DEBUG : files: createFilesystemsFiles: createFiles: op(8): file matches expected sum of: 40daf2a9b9e666c14b10e627da931bd79978628b1f23ef6429c1cb4fcba261f86ccff440c0dbb0070ee760fe55772b4fd279c4582dfbb17fa30bc94b7f00126b Feb 9 08:58:40.023112 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/opt/bin/kubelet" Feb 9 08:58:40.025294 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/opt/bin/kubeadm" Feb 9 08:58:40.025294 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(9): GET https://dl.k8s.io/release/v1.26.5/bin/linux/amd64/kubeadm: attempt #1 Feb 9 08:58:40.071288 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(9): GET result: OK Feb 9 08:58:40.368070 ignition[813]: DEBUG : files: createFilesystemsFiles: createFiles: op(9): file matches expected sum of: 1c324cd645a7bf93d19d24c87498d9a17878eb1cc927e2680200ffeab2f85051ddec47d85b79b8e774042dc6726299ad3d7caf52c060701f00deba30dc33f660 Feb 9 08:58:40.368070 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/opt/bin/kubeadm" Feb 9 08:58:40.368070 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Feb 9 08:58:40.370583 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-amd64.tar.gz: attempt #1 Feb 9 08:58:40.776988 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Feb 9 08:58:40.856621 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/home/core/install.sh" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/home/core/install.sh" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/home/core/nginx.yaml" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/home/core/nginx.yaml" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Feb 9 08:58:40.857771 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/etc/flatcar/update.conf" Feb 9 08:58:40.865621 ignition[813]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/etc/flatcar/update.conf" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(10): [started] processing unit "coreos-metadata-sshkeys@.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(10): [finished] processing unit "coreos-metadata-sshkeys@.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(11): [started] processing unit "prepare-cni-plugins.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(11): op(12): [started] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(11): op(12): [finished] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(11): [finished] processing unit "prepare-cni-plugins.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(13): [started] processing unit "prepare-critools.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(13): op(14): [started] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(13): op(14): [finished] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(13): [finished] processing unit "prepare-critools.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(15): [started] processing unit "prepare-helm.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(15): op(16): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(15): op(16): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(15): [finished] processing unit "prepare-helm.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(17): [started] setting preset to enabled for "prepare-cni-plugins.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(17): [finished] setting preset to enabled for "prepare-cni-plugins.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(18): [started] setting preset to enabled for "prepare-critools.service" Feb 9 08:58:40.865621 ignition[813]: INFO : files: op(18): [finished] setting preset to enabled for "prepare-critools.service" Feb 9 08:58:40.908501 kernel: kauditd_printk_skb: 27 callbacks suppressed Feb 9 08:58:40.908547 kernel: audit: type=1130 audit(1707469120.871:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.908569 kernel: audit: type=1130 audit(1707469120.883:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.908588 kernel: audit: type=1131 audit(1707469120.883:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.908607 kernel: audit: type=1130 audit(1707469120.899:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.871000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.883000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.883000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.899000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.908862 ignition[813]: INFO : files: op(19): [started] setting preset to enabled for "prepare-helm.service" Feb 9 08:58:40.908862 ignition[813]: INFO : files: op(19): [finished] setting preset to enabled for "prepare-helm.service" Feb 9 08:58:40.908862 ignition[813]: INFO : files: op(1a): [started] setting preset to enabled for "coreos-metadata-sshkeys@.service " Feb 9 08:58:40.908862 ignition[813]: INFO : files: op(1a): [finished] setting preset to enabled for "coreos-metadata-sshkeys@.service " Feb 9 08:58:40.908862 ignition[813]: INFO : files: createResultFile: createFiles: op(1b): [started] writing file "/sysroot/etc/.ignition-result.json" Feb 9 08:58:40.908862 ignition[813]: INFO : files: createResultFile: createFiles: op(1b): [finished] writing file "/sysroot/etc/.ignition-result.json" Feb 9 08:58:40.908862 ignition[813]: INFO : files: files passed Feb 9 08:58:40.908862 ignition[813]: INFO : Ignition finished successfully Feb 9 08:58:40.869829 systemd[1]: Finished ignition-files.service. Feb 9 08:58:40.873356 systemd[1]: Starting initrd-setup-root-after-ignition.service... Feb 9 08:58:40.916427 initrd-setup-root-after-ignition[838]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Feb 9 08:58:40.876367 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Feb 9 08:58:40.877556 systemd[1]: Starting ignition-quench.service... Feb 9 08:58:40.883315 systemd[1]: ignition-quench.service: Deactivated successfully. Feb 9 08:58:40.883464 systemd[1]: Finished ignition-quench.service. Feb 9 08:58:40.898194 systemd[1]: Finished initrd-setup-root-after-ignition.service. Feb 9 08:58:40.900163 systemd[1]: Reached target ignition-complete.target. Feb 9 08:58:40.905329 systemd[1]: Starting initrd-parse-etc.service... Feb 9 08:58:40.928465 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Feb 9 08:58:40.937144 kernel: audit: type=1130 audit(1707469120.928:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.937189 kernel: audit: type=1131 audit(1707469120.928:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.928000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.928000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.928626 systemd[1]: Finished initrd-parse-etc.service. Feb 9 08:58:40.929417 systemd[1]: Reached target initrd-fs.target. Feb 9 08:58:40.937575 systemd[1]: Reached target initrd.target. Feb 9 08:58:40.938586 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Feb 9 08:58:40.940200 systemd[1]: Starting dracut-pre-pivot.service... Feb 9 08:58:40.960161 systemd[1]: Finished dracut-pre-pivot.service. Feb 9 08:58:40.960000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.962329 systemd[1]: Starting initrd-cleanup.service... Feb 9 08:58:40.966934 kernel: audit: type=1130 audit(1707469120.960:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.978747 systemd[1]: Stopped target nss-lookup.target. Feb 9 08:58:40.980171 systemd[1]: Stopped target remote-cryptsetup.target. Feb 9 08:58:40.981521 systemd[1]: Stopped target timers.target. Feb 9 08:58:40.982152 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Feb 9 08:58:40.987138 kernel: audit: type=1131 audit(1707469120.982:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.982000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.982346 systemd[1]: Stopped dracut-pre-pivot.service. Feb 9 08:58:40.983328 systemd[1]: Stopped target initrd.target. Feb 9 08:58:40.987529 systemd[1]: Stopped target basic.target. Feb 9 08:58:40.988293 systemd[1]: Stopped target ignition-complete.target. Feb 9 08:58:40.989123 systemd[1]: Stopped target ignition-diskful.target. Feb 9 08:58:40.990000 systemd[1]: Stopped target initrd-root-device.target. Feb 9 08:58:40.990907 systemd[1]: Stopped target remote-fs.target. Feb 9 08:58:40.991769 systemd[1]: Stopped target remote-fs-pre.target. Feb 9 08:58:40.992555 systemd[1]: Stopped target sysinit.target. Feb 9 08:58:40.993429 systemd[1]: Stopped target local-fs.target. Feb 9 08:58:40.994123 systemd[1]: Stopped target local-fs-pre.target. Feb 9 08:58:40.994971 systemd[1]: Stopped target swap.target. Feb 9 08:58:41.000335 kernel: audit: type=1131 audit(1707469120.996:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.996000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.995803 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Feb 9 08:58:40.995933 systemd[1]: Stopped dracut-pre-mount.service. Feb 9 08:58:41.005254 kernel: audit: type=1131 audit(1707469121.000:47): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.000000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:40.996895 systemd[1]: Stopped target cryptsetup.target. Feb 9 08:58:41.004000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.000670 systemd[1]: dracut-initqueue.service: Deactivated successfully. Feb 9 08:58:41.005000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.000839 systemd[1]: Stopped dracut-initqueue.service. Feb 9 08:58:41.006000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.001664 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Feb 9 08:58:41.001849 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Feb 9 08:58:41.015000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.005791 systemd[1]: ignition-files.service: Deactivated successfully. Feb 9 08:58:41.005904 systemd[1]: Stopped ignition-files.service. Feb 9 08:58:41.018000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.006486 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Feb 9 08:58:41.006636 systemd[1]: Stopped flatcar-metadata-hostname.service. Feb 9 08:58:41.008597 systemd[1]: Stopping ignition-mount.service... Feb 9 08:58:41.026068 ignition[851]: INFO : Ignition 2.14.0 Feb 9 08:58:41.026068 ignition[851]: INFO : Stage: umount Feb 9 08:58:41.026068 ignition[851]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Feb 9 08:58:41.026068 ignition[851]: DEBUG : parsing config with SHA512: 865c03baa79b8c74023d13a0b3666474fa06a165421a1e05731b76e0f557d42c5c89d4870a0b9c4182ad7d4d8209de20dca9c9da63d637e0410fbd60314cac6c Feb 9 08:58:41.031000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.011488 systemd[1]: Stopping iscsiuio.service... Feb 9 08:58:41.035266 ignition[851]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/digitalocean" Feb 9 08:58:41.014252 systemd[1]: Stopping sysroot-boot.service... Feb 9 08:58:41.035000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.035000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.038330 ignition[851]: INFO : umount: umount passed Feb 9 08:58:41.038330 ignition[851]: INFO : Ignition finished successfully Feb 9 08:58:41.014781 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Feb 9 08:58:41.040000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.041000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.042000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.015099 systemd[1]: Stopped systemd-udev-trigger.service. Feb 9 08:58:41.043000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.015959 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Feb 9 08:58:41.016733 systemd[1]: Stopped dracut-pre-trigger.service. Feb 9 08:58:41.045000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.029641 systemd[1]: iscsiuio.service: Deactivated successfully. Feb 9 08:58:41.030686 systemd[1]: Stopped iscsiuio.service. Feb 9 08:58:41.034134 systemd[1]: initrd-cleanup.service: Deactivated successfully. Feb 9 08:58:41.034859 systemd[1]: Finished initrd-cleanup.service. Feb 9 08:58:41.039242 systemd[1]: ignition-mount.service: Deactivated successfully. Feb 9 08:58:41.039377 systemd[1]: Stopped ignition-mount.service. Feb 9 08:58:41.041349 systemd[1]: ignition-disks.service: Deactivated successfully. Feb 9 08:58:41.041418 systemd[1]: Stopped ignition-disks.service. Feb 9 08:58:41.041935 systemd[1]: ignition-kargs.service: Deactivated successfully. Feb 9 08:58:41.041987 systemd[1]: Stopped ignition-kargs.service. Feb 9 08:58:41.043160 systemd[1]: ignition-fetch.service: Deactivated successfully. Feb 9 08:58:41.043219 systemd[1]: Stopped ignition-fetch.service. Feb 9 08:58:41.044274 systemd[1]: Stopped target network.target. Feb 9 08:58:41.064000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.045405 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Feb 9 08:58:41.045473 systemd[1]: Stopped ignition-fetch-offline.service. Feb 9 08:58:41.046589 systemd[1]: Stopped target paths.target. Feb 9 08:58:41.047675 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Feb 9 08:58:41.050795 systemd[1]: Stopped systemd-ask-password-console.path. Feb 9 08:58:41.051530 systemd[1]: Stopped target slices.target. Feb 9 08:58:41.052264 systemd[1]: Stopped target sockets.target. Feb 9 08:58:41.053056 systemd[1]: iscsid.socket: Deactivated successfully. Feb 9 08:58:41.053102 systemd[1]: Closed iscsid.socket. Feb 9 08:58:41.063524 systemd[1]: iscsiuio.socket: Deactivated successfully. Feb 9 08:58:41.063590 systemd[1]: Closed iscsiuio.socket. Feb 9 08:58:41.072000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.064256 systemd[1]: ignition-setup.service: Deactivated successfully. Feb 9 08:58:41.074000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.064329 systemd[1]: Stopped ignition-setup.service. Feb 9 08:58:41.065535 systemd[1]: Stopping systemd-networkd.service... Feb 9 08:58:41.076000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.076000 audit: BPF prog-id=6 op=UNLOAD Feb 9 08:58:41.066579 systemd[1]: Stopping systemd-resolved.service... Feb 9 08:58:41.067807 systemd-networkd[688]: eth1: DHCPv6 lease lost Feb 9 08:58:41.078000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.068027 systemd-networkd[688]: eth0: DHCPv6 lease lost Feb 9 08:58:41.078000 audit: BPF prog-id=9 op=UNLOAD Feb 9 08:58:41.070425 systemd[1]: sysroot-boot.mount: Deactivated successfully. Feb 9 08:58:41.071223 systemd[1]: systemd-networkd.service: Deactivated successfully. Feb 9 08:58:41.081000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.082000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.071360 systemd[1]: Stopped systemd-networkd.service. Feb 9 08:58:41.083000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.074057 systemd[1]: systemd-resolved.service: Deactivated successfully. Feb 9 08:58:41.074183 systemd[1]: Stopped systemd-resolved.service. Feb 9 08:58:41.076113 systemd[1]: sysroot-boot.service: Deactivated successfully. Feb 9 08:58:41.076241 systemd[1]: Stopped sysroot-boot.service. Feb 9 08:58:41.077335 systemd[1]: systemd-networkd.socket: Deactivated successfully. Feb 9 08:58:41.077384 systemd[1]: Closed systemd-networkd.socket. Feb 9 08:58:41.078033 systemd[1]: initrd-setup-root.service: Deactivated successfully. Feb 9 08:58:41.078092 systemd[1]: Stopped initrd-setup-root.service. Feb 9 08:58:41.080240 systemd[1]: Stopping network-cleanup.service... Feb 9 08:58:41.081838 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Feb 9 08:58:41.096000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.081916 systemd[1]: Stopped parse-ip-for-networkd.service. Feb 9 08:58:41.098000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.082461 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 9 08:58:41.082515 systemd[1]: Stopped systemd-sysctl.service. Feb 9 08:58:41.083407 systemd[1]: systemd-modules-load.service: Deactivated successfully. Feb 9 08:58:41.101000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.083464 systemd[1]: Stopped systemd-modules-load.service. Feb 9 08:58:41.102000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.084458 systemd[1]: Stopping systemd-udevd.service... Feb 9 08:58:41.102000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.092119 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Feb 9 08:58:41.096763 systemd[1]: systemd-udevd.service: Deactivated successfully. Feb 9 08:58:41.097004 systemd[1]: Stopped systemd-udevd.service. Feb 9 08:58:41.098254 systemd[1]: network-cleanup.service: Deactivated successfully. Feb 9 08:58:41.098381 systemd[1]: Stopped network-cleanup.service. Feb 9 08:58:41.099274 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Feb 9 08:58:41.099328 systemd[1]: Closed systemd-udevd-control.socket. Feb 9 08:58:41.100035 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Feb 9 08:58:41.100098 systemd[1]: Closed systemd-udevd-kernel.socket. Feb 9 08:58:41.100906 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Feb 9 08:58:41.100967 systemd[1]: Stopped dracut-pre-udev.service. Feb 9 08:58:41.101874 systemd[1]: dracut-cmdline.service: Deactivated successfully. Feb 9 08:58:41.115000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.101946 systemd[1]: Stopped dracut-cmdline.service. Feb 9 08:58:41.116000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.102819 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Feb 9 08:58:41.117000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.102870 systemd[1]: Stopped dracut-cmdline-ask.service. Feb 9 08:58:41.104928 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Feb 9 08:58:41.115224 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Feb 9 08:58:41.115332 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Feb 9 08:58:41.116490 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Feb 9 08:58:41.116579 systemd[1]: Stopped kmod-static-nodes.service. Feb 9 08:58:41.121000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.121000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:41.117136 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Feb 9 08:58:41.117189 systemd[1]: Stopped systemd-vconsole-setup.service. Feb 9 08:58:41.119461 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Feb 9 08:58:41.120242 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Feb 9 08:58:41.120376 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Feb 9 08:58:41.122267 systemd[1]: Reached target initrd-switch-root.target. Feb 9 08:58:41.124268 systemd[1]: Starting initrd-switch-root.service... Feb 9 08:58:41.140756 systemd[1]: Switching root. Feb 9 08:58:41.165353 iscsid[698]: iscsid shutting down. Feb 9 08:58:41.166023 systemd-journald[184]: Received SIGTERM from PID 1 (n/a). Feb 9 08:58:41.166106 systemd-journald[184]: Journal stopped Feb 9 08:58:45.139256 kernel: SELinux: Class mctp_socket not defined in policy. Feb 9 08:58:45.139356 kernel: SELinux: Class anon_inode not defined in policy. Feb 9 08:58:45.139382 kernel: SELinux: the above unknown classes and permissions will be allowed Feb 9 08:58:45.139400 kernel: SELinux: policy capability network_peer_controls=1 Feb 9 08:58:45.139416 kernel: SELinux: policy capability open_perms=1 Feb 9 08:58:45.139429 kernel: SELinux: policy capability extended_socket_class=1 Feb 9 08:58:45.139445 kernel: SELinux: policy capability always_check_network=0 Feb 9 08:58:45.139462 kernel: SELinux: policy capability cgroup_seclabel=1 Feb 9 08:58:45.139474 kernel: SELinux: policy capability nnp_nosuid_transition=1 Feb 9 08:58:45.139487 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Feb 9 08:58:45.139499 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Feb 9 08:58:45.139521 systemd[1]: Successfully loaded SELinux policy in 57.525ms. Feb 9 08:58:45.139545 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.764ms. Feb 9 08:58:45.139561 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Feb 9 08:58:45.139577 systemd[1]: Detected virtualization kvm. Feb 9 08:58:45.139590 systemd[1]: Detected architecture x86-64. Feb 9 08:58:45.139604 systemd[1]: Detected first boot. Feb 9 08:58:45.139617 systemd[1]: Hostname set to . Feb 9 08:58:45.139630 systemd[1]: Initializing machine ID from VM UUID. Feb 9 08:58:45.139643 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Feb 9 08:58:45.139657 systemd[1]: Populated /etc with preset unit settings. Feb 9 08:58:45.139676 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 08:58:45.139693 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 08:58:45.139707 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 08:58:45.139740 systemd[1]: iscsid.service: Deactivated successfully. Feb 9 08:58:45.139753 systemd[1]: Stopped iscsid.service. Feb 9 08:58:45.139767 systemd[1]: initrd-switch-root.service: Deactivated successfully. Feb 9 08:58:45.139780 systemd[1]: Stopped initrd-switch-root.service. Feb 9 08:58:45.139793 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Feb 9 08:58:45.139807 systemd[1]: Created slice system-addon\x2dconfig.slice. Feb 9 08:58:45.139824 systemd[1]: Created slice system-addon\x2drun.slice. Feb 9 08:58:45.139837 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice. Feb 9 08:58:45.139851 systemd[1]: Created slice system-getty.slice. Feb 9 08:58:45.139865 systemd[1]: Created slice system-modprobe.slice. Feb 9 08:58:45.139879 systemd[1]: Created slice system-serial\x2dgetty.slice. Feb 9 08:58:45.139893 systemd[1]: Created slice system-system\x2dcloudinit.slice. Feb 9 08:58:45.139906 systemd[1]: Created slice system-systemd\x2dfsck.slice. Feb 9 08:58:45.139921 systemd[1]: Created slice user.slice. Feb 9 08:58:45.139934 systemd[1]: Started systemd-ask-password-console.path. Feb 9 08:58:45.139950 systemd[1]: Started systemd-ask-password-wall.path. Feb 9 08:58:45.139963 systemd[1]: Set up automount boot.automount. Feb 9 08:58:45.139977 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Feb 9 08:58:45.139990 systemd[1]: Stopped target initrd-switch-root.target. Feb 9 08:58:45.140004 systemd[1]: Stopped target initrd-fs.target. Feb 9 08:58:45.140016 systemd[1]: Stopped target initrd-root-fs.target. Feb 9 08:58:45.140032 systemd[1]: Reached target integritysetup.target. Feb 9 08:58:45.140046 systemd[1]: Reached target remote-cryptsetup.target. Feb 9 08:58:45.140059 systemd[1]: Reached target remote-fs.target. Feb 9 08:58:45.140073 systemd[1]: Reached target slices.target. Feb 9 08:58:45.140086 systemd[1]: Reached target swap.target. Feb 9 08:58:45.140100 systemd[1]: Reached target torcx.target. Feb 9 08:58:45.140113 systemd[1]: Reached target veritysetup.target. Feb 9 08:58:45.140128 systemd[1]: Listening on systemd-coredump.socket. Feb 9 08:58:45.140141 systemd[1]: Listening on systemd-initctl.socket. Feb 9 08:58:45.140161 systemd[1]: Listening on systemd-networkd.socket. Feb 9 08:58:45.140185 systemd[1]: Listening on systemd-udevd-control.socket. Feb 9 08:58:45.140204 systemd[1]: Listening on systemd-udevd-kernel.socket. Feb 9 08:58:45.140217 systemd[1]: Listening on systemd-userdbd.socket. Feb 9 08:58:45.140231 systemd[1]: Mounting dev-hugepages.mount... Feb 9 08:58:45.140244 systemd[1]: Mounting dev-mqueue.mount... Feb 9 08:58:45.140256 systemd[1]: Mounting media.mount... Feb 9 08:58:45.140269 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 08:58:45.140288 systemd[1]: Mounting sys-kernel-debug.mount... Feb 9 08:58:45.140307 systemd[1]: Mounting sys-kernel-tracing.mount... Feb 9 08:58:45.140328 systemd[1]: Mounting tmp.mount... Feb 9 08:58:45.140346 systemd[1]: Starting flatcar-tmpfiles.service... Feb 9 08:58:45.140364 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Feb 9 08:58:45.140384 systemd[1]: Starting kmod-static-nodes.service... Feb 9 08:58:45.140405 systemd[1]: Starting modprobe@configfs.service... Feb 9 08:58:45.140423 systemd[1]: Starting modprobe@dm_mod.service... Feb 9 08:58:45.140444 systemd[1]: Starting modprobe@drm.service... Feb 9 08:58:45.140457 systemd[1]: Starting modprobe@efi_pstore.service... Feb 9 08:58:45.140470 systemd[1]: Starting modprobe@fuse.service... Feb 9 08:58:45.140487 systemd[1]: Starting modprobe@loop.service... Feb 9 08:58:45.140500 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 9 08:58:45.140514 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Feb 9 08:58:45.140527 systemd[1]: Stopped systemd-fsck-root.service. Feb 9 08:58:45.140540 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Feb 9 08:58:45.140553 systemd[1]: Stopped systemd-fsck-usr.service. Feb 9 08:58:45.140565 systemd[1]: Stopped systemd-journald.service. Feb 9 08:58:45.140579 systemd[1]: Starting systemd-journald.service... Feb 9 08:58:45.140592 systemd[1]: Starting systemd-modules-load.service... Feb 9 08:58:45.140608 systemd[1]: Starting systemd-network-generator.service... Feb 9 08:58:45.140621 systemd[1]: Starting systemd-remount-fs.service... Feb 9 08:58:45.140635 systemd[1]: Starting systemd-udev-trigger.service... Feb 9 08:58:45.140648 systemd[1]: verity-setup.service: Deactivated successfully. Feb 9 08:58:45.140662 systemd[1]: Stopped verity-setup.service. Feb 9 08:58:45.140681 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 08:58:45.140700 systemd[1]: Mounted dev-hugepages.mount. Feb 9 08:58:45.140713 systemd[1]: Mounted dev-mqueue.mount. Feb 9 08:58:45.140740 systemd[1]: Mounted media.mount. Feb 9 08:58:45.140780 systemd[1]: Mounted sys-kernel-debug.mount. Feb 9 08:58:45.140800 systemd[1]: Mounted sys-kernel-tracing.mount. Feb 9 08:58:45.140822 systemd[1]: Mounted tmp.mount. Feb 9 08:58:45.140841 systemd[1]: Finished kmod-static-nodes.service. Feb 9 08:58:45.140862 systemd[1]: modprobe@configfs.service: Deactivated successfully. Feb 9 08:58:45.140878 systemd[1]: Finished modprobe@configfs.service. Feb 9 08:58:45.140891 kernel: loop: module loaded Feb 9 08:58:45.140904 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 9 08:58:45.140917 systemd[1]: Finished modprobe@dm_mod.service. Feb 9 08:58:45.140930 systemd[1]: modprobe@drm.service: Deactivated successfully. Feb 9 08:58:45.140943 kernel: fuse: init (API version 7.34) Feb 9 08:58:45.140959 systemd[1]: Finished modprobe@drm.service. Feb 9 08:58:45.140973 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 9 08:58:45.140986 systemd[1]: Finished modprobe@efi_pstore.service. Feb 9 08:58:45.140999 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 9 08:58:45.141013 systemd[1]: Finished modprobe@loop.service. Feb 9 08:58:45.141026 systemd[1]: Finished systemd-modules-load.service. Feb 9 08:58:45.141040 systemd[1]: modprobe@fuse.service: Deactivated successfully. Feb 9 08:58:45.141053 systemd[1]: Finished modprobe@fuse.service. Feb 9 08:58:45.141069 systemd[1]: Finished systemd-network-generator.service. Feb 9 08:58:45.141092 systemd[1]: Finished systemd-remount-fs.service. Feb 9 08:58:45.141126 systemd[1]: Reached target network-pre.target. Feb 9 08:58:45.141146 systemd[1]: Mounting sys-fs-fuse-connections.mount... Feb 9 08:58:45.141162 systemd[1]: Mounting sys-kernel-config.mount... Feb 9 08:58:45.141175 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 9 08:58:45.141189 systemd[1]: Starting systemd-hwdb-update.service... Feb 9 08:58:45.141203 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 9 08:58:45.141221 systemd[1]: Starting systemd-random-seed.service... Feb 9 08:58:45.141245 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Feb 9 08:58:45.141267 systemd[1]: Starting systemd-sysctl.service... Feb 9 08:58:45.141285 systemd[1]: Mounted sys-fs-fuse-connections.mount. Feb 9 08:58:45.141304 systemd[1]: Mounted sys-kernel-config.mount. Feb 9 08:58:45.141336 systemd[1]: Finished systemd-random-seed.service. Feb 9 08:58:45.141357 systemd[1]: Reached target first-boot-complete.target. Feb 9 08:58:45.141377 systemd[1]: Finished systemd-sysctl.service. Feb 9 08:58:45.141408 systemd-journald[953]: Journal started Feb 9 08:58:45.141502 systemd-journald[953]: Runtime Journal (/run/log/journal/0f2d3f05f4034df2965e592d36475234) is 4.9M, max 39.5M, 34.5M free. Feb 9 08:58:41.338000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Feb 9 08:58:45.149769 systemd[1]: Started systemd-journald.service. Feb 9 08:58:41.403000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Feb 9 08:58:41.403000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Feb 9 08:58:41.403000 audit: BPF prog-id=10 op=LOAD Feb 9 08:58:41.403000 audit: BPF prog-id=10 op=UNLOAD Feb 9 08:58:41.403000 audit: BPF prog-id=11 op=LOAD Feb 9 08:58:41.403000 audit: BPF prog-id=11 op=UNLOAD Feb 9 08:58:41.515000 audit[883]: AVC avc: denied { associate } for pid=883 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Feb 9 08:58:41.515000 audit[883]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c00014d8a2 a1=c0000cedf8 a2=c0000d70c0 a3=32 items=0 ppid=866 pid=883 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 08:58:41.515000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Feb 9 08:58:41.516000 audit[883]: AVC avc: denied { associate } for pid=883 comm="torcx-generator" name="bin" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Feb 9 08:58:41.516000 audit[883]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c00014d979 a2=1ed a3=0 items=2 ppid=866 pid=883 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 08:58:41.516000 audit: CWD cwd="/" Feb 9 08:58:41.516000 audit: PATH item=0 name=(null) inode=2 dev=00:1a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:41.516000 audit: PATH item=1 name=(null) inode=3 dev=00:1a mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:41.516000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Feb 9 08:58:44.850000 audit: BPF prog-id=12 op=LOAD Feb 9 08:58:44.850000 audit: BPF prog-id=3 op=UNLOAD Feb 9 08:58:44.850000 audit: BPF prog-id=13 op=LOAD Feb 9 08:58:44.850000 audit: BPF prog-id=14 op=LOAD Feb 9 08:58:44.850000 audit: BPF prog-id=4 op=UNLOAD Feb 9 08:58:44.850000 audit: BPF prog-id=5 op=UNLOAD Feb 9 08:58:44.852000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.855000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.858000 audit: BPF prog-id=12 op=UNLOAD Feb 9 08:58:44.859000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.859000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.991000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.993000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.994000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.994000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.995000 audit: BPF prog-id=15 op=LOAD Feb 9 08:58:44.996000 audit: BPF prog-id=16 op=LOAD Feb 9 08:58:44.996000 audit: BPF prog-id=17 op=LOAD Feb 9 08:58:44.996000 audit: BPF prog-id=13 op=UNLOAD Feb 9 08:58:44.996000 audit: BPF prog-id=14 op=UNLOAD Feb 9 08:58:45.020000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.037000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.042000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.042000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.045000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.045000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.050000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.050000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.054000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.054000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.057000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.057000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.159082 systemd-journald[953]: Time spent on flushing to /var/log/journal/0f2d3f05f4034df2965e592d36475234 is 32.017ms for 1144 entries. Feb 9 08:58:45.159082 systemd-journald[953]: System Journal (/var/log/journal/0f2d3f05f4034df2965e592d36475234) is 8.0M, max 195.6M, 187.6M free. Feb 9 08:58:45.204671 systemd-journald[953]: Received client request to flush runtime journal. Feb 9 08:58:45.060000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.063000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.063000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.065000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.067000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.112000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.133000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Feb 9 08:58:45.135000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.133000 audit[953]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=5 a1=7ffd876ae1b0 a2=4000 a3=7ffd876ae24c items=0 ppid=1 pid=953 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 08:58:45.133000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Feb 9 08:58:45.142000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.160000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.205000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:44.847438 systemd[1]: Queued start job for default target multi-user.target. Feb 9 08:58:41.513424 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 08:58:44.847466 systemd[1]: Unnecessary job was removed for dev-vda6.device. Feb 9 08:58:41.513829 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Feb 9 08:58:44.852817 systemd[1]: systemd-journald.service: Deactivated successfully. Feb 9 08:58:41.513850 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Feb 9 08:58:45.145474 systemd[1]: Starting systemd-journal-flush.service... Feb 9 08:58:41.513884 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Feb 9 08:58:45.160801 systemd[1]: Finished flatcar-tmpfiles.service. Feb 9 08:58:41.513898 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="skipped missing lower profile" missing profile=oem Feb 9 08:58:45.163295 systemd[1]: Starting systemd-sysusers.service... Feb 9 08:58:41.513946 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Feb 9 08:58:45.206122 systemd[1]: Finished systemd-journal-flush.service. Feb 9 08:58:41.513966 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Feb 9 08:58:41.514254 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Feb 9 08:58:41.514307 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Feb 9 08:58:41.514325 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Feb 9 08:58:41.514929 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Feb 9 08:58:41.514979 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Feb 9 08:58:41.515000 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.2: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.2 Feb 9 08:58:41.515016 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Feb 9 08:58:41.515035 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.2: no such file or directory" path=/var/lib/torcx/store/3510.3.2 Feb 9 08:58:41.515050 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:41Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Feb 9 08:58:44.281588 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:44Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 08:58:44.282067 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:44Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 08:58:44.282233 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:44Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 08:58:44.282540 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:44Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Feb 9 08:58:44.282622 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:44Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Feb 9 08:58:44.282738 /usr/lib/systemd/system-generators/torcx-generator[883]: time="2024-02-09T08:58:44Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Feb 9 08:58:45.229000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.229519 systemd[1]: Finished systemd-sysusers.service. Feb 9 08:58:45.233230 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Feb 9 08:58:45.252000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.253006 systemd[1]: Finished systemd-udev-trigger.service. Feb 9 08:58:45.255378 systemd[1]: Starting systemd-udev-settle.service... Feb 9 08:58:45.276688 udevadm[995]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Feb 9 08:58:45.280133 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Feb 9 08:58:45.279000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.906495 systemd[1]: Finished systemd-hwdb-update.service. Feb 9 08:58:45.917206 kernel: kauditd_printk_skb: 97 callbacks suppressed Feb 9 08:58:45.917356 kernel: audit: type=1130 audit(1707469125.906:136): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.906000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.919769 kernel: audit: type=1334 audit(1707469125.916:137): prog-id=18 op=LOAD Feb 9 08:58:45.916000 audit: BPF prog-id=18 op=LOAD Feb 9 08:58:45.918528 systemd[1]: Starting systemd-udevd.service... Feb 9 08:58:45.916000 audit: BPF prog-id=19 op=LOAD Feb 9 08:58:45.916000 audit: BPF prog-id=7 op=UNLOAD Feb 9 08:58:45.922836 kernel: audit: type=1334 audit(1707469125.916:138): prog-id=19 op=LOAD Feb 9 08:58:45.922899 kernel: audit: type=1334 audit(1707469125.916:139): prog-id=7 op=UNLOAD Feb 9 08:58:45.922928 kernel: audit: type=1334 audit(1707469125.916:140): prog-id=8 op=UNLOAD Feb 9 08:58:45.916000 audit: BPF prog-id=8 op=UNLOAD Feb 9 08:58:45.944401 systemd-udevd[996]: Using default interface naming scheme 'v252'. Feb 9 08:58:45.978662 systemd[1]: Started systemd-udevd.service. Feb 9 08:58:45.978000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.981418 systemd[1]: Starting systemd-networkd.service... Feb 9 08:58:45.985799 kernel: audit: type=1130 audit(1707469125.978:141): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:45.985921 kernel: audit: type=1334 audit(1707469125.979:142): prog-id=20 op=LOAD Feb 9 08:58:45.979000 audit: BPF prog-id=20 op=LOAD Feb 9 08:58:45.993649 kernel: audit: type=1334 audit(1707469125.989:143): prog-id=21 op=LOAD Feb 9 08:58:45.993765 kernel: audit: type=1334 audit(1707469125.990:144): prog-id=22 op=LOAD Feb 9 08:58:45.993789 kernel: audit: type=1334 audit(1707469125.991:145): prog-id=23 op=LOAD Feb 9 08:58:45.989000 audit: BPF prog-id=21 op=LOAD Feb 9 08:58:45.990000 audit: BPF prog-id=22 op=LOAD Feb 9 08:58:45.991000 audit: BPF prog-id=23 op=LOAD Feb 9 08:58:45.993163 systemd[1]: Starting systemd-userdbd.service... Feb 9 08:58:46.040000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.040529 systemd[1]: Started systemd-userdbd.service. Feb 9 08:58:46.069737 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Feb 9 08:58:46.098747 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 08:58:46.098971 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Feb 9 08:58:46.100845 systemd[1]: Starting modprobe@dm_mod.service... Feb 9 08:58:46.104310 systemd[1]: Starting modprobe@efi_pstore.service... Feb 9 08:58:46.110176 systemd[1]: Starting modprobe@loop.service... Feb 9 08:58:46.110637 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Feb 9 08:58:46.110763 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Feb 9 08:58:46.110870 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Feb 9 08:58:46.111595 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Feb 9 08:58:46.111814 systemd[1]: Finished modprobe@dm_mod.service. Feb 9 08:58:46.111000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.111000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.113170 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Feb 9 08:58:46.113394 systemd[1]: Finished modprobe@efi_pstore.service. Feb 9 08:58:46.113000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.113000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.123929 systemd[1]: modprobe@loop.service: Deactivated successfully. Feb 9 08:58:46.124204 systemd[1]: Finished modprobe@loop.service. Feb 9 08:58:46.124000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.124000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.125091 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Feb 9 08:58:46.125174 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Feb 9 08:58:46.203779 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Feb 9 08:58:46.206552 systemd-networkd[997]: lo: Link UP Feb 9 08:58:46.206567 systemd-networkd[997]: lo: Gained carrier Feb 9 08:58:46.207443 systemd-networkd[997]: Enumeration completed Feb 9 08:58:46.207614 systemd[1]: Started systemd-networkd.service. Feb 9 08:58:46.207000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.208383 systemd-networkd[997]: eth0: Configuring with /run/systemd/network/10-4a:f7:25:cd:b1:63.network. Feb 9 08:58:46.210530 systemd-networkd[997]: eth0: Link UP Feb 9 08:58:46.210545 systemd-networkd[997]: eth0: Gained carrier Feb 9 08:58:46.210790 kernel: ACPI: button: Power Button [PWRF] Feb 9 08:58:46.216000 audit[1006]: AVC avc: denied { confidentiality } for pid=1006 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Feb 9 08:58:46.216000 audit[1006]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=560b9a6c9fc0 a1=32194 a2=7f2d72160bc5 a3=5 items=108 ppid=996 pid=1006 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 08:58:46.216000 audit: CWD cwd="/" Feb 9 08:58:46.216000 audit: PATH item=0 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=1 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=2 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=3 name=(null) inode=14462 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=4 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=5 name=(null) inode=14463 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=6 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=7 name=(null) inode=14464 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=8 name=(null) inode=14464 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=9 name=(null) inode=14465 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=10 name=(null) inode=14464 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=11 name=(null) inode=14466 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=12 name=(null) inode=14464 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=13 name=(null) inode=14467 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=14 name=(null) inode=14464 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=15 name=(null) inode=14468 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=16 name=(null) inode=14464 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=17 name=(null) inode=14469 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=18 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=19 name=(null) inode=14470 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=20 name=(null) inode=14470 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=21 name=(null) inode=14471 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=22 name=(null) inode=14470 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=23 name=(null) inode=14472 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=24 name=(null) inode=14470 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=25 name=(null) inode=14473 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=26 name=(null) inode=14470 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=27 name=(null) inode=14474 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=28 name=(null) inode=14470 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=29 name=(null) inode=14475 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=30 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=31 name=(null) inode=14476 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=32 name=(null) inode=14476 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=33 name=(null) inode=14477 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=34 name=(null) inode=14476 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=35 name=(null) inode=14478 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=36 name=(null) inode=14476 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=37 name=(null) inode=14479 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=38 name=(null) inode=14476 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=39 name=(null) inode=14480 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=40 name=(null) inode=14476 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=41 name=(null) inode=14481 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=42 name=(null) inode=14461 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=43 name=(null) inode=14482 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=44 name=(null) inode=14482 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=45 name=(null) inode=14483 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=46 name=(null) inode=14482 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=47 name=(null) inode=14484 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=48 name=(null) inode=14482 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=49 name=(null) inode=14485 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=50 name=(null) inode=14482 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=51 name=(null) inode=14486 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=52 name=(null) inode=14482 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=53 name=(null) inode=14487 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=54 name=(null) inode=45 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=55 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=56 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=57 name=(null) inode=14489 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=58 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=59 name=(null) inode=14490 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=60 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=61 name=(null) inode=14491 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=62 name=(null) inode=14491 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=63 name=(null) inode=14492 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=64 name=(null) inode=14491 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=65 name=(null) inode=14493 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=66 name=(null) inode=14491 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=67 name=(null) inode=14494 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=68 name=(null) inode=14491 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=69 name=(null) inode=14495 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=70 name=(null) inode=14491 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=71 name=(null) inode=14496 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=72 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=73 name=(null) inode=14497 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=74 name=(null) inode=14497 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=75 name=(null) inode=14498 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=76 name=(null) inode=14497 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=77 name=(null) inode=14499 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=78 name=(null) inode=14497 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=79 name=(null) inode=14500 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=80 name=(null) inode=14497 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=81 name=(null) inode=14501 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=82 name=(null) inode=14497 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=83 name=(null) inode=14502 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=84 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=85 name=(null) inode=14503 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=86 name=(null) inode=14503 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=87 name=(null) inode=14504 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=88 name=(null) inode=14503 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=89 name=(null) inode=14505 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=90 name=(null) inode=14503 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=91 name=(null) inode=14506 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.245189 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0x700, revision 0 Feb 9 08:58:46.216000 audit: PATH item=92 name=(null) inode=14503 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=93 name=(null) inode=14507 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=94 name=(null) inode=14503 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=95 name=(null) inode=14508 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=96 name=(null) inode=14488 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=97 name=(null) inode=14509 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=98 name=(null) inode=14509 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=99 name=(null) inode=14510 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=100 name=(null) inode=14509 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=101 name=(null) inode=14511 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=102 name=(null) inode=14509 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=103 name=(null) inode=14512 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=104 name=(null) inode=14509 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=105 name=(null) inode=14513 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=106 name=(null) inode=14509 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PATH item=107 name=(null) inode=14514 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Feb 9 08:58:46.216000 audit: PROCTITLE proctitle="(udev-worker)" Feb 9 08:58:46.255773 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Feb 9 08:58:46.266743 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Feb 9 08:58:46.328751 kernel: mousedev: PS/2 mouse device common for all mice Feb 9 08:58:46.329315 systemd-networkd[997]: eth1: Configuring with /run/systemd/network/10-c6:24:ae:7c:3d:6e.network. Feb 9 08:58:46.330062 systemd-networkd[997]: eth1: Link UP Feb 9 08:58:46.330074 systemd-networkd[997]: eth1: Gained carrier Feb 9 08:58:46.445752 kernel: EDAC MC: Ver: 3.0.0 Feb 9 08:58:46.460864 systemd[1]: Finished systemd-udev-settle.service. Feb 9 08:58:46.460000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.463435 systemd[1]: Starting lvm2-activation-early.service... Feb 9 08:58:46.486462 lvm[1034]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 9 08:58:46.520286 systemd[1]: Finished lvm2-activation-early.service. Feb 9 08:58:46.520000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.520967 systemd[1]: Reached target cryptsetup.target. Feb 9 08:58:46.523289 systemd[1]: Starting lvm2-activation.service... Feb 9 08:58:46.533782 lvm[1035]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Feb 9 08:58:46.566271 systemd[1]: Finished lvm2-activation.service. Feb 9 08:58:46.566000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.567285 systemd[1]: Reached target local-fs-pre.target. Feb 9 08:58:46.570871 systemd[1]: Mounting media-configdrive.mount... Feb 9 08:58:46.571667 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Feb 9 08:58:46.571833 systemd[1]: Reached target machines.target. Feb 9 08:58:46.576559 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Feb 9 08:58:46.593000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.593261 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Feb 9 08:58:46.596733 kernel: ISO 9660 Extensions: RRIP_1991A Feb 9 08:58:46.598710 systemd[1]: Mounted media-configdrive.mount. Feb 9 08:58:46.599451 systemd[1]: Reached target local-fs.target. Feb 9 08:58:46.601785 systemd[1]: Starting ldconfig.service... Feb 9 08:58:46.603485 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Feb 9 08:58:46.603564 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Feb 9 08:58:46.606146 systemd[1]: Starting systemd-boot-update.service... Feb 9 08:58:46.609662 systemd[1]: Starting systemd-machine-id-commit.service... Feb 9 08:58:46.610899 systemd[1]: systemd-sysext.service was skipped because no trigger condition checks were met. Feb 9 08:58:46.611084 systemd[1]: ensure-sysext.service was skipped because no trigger condition checks were met. Feb 9 08:58:46.613585 systemd[1]: Starting systemd-tmpfiles-setup.service... Feb 9 08:58:46.640081 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1041 (bootctl) Feb 9 08:58:46.644071 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Feb 9 08:58:46.657702 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Feb 9 08:58:46.659261 systemd[1]: Finished systemd-machine-id-commit.service. Feb 9 08:58:46.659000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.680776 systemd-tmpfiles[1043]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Feb 9 08:58:46.685796 systemd-tmpfiles[1043]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Feb 9 08:58:46.695419 systemd-tmpfiles[1043]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Feb 9 08:58:46.749735 systemd-fsck[1047]: fsck.fat 4.2 (2021-01-31) Feb 9 08:58:46.749735 systemd-fsck[1047]: /dev/vda1: 789 files, 115332/258078 clusters Feb 9 08:58:46.754333 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Feb 9 08:58:46.754000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.756582 systemd[1]: Mounting boot.mount... Feb 9 08:58:46.768896 systemd[1]: Mounted boot.mount. Feb 9 08:58:46.806542 systemd[1]: Finished systemd-boot-update.service. Feb 9 08:58:46.806000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.955454 systemd[1]: Finished systemd-tmpfiles-setup.service. Feb 9 08:58:46.955000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.958128 systemd[1]: Starting audit-rules.service... Feb 9 08:58:46.961201 systemd[1]: Starting clean-ca-certificates.service... Feb 9 08:58:46.965316 systemd[1]: Starting systemd-journal-catalog-update.service... Feb 9 08:58:46.967000 audit: BPF prog-id=24 op=LOAD Feb 9 08:58:46.970521 systemd[1]: Starting systemd-resolved.service... Feb 9 08:58:46.976000 audit: BPF prog-id=25 op=LOAD Feb 9 08:58:46.979029 systemd[1]: Starting systemd-timesyncd.service... Feb 9 08:58:46.983470 systemd[1]: Starting systemd-update-utmp.service... Feb 9 08:58:46.998629 systemd[1]: Finished clean-ca-certificates.service. Feb 9 08:58:46.998000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:46.999358 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Feb 9 08:58:47.006000 audit[1055]: SYSTEM_BOOT pid=1055 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Feb 9 08:58:47.010289 systemd[1]: Finished systemd-update-utmp.service. Feb 9 08:58:47.010000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:47.075417 ldconfig[1040]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Feb 9 08:58:47.083604 systemd[1]: Finished ldconfig.service. Feb 9 08:58:47.083000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Feb 9 08:58:47.106000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Feb 9 08:58:47.106000 audit[1070]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffc40ac1710 a2=420 a3=0 items=0 ppid=1050 pid=1070 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Feb 9 08:58:47.106000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Feb 9 08:58:47.110011 augenrules[1070]: No rules Feb 9 08:58:47.110161 systemd[1]: Finished audit-rules.service. Feb 9 08:58:47.120076 systemd[1]: Finished systemd-journal-catalog-update.service. Feb 9 08:58:47.122152 systemd[1]: Starting systemd-update-done.service... Feb 9 08:58:47.136266 systemd[1]: Finished systemd-update-done.service. Feb 9 08:58:47.143832 systemd-resolved[1053]: Positive Trust Anchors: Feb 9 08:58:47.143851 systemd-resolved[1053]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Feb 9 08:58:47.143888 systemd-resolved[1053]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Feb 9 08:58:47.150947 systemd-resolved[1053]: Using system hostname 'ci-3510.3.2-5-6e53ffd492'. Feb 9 08:58:47.153382 systemd[1]: Started systemd-resolved.service. Feb 9 08:58:47.153986 systemd[1]: Reached target network.target. Feb 9 08:58:47.154323 systemd[1]: Reached target nss-lookup.target. Feb 9 08:58:47.182258 systemd[1]: Started systemd-timesyncd.service. Feb 9 08:58:47.182863 systemd[1]: Reached target sysinit.target. Feb 9 08:58:47.183381 systemd[1]: Started motdgen.path. Feb 9 08:58:47.183759 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Feb 9 08:58:47.184174 systemd[1]: Started systemd-tmpfiles-clean.timer. Feb 9 08:58:47.184491 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Feb 9 08:58:47.184528 systemd[1]: Reached target paths.target. Feb 9 08:58:47.184828 systemd[1]: Reached target time-set.target. Feb 9 08:58:47.185440 systemd[1]: Started logrotate.timer. Feb 9 08:58:47.185922 systemd[1]: Started mdadm.timer. Feb 9 08:58:47.186312 systemd[1]: Reached target timers.target. Feb 9 08:58:47.187052 systemd[1]: Listening on dbus.socket. Feb 9 08:58:47.189376 systemd[1]: Starting docker.socket... Feb 9 08:58:47.193697 systemd[1]: Listening on sshd.socket. Feb 9 08:58:47.194298 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Feb 9 08:58:47.194971 systemd[1]: Listening on docker.socket. Feb 9 08:58:47.195660 systemd[1]: Reached target sockets.target. Feb 9 08:58:47.196024 systemd[1]: Reached target basic.target. Feb 9 08:58:47.196595 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Feb 9 08:58:47.196625 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Feb 9 08:58:47.198178 systemd[1]: Starting containerd.service... Feb 9 08:58:47.199969 systemd[1]: Starting coreos-metadata-sshkeys@core.service... Feb 9 08:58:47.205015 systemd[1]: Starting dbus.service... Feb 9 08:58:47.207289 systemd[1]: Starting enable-oem-cloudinit.service... Feb 9 08:58:48.095588 systemd-timesyncd[1054]: Contacted time server 216.240.36.24:123 (0.flatcar.pool.ntp.org). Feb 9 08:58:48.095668 systemd-timesyncd[1054]: Initial clock synchronization to Fri 2024-02-09 08:58:48.094408 UTC. Feb 9 08:58:48.095989 systemd[1]: Starting extend-filesystems.service... Feb 9 08:58:48.096787 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Feb 9 08:58:48.100458 systemd[1]: Starting motdgen.service... Feb 9 08:58:48.102635 systemd[1]: Starting prepare-cni-plugins.service... Feb 9 08:58:48.107308 systemd[1]: Starting prepare-critools.service... Feb 9 08:58:48.108492 systemd-resolved[1053]: Clock change detected. Flushing caches. Feb 9 08:58:48.109702 systemd[1]: Starting prepare-helm.service... Feb 9 08:58:48.111647 jq[1084]: false Feb 9 08:58:48.113777 systemd[1]: Starting ssh-key-proc-cmdline.service... Feb 9 08:58:48.121559 systemd[1]: Starting sshd-keygen.service... Feb 9 08:58:48.125809 systemd[1]: Starting systemd-logind.service... Feb 9 08:58:48.126301 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Feb 9 08:58:48.126399 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Feb 9 08:58:48.129499 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Feb 9 08:58:48.130707 systemd[1]: Starting update-engine.service... Feb 9 08:58:48.133827 systemd[1]: Starting update-ssh-keys-after-ignition.service... Feb 9 08:58:48.142842 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Feb 9 08:58:48.143200 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Feb 9 08:58:48.147749 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Feb 9 08:58:48.147986 systemd[1]: Finished ssh-key-proc-cmdline.service. Feb 9 08:58:48.168288 tar[1099]: ./ Feb 9 08:58:48.168288 tar[1099]: ./macvlan Feb 9 08:58:48.170851 tar[1100]: linux-amd64/helm Feb 9 08:58:48.174848 jq[1097]: true Feb 9 08:58:48.196007 tar[1101]: crictl Feb 9 08:58:48.227077 dbus-daemon[1083]: [system] SELinux support is enabled Feb 9 08:58:48.227480 systemd[1]: Started dbus.service. Feb 9 08:58:48.230735 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Feb 9 08:58:48.230803 systemd[1]: Reached target system-config.target. Feb 9 08:58:48.231380 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Feb 9 08:58:48.233171 systemd[1]: Starting user-configdrive.service... Feb 9 08:58:48.253638 extend-filesystems[1085]: Found vda Feb 9 08:58:48.259755 extend-filesystems[1085]: Found vda1 Feb 9 08:58:48.260168 systemd[1]: motdgen.service: Deactivated successfully. Feb 9 08:58:48.260358 systemd[1]: Finished motdgen.service. Feb 9 08:58:48.262221 extend-filesystems[1085]: Found vda2 Feb 9 08:58:48.265018 extend-filesystems[1085]: Found vda3 Feb 9 08:58:48.265680 extend-filesystems[1085]: Found usr Feb 9 08:58:48.266262 extend-filesystems[1085]: Found vda4 Feb 9 08:58:48.266832 extend-filesystems[1085]: Found vda6 Feb 9 08:58:48.270190 extend-filesystems[1085]: Found vda7 Feb 9 08:58:48.271133 extend-filesystems[1085]: Found vda9 Feb 9 08:58:48.271133 extend-filesystems[1085]: Checking size of /dev/vda9 Feb 9 08:58:48.274414 jq[1108]: true Feb 9 08:58:48.351088 update_engine[1096]: I0209 08:58:48.349264 1096 main.cc:92] Flatcar Update Engine starting Feb 9 08:58:48.356208 coreos-cloudinit[1116]: 2024/02/09 08:58:48 Checking availability of "cloud-drive" Feb 9 08:58:48.356625 coreos-cloudinit[1116]: 2024/02/09 08:58:48 Fetching user-data from datasource of type "cloud-drive" Feb 9 08:58:48.356625 coreos-cloudinit[1116]: 2024/02/09 08:58:48 Attempting to read from "/media/configdrive/openstack/latest/user_data" Feb 9 08:58:48.356874 systemd[1]: Started update-engine.service. Feb 9 08:58:48.359810 systemd[1]: Started locksmithd.service. Feb 9 08:58:48.363220 extend-filesystems[1085]: Resized partition /dev/vda9 Feb 9 08:58:48.363981 update_engine[1096]: I0209 08:58:48.363283 1096 update_check_scheduler.cc:74] Next update check in 9m55s Feb 9 08:58:48.367411 coreos-cloudinit[1116]: 2024/02/09 08:58:48 Fetching meta-data from datasource of type "cloud-drive" Feb 9 08:58:48.367411 coreos-cloudinit[1116]: 2024/02/09 08:58:48 Attempting to read from "/media/configdrive/openstack/latest/meta_data.json" Feb 9 08:58:48.371900 extend-filesystems[1132]: resize2fs 1.46.5 (30-Dec-2021) Feb 9 08:58:48.380070 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 15121403 blocks Feb 9 08:58:48.381813 coreos-cloudinit[1116]: Detected an Ignition config. Exiting... Feb 9 08:58:48.424088 systemd[1]: Finished user-configdrive.service. Feb 9 08:58:48.426280 systemd[1]: Reached target user-config.target. Feb 9 08:58:48.455091 kernel: EXT4-fs (vda9): resized filesystem to 15121403 Feb 9 08:58:48.471310 extend-filesystems[1132]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Feb 9 08:58:48.471310 extend-filesystems[1132]: old_desc_blocks = 1, new_desc_blocks = 8 Feb 9 08:58:48.471310 extend-filesystems[1132]: The filesystem on /dev/vda9 is now 15121403 (4k) blocks long. Feb 9 08:58:48.474276 extend-filesystems[1085]: Resized filesystem in /dev/vda9 Feb 9 08:58:48.474276 extend-filesystems[1085]: Found vdb Feb 9 08:58:48.471871 systemd[1]: extend-filesystems.service: Deactivated successfully. Feb 9 08:58:48.472061 systemd[1]: Finished extend-filesystems.service. Feb 9 08:58:48.487301 systemd-networkd[997]: eth1: Gained IPv6LL Feb 9 08:58:48.517433 bash[1146]: Updated "/home/core/.ssh/authorized_keys" Feb 9 08:58:48.519162 systemd[1]: Finished update-ssh-keys-after-ignition.service. Feb 9 08:58:48.522578 systemd-logind[1095]: Watching system buttons on /dev/input/event1 (Power Button) Feb 9 08:58:48.523077 systemd-logind[1095]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Feb 9 08:58:48.524321 env[1102]: time="2024-02-09T08:58:48.524222823Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Feb 9 08:58:48.526232 systemd-logind[1095]: New seat seat0. Feb 9 08:58:48.540252 systemd[1]: Started systemd-logind.service. Feb 9 08:58:48.553470 tar[1099]: ./static Feb 9 08:58:48.564641 coreos-metadata[1080]: Feb 09 08:58:48.564 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 08:58:48.598454 coreos-metadata[1080]: Feb 09 08:58:48.598 INFO Fetch successful Feb 9 08:58:48.613698 unknown[1080]: wrote ssh authorized keys file for user: core Feb 9 08:58:48.646676 update-ssh-keys[1152]: Updated "/home/core/.ssh/authorized_keys" Feb 9 08:58:48.647408 systemd[1]: Finished coreos-metadata-sshkeys@core.service. Feb 9 08:58:48.696538 env[1102]: time="2024-02-09T08:58:48.695305222Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Feb 9 08:58:48.696673 env[1102]: time="2024-02-09T08:58:48.696631807Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Feb 9 08:58:48.698761 env[1102]: time="2024-02-09T08:58:48.698702270Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.148-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Feb 9 08:58:48.698761 env[1102]: time="2024-02-09T08:58:48.698754539Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Feb 9 08:58:48.700299 env[1102]: time="2024-02-09T08:58:48.700248352Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 9 08:58:48.700299 env[1102]: time="2024-02-09T08:58:48.700291263Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Feb 9 08:58:48.700467 env[1102]: time="2024-02-09T08:58:48.700314419Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Feb 9 08:58:48.700467 env[1102]: time="2024-02-09T08:58:48.700333652Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Feb 9 08:58:48.700556 env[1102]: time="2024-02-09T08:58:48.700477429Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Feb 9 08:58:48.703066 env[1102]: time="2024-02-09T08:58:48.703002657Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Feb 9 08:58:48.703406 env[1102]: time="2024-02-09T08:58:48.703307325Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Feb 9 08:58:48.703406 env[1102]: time="2024-02-09T08:58:48.703402563Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Feb 9 08:58:48.703537 env[1102]: time="2024-02-09T08:58:48.703514215Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Feb 9 08:58:48.703593 env[1102]: time="2024-02-09T08:58:48.703538922Z" level=info msg="metadata content store policy set" policy=shared Feb 9 08:58:48.709159 env[1102]: time="2024-02-09T08:58:48.709103554Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Feb 9 08:58:48.709159 env[1102]: time="2024-02-09T08:58:48.709164381Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Feb 9 08:58:48.709301 env[1102]: time="2024-02-09T08:58:48.709186733Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Feb 9 08:58:48.709301 env[1102]: time="2024-02-09T08:58:48.709252252Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709301 env[1102]: time="2024-02-09T08:58:48.709276660Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709407 env[1102]: time="2024-02-09T08:58:48.709380476Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709446 env[1102]: time="2024-02-09T08:58:48.709404088Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709446 env[1102]: time="2024-02-09T08:58:48.709428552Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709497 env[1102]: time="2024-02-09T08:58:48.709450153Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709497 env[1102]: time="2024-02-09T08:58:48.709472660Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709548 env[1102]: time="2024-02-09T08:58:48.709493500Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.709548 env[1102]: time="2024-02-09T08:58:48.709515402Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Feb 9 08:58:48.709733 env[1102]: time="2024-02-09T08:58:48.709704283Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Feb 9 08:58:48.709851 env[1102]: time="2024-02-09T08:58:48.709828982Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Feb 9 08:58:48.710412 env[1102]: time="2024-02-09T08:58:48.710384341Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Feb 9 08:58:48.710461 env[1102]: time="2024-02-09T08:58:48.710434054Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710500 env[1102]: time="2024-02-09T08:58:48.710459206Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Feb 9 08:58:48.710557 env[1102]: time="2024-02-09T08:58:48.710534914Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710625 env[1102]: time="2024-02-09T08:58:48.710560103Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710675 env[1102]: time="2024-02-09T08:58:48.710648896Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710719 env[1102]: time="2024-02-09T08:58:48.710684526Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710719 env[1102]: time="2024-02-09T08:58:48.710707729Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710790 env[1102]: time="2024-02-09T08:58:48.710727916Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710790 env[1102]: time="2024-02-09T08:58:48.710747256Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710790 env[1102]: time="2024-02-09T08:58:48.710766224Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.710917 env[1102]: time="2024-02-09T08:58:48.710790783Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Feb 9 08:58:48.711013 env[1102]: time="2024-02-09T08:58:48.710981616Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.711089 env[1102]: time="2024-02-09T08:58:48.711015841Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.711089 env[1102]: time="2024-02-09T08:58:48.711039622Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.711089 env[1102]: time="2024-02-09T08:58:48.711082003Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Feb 9 08:58:48.711220 env[1102]: time="2024-02-09T08:58:48.711107782Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Feb 9 08:58:48.711220 env[1102]: time="2024-02-09T08:58:48.711127066Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Feb 9 08:58:48.711220 env[1102]: time="2024-02-09T08:58:48.711155724Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Feb 9 08:58:48.711220 env[1102]: time="2024-02-09T08:58:48.711204636Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Feb 9 08:58:48.711609 env[1102]: time="2024-02-09T08:58:48.711521844Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Feb 9 08:58:48.713862 env[1102]: time="2024-02-09T08:58:48.711618846Z" level=info msg="Connect containerd service" Feb 9 08:58:48.713862 env[1102]: time="2024-02-09T08:58:48.711663386Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Feb 9 08:58:48.713862 env[1102]: time="2024-02-09T08:58:48.712751643Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 9 08:58:48.715352 env[1102]: time="2024-02-09T08:58:48.715295962Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Feb 9 08:58:48.715463 env[1102]: time="2024-02-09T08:58:48.715401695Z" level=info msg=serving... address=/run/containerd/containerd.sock Feb 9 08:58:48.715593 systemd[1]: Started containerd.service. Feb 9 08:58:48.719145 env[1102]: time="2024-02-09T08:58:48.719094092Z" level=info msg="containerd successfully booted in 0.222896s" Feb 9 08:58:48.721075 env[1102]: time="2024-02-09T08:58:48.719661001Z" level=info msg="Start subscribing containerd event" Feb 9 08:58:48.721075 env[1102]: time="2024-02-09T08:58:48.719879548Z" level=info msg="Start recovering state" Feb 9 08:58:48.721075 env[1102]: time="2024-02-09T08:58:48.719986604Z" level=info msg="Start event monitor" Feb 9 08:58:48.721075 env[1102]: time="2024-02-09T08:58:48.720038103Z" level=info msg="Start snapshots syncer" Feb 9 08:58:48.721075 env[1102]: time="2024-02-09T08:58:48.720075764Z" level=info msg="Start cni network conf syncer for default" Feb 9 08:58:48.721075 env[1102]: time="2024-02-09T08:58:48.720127116Z" level=info msg="Start streaming server" Feb 9 08:58:48.763575 tar[1099]: ./vlan Feb 9 08:58:48.895819 tar[1099]: ./portmap Feb 9 08:58:48.935310 systemd-networkd[997]: eth0: Gained IPv6LL Feb 9 08:58:48.998990 tar[1099]: ./host-local Feb 9 08:58:49.084510 tar[1099]: ./vrf Feb 9 08:58:49.185200 tar[1099]: ./bridge Feb 9 08:58:49.284074 tar[1099]: ./tuning Feb 9 08:58:49.392229 tar[1099]: ./firewall Feb 9 08:58:49.502951 tar[1099]: ./host-device Feb 9 08:58:49.593706 tar[1099]: ./sbr Feb 9 08:58:49.681318 tar[1099]: ./loopback Feb 9 08:58:49.759896 tar[1099]: ./dhcp Feb 9 08:58:49.779765 tar[1100]: linux-amd64/LICENSE Feb 9 08:58:49.779765 tar[1100]: linux-amd64/README.md Feb 9 08:58:49.796150 systemd[1]: Finished prepare-helm.service. Feb 9 08:58:49.806942 systemd[1]: Finished prepare-critools.service. Feb 9 08:58:49.914651 tar[1099]: ./ptp Feb 9 08:58:49.963402 tar[1099]: ./ipvlan Feb 9 08:58:50.011153 tar[1099]: ./bandwidth Feb 9 08:58:50.037895 systemd[1]: Created slice system-sshd.slice. Feb 9 08:58:50.075004 systemd[1]: Finished prepare-cni-plugins.service. Feb 9 08:58:50.083877 locksmithd[1129]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Feb 9 08:58:51.612342 sshd_keygen[1114]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Feb 9 08:58:51.640819 systemd[1]: Finished sshd-keygen.service. Feb 9 08:58:51.643450 systemd[1]: Starting issuegen.service... Feb 9 08:58:51.645606 systemd[1]: Started sshd@0-64.23.148.25:22-139.178.89.65:58998.service. Feb 9 08:58:51.652330 systemd[1]: issuegen.service: Deactivated successfully. Feb 9 08:58:51.652614 systemd[1]: Finished issuegen.service. Feb 9 08:58:51.656550 systemd[1]: Starting systemd-user-sessions.service... Feb 9 08:58:51.667830 systemd[1]: Finished systemd-user-sessions.service. Feb 9 08:58:51.670650 systemd[1]: Started getty@tty1.service. Feb 9 08:58:51.673218 systemd[1]: Started serial-getty@ttyS0.service. Feb 9 08:58:51.673905 systemd[1]: Reached target getty.target. Feb 9 08:58:51.674598 systemd[1]: Reached target multi-user.target. Feb 9 08:58:51.676997 systemd[1]: Starting systemd-update-utmp-runlevel.service... Feb 9 08:58:51.690309 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Feb 9 08:58:51.690490 systemd[1]: Finished systemd-update-utmp-runlevel.service. Feb 9 08:58:51.691016 systemd[1]: Startup finished in 965ms (kernel) + 7.493s (initrd) + 9.538s (userspace) = 17.996s. Feb 9 08:58:51.741700 sshd[1171]: Accepted publickey for core from 139.178.89.65 port 58998 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 08:58:51.744814 sshd[1171]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 08:58:51.757591 systemd[1]: Created slice user-500.slice. Feb 9 08:58:51.759240 systemd[1]: Starting user-runtime-dir@500.service... Feb 9 08:58:51.764034 systemd-logind[1095]: New session 1 of user core. Feb 9 08:58:51.773583 systemd[1]: Finished user-runtime-dir@500.service. Feb 9 08:58:51.775415 systemd[1]: Starting user@500.service... Feb 9 08:58:51.779945 (systemd)[1182]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Feb 9 08:58:51.877820 systemd[1182]: Queued start job for default target default.target. Feb 9 08:58:51.878586 systemd[1182]: Reached target paths.target. Feb 9 08:58:51.878613 systemd[1182]: Reached target sockets.target. Feb 9 08:58:51.878628 systemd[1182]: Reached target timers.target. Feb 9 08:58:51.878641 systemd[1182]: Reached target basic.target. Feb 9 08:58:51.878808 systemd[1]: Started user@500.service. Feb 9 08:58:51.880459 systemd[1]: Started session-1.scope. Feb 9 08:58:51.881424 systemd[1182]: Reached target default.target. Feb 9 08:58:51.881668 systemd[1182]: Startup finished in 92ms. Feb 9 08:58:51.949506 systemd[1]: Started sshd@1-64.23.148.25:22-139.178.89.65:59010.service. Feb 9 08:58:51.999906 sshd[1191]: Accepted publickey for core from 139.178.89.65 port 59010 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 08:58:52.001864 sshd[1191]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 08:58:52.007153 systemd-logind[1095]: New session 2 of user core. Feb 9 08:58:52.008185 systemd[1]: Started session-2.scope. Feb 9 08:58:52.077243 sshd[1191]: pam_unix(sshd:session): session closed for user core Feb 9 08:58:52.084288 systemd[1]: Started sshd@2-64.23.148.25:22-139.178.89.65:59014.service. Feb 9 08:58:52.085101 systemd[1]: sshd@1-64.23.148.25:22-139.178.89.65:59010.service: Deactivated successfully. Feb 9 08:58:52.086294 systemd[1]: session-2.scope: Deactivated successfully. Feb 9 08:58:52.087030 systemd-logind[1095]: Session 2 logged out. Waiting for processes to exit. Feb 9 08:58:52.088423 systemd-logind[1095]: Removed session 2. Feb 9 08:58:52.134213 sshd[1196]: Accepted publickey for core from 139.178.89.65 port 59014 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 08:58:52.135810 sshd[1196]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 08:58:52.142409 systemd-logind[1095]: New session 3 of user core. Feb 9 08:58:52.142979 systemd[1]: Started session-3.scope. Feb 9 08:58:52.201577 sshd[1196]: pam_unix(sshd:session): session closed for user core Feb 9 08:58:52.206745 systemd[1]: sshd@2-64.23.148.25:22-139.178.89.65:59014.service: Deactivated successfully. Feb 9 08:58:52.207526 systemd[1]: session-3.scope: Deactivated successfully. Feb 9 08:58:52.208141 systemd-logind[1095]: Session 3 logged out. Waiting for processes to exit. Feb 9 08:58:52.209609 systemd[1]: Started sshd@3-64.23.148.25:22-139.178.89.65:59020.service. Feb 9 08:58:52.210887 systemd-logind[1095]: Removed session 3. Feb 9 08:58:52.260837 sshd[1203]: Accepted publickey for core from 139.178.89.65 port 59020 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 08:58:52.263439 sshd[1203]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 08:58:52.269691 systemd[1]: Started session-4.scope. Feb 9 08:58:52.270143 systemd-logind[1095]: New session 4 of user core. Feb 9 08:58:52.337177 sshd[1203]: pam_unix(sshd:session): session closed for user core Feb 9 08:58:52.342532 systemd[1]: sshd@3-64.23.148.25:22-139.178.89.65:59020.service: Deactivated successfully. Feb 9 08:58:52.343506 systemd[1]: session-4.scope: Deactivated successfully. Feb 9 08:58:52.344246 systemd-logind[1095]: Session 4 logged out. Waiting for processes to exit. Feb 9 08:58:52.346154 systemd[1]: Started sshd@4-64.23.148.25:22-139.178.89.65:59024.service. Feb 9 08:58:52.348159 systemd-logind[1095]: Removed session 4. Feb 9 08:58:52.405711 sshd[1209]: Accepted publickey for core from 139.178.89.65 port 59024 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 08:58:52.407569 sshd[1209]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 08:58:52.414735 systemd[1]: Started session-5.scope. Feb 9 08:58:52.415630 systemd-logind[1095]: New session 5 of user core. Feb 9 08:58:52.489495 sudo[1212]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Feb 9 08:58:52.489842 sudo[1212]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Feb 9 08:58:53.055132 systemd[1]: Starting systemd-networkd-wait-online.service... Feb 9 08:58:53.064798 systemd[1]: Finished systemd-networkd-wait-online.service. Feb 9 08:58:53.065333 systemd[1]: Reached target network-online.target. Feb 9 08:58:53.067524 systemd[1]: Starting docker.service... Feb 9 08:58:53.133362 env[1228]: time="2024-02-09T08:58:53.133275691Z" level=info msg="Starting up" Feb 9 08:58:53.136779 env[1228]: time="2024-02-09T08:58:53.136729065Z" level=info msg="parsed scheme: \"unix\"" module=grpc Feb 9 08:58:53.136779 env[1228]: time="2024-02-09T08:58:53.136765145Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Feb 9 08:58:53.136985 env[1228]: time="2024-02-09T08:58:53.136800315Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Feb 9 08:58:53.136985 env[1228]: time="2024-02-09T08:58:53.136818589Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Feb 9 08:58:53.139443 env[1228]: time="2024-02-09T08:58:53.139403016Z" level=info msg="parsed scheme: \"unix\"" module=grpc Feb 9 08:58:53.139642 env[1228]: time="2024-02-09T08:58:53.139620452Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Feb 9 08:58:53.139755 env[1228]: time="2024-02-09T08:58:53.139733110Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Feb 9 08:58:53.139838 env[1228]: time="2024-02-09T08:58:53.139821388Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Feb 9 08:58:53.147468 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport3105880464-merged.mount: Deactivated successfully. Feb 9 08:58:53.215064 env[1228]: time="2024-02-09T08:58:53.214975287Z" level=info msg="Loading containers: start." Feb 9 08:58:53.333072 kernel: Initializing XFRM netlink socket Feb 9 08:58:53.378943 env[1228]: time="2024-02-09T08:58:53.378884033Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Feb 9 08:58:53.467733 systemd-networkd[997]: docker0: Link UP Feb 9 08:58:53.480784 env[1228]: time="2024-02-09T08:58:53.480735670Z" level=info msg="Loading containers: done." Feb 9 08:58:53.497140 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1462710703-merged.mount: Deactivated successfully. Feb 9 08:58:53.501437 env[1228]: time="2024-02-09T08:58:53.501377183Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Feb 9 08:58:53.502139 env[1228]: time="2024-02-09T08:58:53.502105520Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Feb 9 08:58:53.502529 env[1228]: time="2024-02-09T08:58:53.502501209Z" level=info msg="Daemon has completed initialization" Feb 9 08:58:53.519840 systemd[1]: Started docker.service. Feb 9 08:58:53.533537 env[1228]: time="2024-02-09T08:58:53.533463059Z" level=info msg="API listen on /run/docker.sock" Feb 9 08:58:53.558471 systemd[1]: Starting coreos-metadata.service... Feb 9 08:58:53.614110 coreos-metadata[1346]: Feb 09 08:58:53.613 INFO Fetching http://169.254.169.254/metadata/v1.json: Attempt #1 Feb 9 08:58:53.627588 coreos-metadata[1346]: Feb 09 08:58:53.627 INFO Fetch successful Feb 9 08:58:53.644734 systemd[1]: Finished coreos-metadata.service. Feb 9 08:58:53.660547 systemd[1]: Reloading. Feb 9 08:58:53.730698 /usr/lib/systemd/system-generators/torcx-generator[1379]: time="2024-02-09T08:58:53Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 08:58:53.730738 /usr/lib/systemd/system-generators/torcx-generator[1379]: time="2024-02-09T08:58:53Z" level=info msg="torcx already run" Feb 9 08:58:53.872334 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 08:58:53.872362 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 08:58:53.909643 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 08:58:54.054526 systemd[1]: Started kubelet.service. Feb 9 08:58:54.170610 kubelet[1424]: E0209 08:58:54.170441 1424 run.go:74] "command failed" err="failed to validate kubelet flags: the container runtime endpoint address was not specified or empty, use --container-runtime-endpoint to set" Feb 9 08:58:54.173824 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 9 08:58:54.174020 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 9 08:58:54.461939 systemd[1]: Started sshd@5-64.23.148.25:22-141.98.11.90:52826.service. Feb 9 08:58:54.616428 env[1102]: time="2024-02-09T08:58:54.616353473Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.26.13\"" Feb 9 08:58:55.212066 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount148176708.mount: Deactivated successfully. Feb 9 08:58:55.857387 sshd[1437]: Invalid user admin from 141.98.11.90 port 52826 Feb 9 08:58:56.127284 sshd[1437]: pam_faillock(sshd:auth): User unknown Feb 9 08:58:56.128202 sshd[1437]: pam_unix(sshd:auth): check pass; user unknown Feb 9 08:58:56.128265 sshd[1437]: pam_unix(sshd:auth): authentication failure; logname= uid=0 euid=0 tty=ssh ruser= rhost=141.98.11.90 Feb 9 08:58:56.128861 sshd[1437]: pam_faillock(sshd:auth): User unknown Feb 9 08:58:57.273709 env[1102]: time="2024-02-09T08:58:57.273645108Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:57.275659 env[1102]: time="2024-02-09T08:58:57.275588767Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:84900298406b2df97ade16b73c49c2b73265ded8735ac19a4e20c2a4ad65853f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:57.278218 env[1102]: time="2024-02-09T08:58:57.278172432Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:57.280697 env[1102]: time="2024-02-09T08:58:57.280659005Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:2f28bed4096abd572a56595ac0304238bdc271dcfe22c650707c09bf97ec16fd,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:57.281947 env[1102]: time="2024-02-09T08:58:57.281907177Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.26.13\" returns image reference \"sha256:84900298406b2df97ade16b73c49c2b73265ded8735ac19a4e20c2a4ad65853f\"" Feb 9 08:58:57.301081 env[1102]: time="2024-02-09T08:58:57.301020061Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.26.13\"" Feb 9 08:58:58.407923 sshd[1437]: Failed password for invalid user admin from 141.98.11.90 port 52826 ssh2 Feb 9 08:58:59.562064 env[1102]: time="2024-02-09T08:58:59.561978150Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:59.564256 env[1102]: time="2024-02-09T08:58:59.564214197Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:921f237b560bdb02300f82d3606635d395b20635512fab10f0191cff42079486,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:59.566743 env[1102]: time="2024-02-09T08:58:59.566690665Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:59.569669 env[1102]: time="2024-02-09T08:58:59.569624952Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:fda420c6c15cdd01c4eba3404f0662fe486a9c7f38fa13c741a21334673841a2,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:58:59.569991 env[1102]: time="2024-02-09T08:58:59.569964977Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.26.13\" returns image reference \"sha256:921f237b560bdb02300f82d3606635d395b20635512fab10f0191cff42079486\"" Feb 9 08:58:59.594324 env[1102]: time="2024-02-09T08:58:59.594264747Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.26.13\"" Feb 9 08:59:00.744973 sshd[1437]: Connection closed by invalid user admin 141.98.11.90 port 52826 [preauth] Feb 9 08:59:00.746422 systemd[1]: sshd@5-64.23.148.25:22-141.98.11.90:52826.service: Deactivated successfully. Feb 9 08:59:01.330604 env[1102]: time="2024-02-09T08:59:01.330521818Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:01.332482 env[1102]: time="2024-02-09T08:59:01.332428923Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:4fe82b56f06250b6b7eb3d5a879cd2cfabf41cb3e45b24af6059eadbc3b8026e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:01.335349 env[1102]: time="2024-02-09T08:59:01.335257775Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:01.337705 env[1102]: time="2024-02-09T08:59:01.337656242Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:c3c7303ee6d01c8e5a769db28661cf854b55175aa72c67e9b6a7b9d47ac42af3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:01.338896 env[1102]: time="2024-02-09T08:59:01.338848829Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.26.13\" returns image reference \"sha256:4fe82b56f06250b6b7eb3d5a879cd2cfabf41cb3e45b24af6059eadbc3b8026e\"" Feb 9 08:59:01.357925 env[1102]: time="2024-02-09T08:59:01.357859462Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.26.13\"" Feb 9 08:59:02.653434 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount661048639.mount: Deactivated successfully. Feb 9 08:59:03.455805 env[1102]: time="2024-02-09T08:59:03.455735543Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:03.457329 env[1102]: time="2024-02-09T08:59:03.457272427Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:5a7325fa2b6e8d712e4a770abb4a5a5852e87b6de8df34552d67853e9bfb9f9f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:03.459028 env[1102]: time="2024-02-09T08:59:03.458986172Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.26.13,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:03.461079 env[1102]: time="2024-02-09T08:59:03.460988162Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:f6e0de32a002b910b9b2e0e8d769e2d7b05208240559c745ce4781082ab15f22,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:03.461879 env[1102]: time="2024-02-09T08:59:03.461821614Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.26.13\" returns image reference \"sha256:5a7325fa2b6e8d712e4a770abb4a5a5852e87b6de8df34552d67853e9bfb9f9f\"" Feb 9 08:59:03.478920 env[1102]: time="2024-02-09T08:59:03.478863627Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Feb 9 08:59:04.004310 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2480603105.mount: Deactivated successfully. Feb 9 08:59:04.006430 env[1102]: time="2024-02-09T08:59:04.006386106Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:04.007494 env[1102]: time="2024-02-09T08:59:04.007462275Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:04.009364 env[1102]: time="2024-02-09T08:59:04.009335335Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:04.011157 env[1102]: time="2024-02-09T08:59:04.011121682Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:04.012100 env[1102]: time="2024-02-09T08:59:04.012066625Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Feb 9 08:59:04.027235 env[1102]: time="2024-02-09T08:59:04.027187778Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.6-0\"" Feb 9 08:59:04.425030 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Feb 9 08:59:04.425278 systemd[1]: Stopped kubelet.service. Feb 9 08:59:04.427965 systemd[1]: Started kubelet.service. Feb 9 08:59:04.523452 kubelet[1473]: E0209 08:59:04.523372 1473 run.go:74] "command failed" err="failed to validate kubelet flags: the container runtime endpoint address was not specified or empty, use --container-runtime-endpoint to set" Feb 9 08:59:04.530511 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 9 08:59:04.530730 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 9 08:59:04.929478 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3030111458.mount: Deactivated successfully. Feb 9 08:59:10.249759 env[1102]: time="2024-02-09T08:59:10.249695299Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.6-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:10.251301 env[1102]: time="2024-02-09T08:59:10.251254863Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fce326961ae2d51a5f726883fd59d2a8c2ccc3e45d3bb859882db58e422e59e7,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:10.252979 env[1102]: time="2024-02-09T08:59:10.252944031Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.6-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:10.254599 env[1102]: time="2024-02-09T08:59:10.254566569Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:dd75ec974b0a2a6f6bb47001ba09207976e625db898d1b16735528c009cb171c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:10.255454 env[1102]: time="2024-02-09T08:59:10.255421630Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.6-0\" returns image reference \"sha256:fce326961ae2d51a5f726883fd59d2a8c2ccc3e45d3bb859882db58e422e59e7\"" Feb 9 08:59:10.267396 env[1102]: time="2024-02-09T08:59:10.267338859Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.9.3\"" Feb 9 08:59:10.799599 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2727329302.mount: Deactivated successfully. Feb 9 08:59:11.466931 env[1102]: time="2024-02-09T08:59:11.466873567Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.9.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:11.468441 env[1102]: time="2024-02-09T08:59:11.468390637Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:5185b96f0becf59032b8e3646e99f84d9655dff3ac9e2605e0dc77f9c441ae4a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:11.469748 env[1102]: time="2024-02-09T08:59:11.469719431Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.9.3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:11.471613 env[1102]: time="2024-02-09T08:59:11.471572095Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:8e352a029d304ca7431c6507b56800636c321cb52289686a581ab70aaa8a2e2a,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:11.472271 env[1102]: time="2024-02-09T08:59:11.472237955Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.9.3\" returns image reference \"sha256:5185b96f0becf59032b8e3646e99f84d9655dff3ac9e2605e0dc77f9c441ae4a\"" Feb 9 08:59:14.681948 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Feb 9 08:59:14.682178 systemd[1]: Stopped kubelet.service. Feb 9 08:59:14.684101 systemd[1]: Started kubelet.service. Feb 9 08:59:14.777624 kubelet[1543]: E0209 08:59:14.777544 1543 run.go:74] "command failed" err="failed to validate kubelet flags: the container runtime endpoint address was not specified or empty, use --container-runtime-endpoint to set" Feb 9 08:59:14.780577 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Feb 9 08:59:14.780764 systemd[1]: kubelet.service: Failed with result 'exit-code'. Feb 9 08:59:15.787830 systemd[1]: Stopped kubelet.service. Feb 9 08:59:15.805457 systemd[1]: Reloading. Feb 9 08:59:15.890323 /usr/lib/systemd/system-generators/torcx-generator[1572]: time="2024-02-09T08:59:15Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 08:59:15.890356 /usr/lib/systemd/system-generators/torcx-generator[1572]: time="2024-02-09T08:59:15Z" level=info msg="torcx already run" Feb 9 08:59:15.994073 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 08:59:15.994102 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 08:59:16.018736 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 08:59:16.114318 systemd[1]: Started kubelet.service. Feb 9 08:59:16.185325 kubelet[1619]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Feb 9 08:59:16.185325 kubelet[1619]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 9 08:59:16.185728 kubelet[1619]: I0209 08:59:16.185420 1619 server.go:198] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 9 08:59:16.187335 kubelet[1619]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Feb 9 08:59:16.187335 kubelet[1619]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 9 08:59:16.674295 kubelet[1619]: I0209 08:59:16.674260 1619 server.go:412] "Kubelet version" kubeletVersion="v1.26.5" Feb 9 08:59:16.674481 kubelet[1619]: I0209 08:59:16.674467 1619 server.go:414] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 9 08:59:16.674819 kubelet[1619]: I0209 08:59:16.674801 1619 server.go:836] "Client rotation is on, will bootstrap in background" Feb 9 08:59:16.680159 kubelet[1619]: I0209 08:59:16.680133 1619 server.go:659] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 9 08:59:16.680527 kubelet[1619]: I0209 08:59:16.680513 1619 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 9 08:59:16.681603 kubelet[1619]: I0209 08:59:16.681576 1619 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] CPUManagerPolicy:none CPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container CPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none ExperimentalTopologyManagerPolicyOptions:map[]} Feb 9 08:59:16.681817 kubelet[1619]: I0209 08:59:16.681800 1619 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Feb 9 08:59:16.681900 kubelet[1619]: I0209 08:59:16.681890 1619 container_manager_linux.go:308] "Creating device plugin manager" Feb 9 08:59:16.682028 kubelet[1619]: E0209 08:59:16.681450 1619 certificate_manager.go:471] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://64.23.148.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.682108 kubelet[1619]: I0209 08:59:16.681491 1619 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 9 08:59:16.682196 kubelet[1619]: I0209 08:59:16.682183 1619 state_mem.go:36] "Initialized new in-memory state store" Feb 9 08:59:16.685448 kubelet[1619]: I0209 08:59:16.685430 1619 kubelet.go:398] "Attempting to sync node with API server" Feb 9 08:59:16.685885 kubelet[1619]: I0209 08:59:16.685870 1619 kubelet.go:286] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 9 08:59:16.685997 kubelet[1619]: W0209 08:59:16.685959 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://64.23.148.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.2-5-6e53ffd492&limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.686038 kubelet[1619]: E0209 08:59:16.686008 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://64.23.148.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.2-5-6e53ffd492&limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.686038 kubelet[1619]: I0209 08:59:16.685988 1619 kubelet.go:297] "Adding apiserver pod source" Feb 9 08:59:16.686038 kubelet[1619]: I0209 08:59:16.686036 1619 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 9 08:59:16.686807 kubelet[1619]: I0209 08:59:16.686782 1619 kuberuntime_manager.go:244] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Feb 9 08:59:16.687117 kubelet[1619]: W0209 08:59:16.687101 1619 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Feb 9 08:59:16.687456 kubelet[1619]: I0209 08:59:16.687440 1619 server.go:1186] "Started kubelet" Feb 9 08:59:16.687547 kubelet[1619]: W0209 08:59:16.687520 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://64.23.148.25:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.687611 kubelet[1619]: E0209 08:59:16.687552 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://64.23.148.25:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.692458 kubelet[1619]: I0209 08:59:16.692439 1619 server.go:161] "Starting to listen" address="0.0.0.0" port=10250 Feb 9 08:59:16.693659 kubelet[1619]: E0209 08:59:16.693542 1619 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ci-3510.3.2-5-6e53ffd492.17b22626c54a1ca7", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ci-3510.3.2-5-6e53ffd492", UID:"ci-3510.3.2-5-6e53ffd492", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"ci-3510.3.2-5-6e53ffd492"}, FirstTimestamp:time.Date(2024, time.February, 9, 8, 59, 16, 687420583, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 8, 59, 16, 687420583, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://64.23.148.25:6443/api/v1/namespaces/default/events": dial tcp 64.23.148.25:6443: connect: connection refused'(may retry after sleeping) Feb 9 08:59:16.695492 kubelet[1619]: E0209 08:59:16.695475 1619 cri_stats_provider.go:455] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Feb 9 08:59:16.695589 kubelet[1619]: E0209 08:59:16.695578 1619 kubelet.go:1386] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 9 08:59:16.697225 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Feb 9 08:59:16.697378 kubelet[1619]: I0209 08:59:16.697359 1619 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 9 08:59:16.697636 kubelet[1619]: I0209 08:59:16.697623 1619 server.go:451] "Adding debug handlers to kubelet server" Feb 9 08:59:16.701311 kubelet[1619]: E0209 08:59:16.701295 1619 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"ci-3510.3.2-5-6e53ffd492\" not found" Feb 9 08:59:16.701439 kubelet[1619]: I0209 08:59:16.701428 1619 volume_manager.go:293] "Starting Kubelet Volume Manager" Feb 9 08:59:16.701564 kubelet[1619]: I0209 08:59:16.701553 1619 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Feb 9 08:59:16.701988 kubelet[1619]: W0209 08:59:16.701957 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://64.23.148.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.702127 kubelet[1619]: E0209 08:59:16.702113 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://64.23.148.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.702401 kubelet[1619]: E0209 08:59:16.702382 1619 controller.go:146] failed to ensure lease exists, will retry in 200ms, error: Get "https://64.23.148.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.2-5-6e53ffd492?timeout=10s": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.737902 kubelet[1619]: I0209 08:59:16.737872 1619 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 9 08:59:16.738097 kubelet[1619]: I0209 08:59:16.738084 1619 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 9 08:59:16.738198 kubelet[1619]: I0209 08:59:16.738189 1619 state_mem.go:36] "Initialized new in-memory state store" Feb 9 08:59:16.739780 kubelet[1619]: I0209 08:59:16.739759 1619 policy_none.go:49] "None policy: Start" Feb 9 08:59:16.740425 kubelet[1619]: I0209 08:59:16.740406 1619 memory_manager.go:169] "Starting memorymanager" policy="None" Feb 9 08:59:16.740513 kubelet[1619]: I0209 08:59:16.740431 1619 state_mem.go:35] "Initializing new in-memory state store" Feb 9 08:59:16.745000 systemd[1]: Created slice kubepods.slice. Feb 9 08:59:16.749547 kubelet[1619]: I0209 08:59:16.749314 1619 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Feb 9 08:59:16.749434 systemd[1]: Created slice kubepods-burstable.slice. Feb 9 08:59:16.753871 systemd[1]: Created slice kubepods-besteffort.slice. Feb 9 08:59:16.760030 kubelet[1619]: I0209 08:59:16.760003 1619 manager.go:455] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 9 08:59:16.760717 kubelet[1619]: I0209 08:59:16.760679 1619 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 9 08:59:16.761621 kubelet[1619]: E0209 08:59:16.761604 1619 eviction_manager.go:261] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510.3.2-5-6e53ffd492\" not found" Feb 9 08:59:16.784823 kubelet[1619]: I0209 08:59:16.784790 1619 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Feb 9 08:59:16.785056 kubelet[1619]: I0209 08:59:16.785030 1619 status_manager.go:176] "Starting to sync pod status with apiserver" Feb 9 08:59:16.785291 kubelet[1619]: I0209 08:59:16.785278 1619 kubelet.go:2113] "Starting kubelet main sync loop" Feb 9 08:59:16.785449 kubelet[1619]: E0209 08:59:16.785439 1619 kubelet.go:2137] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Feb 9 08:59:16.786082 kubelet[1619]: W0209 08:59:16.786022 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://64.23.148.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.786221 kubelet[1619]: E0209 08:59:16.786209 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://64.23.148.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.802941 kubelet[1619]: I0209 08:59:16.802913 1619 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:16.803565 kubelet[1619]: E0209 08:59:16.803531 1619 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://64.23.148.25:6443/api/v1/nodes\": dial tcp 64.23.148.25:6443: connect: connection refused" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:16.885952 kubelet[1619]: I0209 08:59:16.885868 1619 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:16.886991 kubelet[1619]: I0209 08:59:16.886969 1619 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:16.888068 kubelet[1619]: I0209 08:59:16.888036 1619 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:16.888888 kubelet[1619]: I0209 08:59:16.888867 1619 status_manager.go:698] "Failed to get status for pod" podUID=4c1af3803e6ef0d2c5bf3b768d7044e7 pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" err="Get \"https://64.23.148.25:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-ci-3510.3.2-5-6e53ffd492\": dial tcp 64.23.148.25:6443: connect: connection refused" Feb 9 08:59:16.889600 kubelet[1619]: I0209 08:59:16.889561 1619 status_manager.go:698] "Failed to get status for pod" podUID=b9bac6a0ccc5042be8f5d46f6272401c pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" err="Get \"https://64.23.148.25:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ci-3510.3.2-5-6e53ffd492\": dial tcp 64.23.148.25:6443: connect: connection refused" Feb 9 08:59:16.890827 kubelet[1619]: I0209 08:59:16.890808 1619 status_manager.go:698] "Failed to get status for pod" podUID=297285cbca3cdb75be03345ce23f82d0 pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" err="Get \"https://64.23.148.25:6443/api/v1/namespaces/kube-system/pods/kube-scheduler-ci-3510.3.2-5-6e53ffd492\": dial tcp 64.23.148.25:6443: connect: connection refused" Feb 9 08:59:16.894140 systemd[1]: Created slice kubepods-burstable-pod4c1af3803e6ef0d2c5bf3b768d7044e7.slice. Feb 9 08:59:16.903075 kubelet[1619]: E0209 08:59:16.903005 1619 controller.go:146] failed to ensure lease exists, will retry in 400ms, error: Get "https://64.23.148.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.2-5-6e53ffd492?timeout=10s": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:16.904622 systemd[1]: Created slice kubepods-burstable-podb9bac6a0ccc5042be8f5d46f6272401c.slice. Feb 9 08:59:16.915427 systemd[1]: Created slice kubepods-burstable-pod297285cbca3cdb75be03345ce23f82d0.slice. Feb 9 08:59:17.002606 kubelet[1619]: I0209 08:59:17.002553 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.002922 kubelet[1619]: I0209 08:59:17.002860 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/297285cbca3cdb75be03345ce23f82d0-kubeconfig\") pod \"kube-scheduler-ci-3510.3.2-5-6e53ffd492\" (UID: \"297285cbca3cdb75be03345ce23f82d0\") " pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003088 kubelet[1619]: I0209 08:59:17.003038 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4c1af3803e6ef0d2c5bf3b768d7044e7-ca-certs\") pod \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" (UID: \"4c1af3803e6ef0d2c5bf3b768d7044e7\") " pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003239 kubelet[1619]: I0209 08:59:17.003227 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4c1af3803e6ef0d2c5bf3b768d7044e7-k8s-certs\") pod \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" (UID: \"4c1af3803e6ef0d2c5bf3b768d7044e7\") " pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003341 kubelet[1619]: I0209 08:59:17.003331 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4c1af3803e6ef0d2c5bf3b768d7044e7-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" (UID: \"4c1af3803e6ef0d2c5bf3b768d7044e7\") " pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003461 kubelet[1619]: I0209 08:59:17.003452 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-ca-certs\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003558 kubelet[1619]: I0209 08:59:17.003549 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003688 kubelet[1619]: I0209 08:59:17.003676 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.003791 kubelet[1619]: I0209 08:59:17.003782 1619 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.004576 kubelet[1619]: I0209 08:59:17.004529 1619 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.004921 kubelet[1619]: E0209 08:59:17.004904 1619 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://64.23.148.25:6443/api/v1/nodes\": dial tcp 64.23.148.25:6443: connect: connection refused" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.202007 kubelet[1619]: E0209 08:59:17.201964 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:17.203915 env[1102]: time="2024-02-09T08:59:17.203521936Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.2-5-6e53ffd492,Uid:4c1af3803e6ef0d2c5bf3b768d7044e7,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:17.207955 kubelet[1619]: E0209 08:59:17.207904 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:17.208667 env[1102]: time="2024-02-09T08:59:17.208610302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.2-5-6e53ffd492,Uid:b9bac6a0ccc5042be8f5d46f6272401c,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:17.218840 kubelet[1619]: E0209 08:59:17.218801 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:17.220092 env[1102]: time="2024-02-09T08:59:17.219750275Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.2-5-6e53ffd492,Uid:297285cbca3cdb75be03345ce23f82d0,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:17.304429 kubelet[1619]: E0209 08:59:17.304271 1619 controller.go:146] failed to ensure lease exists, will retry in 800ms, error: Get "https://64.23.148.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.2-5-6e53ffd492?timeout=10s": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.406830 kubelet[1619]: I0209 08:59:17.406484 1619 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.407166 kubelet[1619]: E0209 08:59:17.406953 1619 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://64.23.148.25:6443/api/v1/nodes\": dial tcp 64.23.148.25:6443: connect: connection refused" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:17.707903 kubelet[1619]: W0209 08:59:17.707792 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://64.23.148.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.2-5-6e53ffd492&limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.707903 kubelet[1619]: E0209 08:59:17.707900 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://64.23.148.25:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.2-5-6e53ffd492&limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.733919 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1420054441.mount: Deactivated successfully. Feb 9 08:59:17.738484 env[1102]: time="2024-02-09T08:59:17.738409022Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.739797 env[1102]: time="2024-02-09T08:59:17.739748033Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.743522 env[1102]: time="2024-02-09T08:59:17.743474174Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.746183 env[1102]: time="2024-02-09T08:59:17.746139281Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.750116 env[1102]: time="2024-02-09T08:59:17.750072397Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.751805 env[1102]: time="2024-02-09T08:59:17.751735528Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.755389 env[1102]: time="2024-02-09T08:59:17.755352150Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.756444 env[1102]: time="2024-02-09T08:59:17.756411285Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.757516 env[1102]: time="2024-02-09T08:59:17.757490596Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.758458 env[1102]: time="2024-02-09T08:59:17.758420456Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.759357 env[1102]: time="2024-02-09T08:59:17.759331052Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.760315 env[1102]: time="2024-02-09T08:59:17.760289952Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:17.783523 env[1102]: time="2024-02-09T08:59:17.780610814Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 08:59:17.783523 env[1102]: time="2024-02-09T08:59:17.780663584Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 08:59:17.783523 env[1102]: time="2024-02-09T08:59:17.780681722Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 08:59:17.783523 env[1102]: time="2024-02-09T08:59:17.780880050Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9889d70fedee20d0c394132c98dc458107e6892170ea32fa22b6c5e3d801c4f8 pid=1696 runtime=io.containerd.runc.v2 Feb 9 08:59:17.793398 env[1102]: time="2024-02-09T08:59:17.793303158Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 08:59:17.793398 env[1102]: time="2024-02-09T08:59:17.793345985Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 08:59:17.793398 env[1102]: time="2024-02-09T08:59:17.793357075Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 08:59:17.793870 env[1102]: time="2024-02-09T08:59:17.793746004Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/d3c5e10de54abd43e544fe48790c342f1a954719725730122e95744568e04df9 pid=1707 runtime=io.containerd.runc.v2 Feb 9 08:59:17.816025 systemd[1]: Started cri-containerd-9889d70fedee20d0c394132c98dc458107e6892170ea32fa22b6c5e3d801c4f8.scope. Feb 9 08:59:17.837100 systemd[1]: Started cri-containerd-d3c5e10de54abd43e544fe48790c342f1a954719725730122e95744568e04df9.scope. Feb 9 08:59:17.857159 env[1102]: time="2024-02-09T08:59:17.853481686Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 08:59:17.857159 env[1102]: time="2024-02-09T08:59:17.853522869Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 08:59:17.857159 env[1102]: time="2024-02-09T08:59:17.853534853Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 08:59:17.857159 env[1102]: time="2024-02-09T08:59:17.853663002Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/36a5454cc1aa1e2e6e0b1007d74aa589aee4dcad445e8a7b08cd22ac604bc62a pid=1743 runtime=io.containerd.runc.v2 Feb 9 08:59:17.876963 kubelet[1619]: W0209 08:59:17.876882 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://64.23.148.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.876963 kubelet[1619]: E0209 08:59:17.876941 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://64.23.148.25:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.882319 systemd[1]: Started cri-containerd-36a5454cc1aa1e2e6e0b1007d74aa589aee4dcad445e8a7b08cd22ac604bc62a.scope. Feb 9 08:59:17.930270 env[1102]: time="2024-02-09T08:59:17.930108634Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.2-5-6e53ffd492,Uid:4c1af3803e6ef0d2c5bf3b768d7044e7,Namespace:kube-system,Attempt:0,} returns sandbox id \"9889d70fedee20d0c394132c98dc458107e6892170ea32fa22b6c5e3d801c4f8\"" Feb 9 08:59:17.931854 kubelet[1619]: E0209 08:59:17.931826 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:17.946077 env[1102]: time="2024-02-09T08:59:17.946000822Z" level=info msg="CreateContainer within sandbox \"9889d70fedee20d0c394132c98dc458107e6892170ea32fa22b6c5e3d801c4f8\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Feb 9 08:59:17.950715 kubelet[1619]: W0209 08:59:17.950577 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://64.23.148.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.950715 kubelet[1619]: E0209 08:59:17.950680 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://64.23.148.25:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:17.970353 env[1102]: time="2024-02-09T08:59:17.969433845Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.2-5-6e53ffd492,Uid:b9bac6a0ccc5042be8f5d46f6272401c,Namespace:kube-system,Attempt:0,} returns sandbox id \"d3c5e10de54abd43e544fe48790c342f1a954719725730122e95744568e04df9\"" Feb 9 08:59:17.971457 kubelet[1619]: E0209 08:59:17.971227 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:17.975616 env[1102]: time="2024-02-09T08:59:17.975542291Z" level=info msg="CreateContainer within sandbox \"d3c5e10de54abd43e544fe48790c342f1a954719725730122e95744568e04df9\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Feb 9 08:59:17.988003 env[1102]: time="2024-02-09T08:59:17.987946497Z" level=info msg="CreateContainer within sandbox \"9889d70fedee20d0c394132c98dc458107e6892170ea32fa22b6c5e3d801c4f8\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b24e301f85f4c843ec05f6b652db71609fcc91f6fcc8d66c7a0b8264bbdb59f3\"" Feb 9 08:59:17.989039 env[1102]: time="2024-02-09T08:59:17.988995066Z" level=info msg="StartContainer for \"b24e301f85f4c843ec05f6b652db71609fcc91f6fcc8d66c7a0b8264bbdb59f3\"" Feb 9 08:59:17.992975 env[1102]: time="2024-02-09T08:59:17.992928468Z" level=info msg="CreateContainer within sandbox \"d3c5e10de54abd43e544fe48790c342f1a954719725730122e95744568e04df9\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"972713abd8fbaa854a5cd4c89ca487668a053249e817bcb77e6326a57b69c8c2\"" Feb 9 08:59:17.993725 env[1102]: time="2024-02-09T08:59:17.993689837Z" level=info msg="StartContainer for \"972713abd8fbaa854a5cd4c89ca487668a053249e817bcb77e6326a57b69c8c2\"" Feb 9 08:59:18.008129 env[1102]: time="2024-02-09T08:59:18.008013020Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.2-5-6e53ffd492,Uid:297285cbca3cdb75be03345ce23f82d0,Namespace:kube-system,Attempt:0,} returns sandbox id \"36a5454cc1aa1e2e6e0b1007d74aa589aee4dcad445e8a7b08cd22ac604bc62a\"" Feb 9 08:59:18.009826 kubelet[1619]: E0209 08:59:18.009523 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:18.012208 env[1102]: time="2024-02-09T08:59:18.012160819Z" level=info msg="CreateContainer within sandbox \"36a5454cc1aa1e2e6e0b1007d74aa589aee4dcad445e8a7b08cd22ac604bc62a\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Feb 9 08:59:18.031421 env[1102]: time="2024-02-09T08:59:18.031339644Z" level=info msg="CreateContainer within sandbox \"36a5454cc1aa1e2e6e0b1007d74aa589aee4dcad445e8a7b08cd22ac604bc62a\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"d5f375e97cc7078c11948f334761bc49fcad2d7c95ffdd86a0c0255f8041e78b\"" Feb 9 08:59:18.032227 env[1102]: time="2024-02-09T08:59:18.032191371Z" level=info msg="StartContainer for \"d5f375e97cc7078c11948f334761bc49fcad2d7c95ffdd86a0c0255f8041e78b\"" Feb 9 08:59:18.040926 systemd[1]: Started cri-containerd-b24e301f85f4c843ec05f6b652db71609fcc91f6fcc8d66c7a0b8264bbdb59f3.scope. Feb 9 08:59:18.065154 systemd[1]: Started cri-containerd-972713abd8fbaa854a5cd4c89ca487668a053249e817bcb77e6326a57b69c8c2.scope. Feb 9 08:59:18.100972 systemd[1]: Started cri-containerd-d5f375e97cc7078c11948f334761bc49fcad2d7c95ffdd86a0c0255f8041e78b.scope. Feb 9 08:59:18.106086 kubelet[1619]: E0209 08:59:18.105441 1619 controller.go:146] failed to ensure lease exists, will retry in 1.6s, error: Get "https://64.23.148.25:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.2-5-6e53ffd492?timeout=10s": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:18.144919 env[1102]: time="2024-02-09T08:59:18.144859076Z" level=info msg="StartContainer for \"b24e301f85f4c843ec05f6b652db71609fcc91f6fcc8d66c7a0b8264bbdb59f3\" returns successfully" Feb 9 08:59:18.180530 env[1102]: time="2024-02-09T08:59:18.180472496Z" level=info msg="StartContainer for \"972713abd8fbaa854a5cd4c89ca487668a053249e817bcb77e6326a57b69c8c2\" returns successfully" Feb 9 08:59:18.208265 kubelet[1619]: I0209 08:59:18.208228 1619 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:18.208733 kubelet[1619]: E0209 08:59:18.208641 1619 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://64.23.148.25:6443/api/v1/nodes\": dial tcp 64.23.148.25:6443: connect: connection refused" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:18.215061 env[1102]: time="2024-02-09T08:59:18.214989070Z" level=info msg="StartContainer for \"d5f375e97cc7078c11948f334761bc49fcad2d7c95ffdd86a0c0255f8041e78b\" returns successfully" Feb 9 08:59:18.234015 kubelet[1619]: W0209 08:59:18.233847 1619 reflector.go:424] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://64.23.148.25:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:18.234015 kubelet[1619]: E0209 08:59:18.233925 1619 reflector.go:140] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://64.23.148.25:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:18.529318 kubelet[1619]: E0209 08:59:18.529100 1619 event.go:276] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ci-3510.3.2-5-6e53ffd492.17b22626c54a1ca7", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ci-3510.3.2-5-6e53ffd492", UID:"ci-3510.3.2-5-6e53ffd492", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"ci-3510.3.2-5-6e53ffd492"}, FirstTimestamp:time.Date(2024, time.February, 9, 8, 59, 16, 687420583, time.Local), LastTimestamp:time.Date(2024, time.February, 9, 8, 59, 16, 687420583, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://64.23.148.25:6443/api/v1/namespaces/default/events": dial tcp 64.23.148.25:6443: connect: connection refused'(may retry after sleeping) Feb 9 08:59:18.796116 kubelet[1619]: E0209 08:59:18.795982 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:18.796564 kubelet[1619]: I0209 08:59:18.796545 1619 status_manager.go:698] "Failed to get status for pod" podUID=b9bac6a0ccc5042be8f5d46f6272401c pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" err="Get \"https://64.23.148.25:6443/api/v1/namespaces/kube-system/pods/kube-controller-manager-ci-3510.3.2-5-6e53ffd492\": dial tcp 64.23.148.25:6443: connect: connection refused" Feb 9 08:59:18.801466 kubelet[1619]: E0209 08:59:18.801429 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:18.801839 kubelet[1619]: I0209 08:59:18.801817 1619 status_manager.go:698] "Failed to get status for pod" podUID=297285cbca3cdb75be03345ce23f82d0 pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" err="Get \"https://64.23.148.25:6443/api/v1/namespaces/kube-system/pods/kube-scheduler-ci-3510.3.2-5-6e53ffd492\": dial tcp 64.23.148.25:6443: connect: connection refused" Feb 9 08:59:18.804931 kubelet[1619]: E0209 08:59:18.804894 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:18.847552 kubelet[1619]: E0209 08:59:18.847493 1619 certificate_manager.go:471] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://64.23.148.25:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 64.23.148.25:6443: connect: connection refused Feb 9 08:59:18.887168 kubelet[1619]: I0209 08:59:18.887124 1619 status_manager.go:698] "Failed to get status for pod" podUID=4c1af3803e6ef0d2c5bf3b768d7044e7 pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" err="Get \"https://64.23.148.25:6443/api/v1/namespaces/kube-system/pods/kube-apiserver-ci-3510.3.2-5-6e53ffd492\": dial tcp 64.23.148.25:6443: connect: connection refused" Feb 9 08:59:19.807745 kubelet[1619]: E0209 08:59:19.807705 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:19.808412 kubelet[1619]: E0209 08:59:19.808390 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:19.809247 kubelet[1619]: E0209 08:59:19.809226 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:19.810246 kubelet[1619]: I0209 08:59:19.810221 1619 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:20.808398 kubelet[1619]: E0209 08:59:20.808356 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:21.242226 kubelet[1619]: E0209 08:59:21.242175 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:21.813663 kubelet[1619]: E0209 08:59:21.813631 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:21.945882 kubelet[1619]: I0209 08:59:21.945841 1619 kubelet_node_status.go:73] "Successfully registered node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:22.045111 kubelet[1619]: E0209 08:59:22.045069 1619 controller.go:146] failed to ensure lease exists, will retry in 3.2s, error: namespaces "kube-node-lease" not found Feb 9 08:59:22.690200 kubelet[1619]: I0209 08:59:22.690137 1619 apiserver.go:52] "Watching apiserver" Feb 9 08:59:22.702057 kubelet[1619]: I0209 08:59:22.701988 1619 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Feb 9 08:59:22.735390 kubelet[1619]: I0209 08:59:22.735338 1619 reconciler.go:41] "Reconciler: start to sync state" Feb 9 08:59:24.589503 kubelet[1619]: E0209 08:59:24.589469 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:24.817515 kubelet[1619]: E0209 08:59:24.817474 1619 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:25.353723 systemd[1]: Reloading. Feb 9 08:59:25.460506 /usr/lib/systemd/system-generators/torcx-generator[1948]: time="2024-02-09T08:59:25Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.2 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.2 /var/lib/torcx/store]" Feb 9 08:59:25.461119 /usr/lib/systemd/system-generators/torcx-generator[1948]: time="2024-02-09T08:59:25Z" level=info msg="torcx already run" Feb 9 08:59:25.583739 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Feb 9 08:59:25.583759 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Feb 9 08:59:25.607162 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Feb 9 08:59:25.721647 kubelet[1619]: I0209 08:59:25.721466 1619 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 9 08:59:25.721757 systemd[1]: Stopping kubelet.service... Feb 9 08:59:25.733959 systemd[1]: kubelet.service: Deactivated successfully. Feb 9 08:59:25.734193 systemd[1]: Stopped kubelet.service. Feb 9 08:59:25.736634 systemd[1]: Started kubelet.service. Feb 9 08:59:25.848547 kubelet[1995]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Feb 9 08:59:25.848939 kubelet[1995]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 9 08:59:25.849362 kubelet[1995]: I0209 08:59:25.849322 1995 server.go:198] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 9 08:59:25.853823 kubelet[1995]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.27. Image garbage collector will get sandbox image information from CRI. Feb 9 08:59:25.853954 kubelet[1995]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 9 08:59:25.859620 kubelet[1995]: I0209 08:59:25.858987 1995 server.go:412] "Kubelet version" kubeletVersion="v1.26.5" Feb 9 08:59:25.859620 kubelet[1995]: I0209 08:59:25.859020 1995 server.go:414] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 9 08:59:25.859620 kubelet[1995]: I0209 08:59:25.859311 1995 server.go:836] "Client rotation is on, will bootstrap in background" Feb 9 08:59:25.860888 kubelet[1995]: I0209 08:59:25.860865 1995 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 9 08:59:25.861754 kubelet[1995]: I0209 08:59:25.861676 1995 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Feb 9 08:59:25.873537 kubelet[1995]: I0209 08:59:25.872874 1995 server.go:659] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Feb 9 08:59:25.873537 kubelet[1995]: I0209 08:59:25.873175 1995 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 9 08:59:25.873537 kubelet[1995]: I0209 08:59:25.873248 1995 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:systemd KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] CPUManagerPolicy:none CPUManagerPolicyOptions:map[] ExperimentalTopologyManagerScope:container CPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] ExperimentalPodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms ExperimentalTopologyManagerPolicy:none ExperimentalTopologyManagerPolicyOptions:map[]} Feb 9 08:59:25.873537 kubelet[1995]: I0209 08:59:25.873270 1995 topology_manager.go:134] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Feb 9 08:59:25.873537 kubelet[1995]: I0209 08:59:25.873283 1995 container_manager_linux.go:308] "Creating device plugin manager" Feb 9 08:59:25.873537 kubelet[1995]: I0209 08:59:25.873324 1995 state_mem.go:36] "Initialized new in-memory state store" Feb 9 08:59:25.879340 kubelet[1995]: I0209 08:59:25.879284 1995 kubelet.go:398] "Attempting to sync node with API server" Feb 9 08:59:25.879340 kubelet[1995]: I0209 08:59:25.879309 1995 kubelet.go:286] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 9 08:59:25.879340 kubelet[1995]: I0209 08:59:25.879332 1995 kubelet.go:297] "Adding apiserver pod source" Feb 9 08:59:25.879340 kubelet[1995]: I0209 08:59:25.879348 1995 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 9 08:59:25.889707 kubelet[1995]: I0209 08:59:25.886706 1995 kuberuntime_manager.go:244] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Feb 9 08:59:25.889707 kubelet[1995]: I0209 08:59:25.887508 1995 server.go:1186] "Started kubelet" Feb 9 08:59:25.890695 kubelet[1995]: I0209 08:59:25.890669 1995 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 9 08:59:25.896124 sudo[2008]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Feb 9 08:59:25.896978 kubelet[1995]: I0209 08:59:25.896211 1995 server.go:161] "Starting to listen" address="0.0.0.0" port=10250 Feb 9 08:59:25.896402 sudo[2008]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Feb 9 08:59:25.898384 kubelet[1995]: I0209 08:59:25.898250 1995 server.go:451] "Adding debug handlers to kubelet server" Feb 9 08:59:25.912073 kubelet[1995]: I0209 08:59:25.907357 1995 volume_manager.go:293] "Starting Kubelet Volume Manager" Feb 9 08:59:25.912073 kubelet[1995]: I0209 08:59:25.907476 1995 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Feb 9 08:59:25.940398 kubelet[1995]: E0209 08:59:25.940364 1995 cri_stats_provider.go:455] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Feb 9 08:59:25.941019 kubelet[1995]: E0209 08:59:25.940905 1995 kubelet.go:1386] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Feb 9 08:59:26.004105 kubelet[1995]: I0209 08:59:26.003628 1995 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Feb 9 08:59:26.033007 kubelet[1995]: I0209 08:59:26.032374 1995 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.048094 kubelet[1995]: I0209 08:59:26.046489 1995 cpu_manager.go:214] "Starting CPU manager" policy="none" Feb 9 08:59:26.048094 kubelet[1995]: I0209 08:59:26.046530 1995 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Feb 9 08:59:26.048094 kubelet[1995]: I0209 08:59:26.046557 1995 state_mem.go:36] "Initialized new in-memory state store" Feb 9 08:59:26.048094 kubelet[1995]: I0209 08:59:26.046987 1995 state_mem.go:88] "Updated default CPUSet" cpuSet="" Feb 9 08:59:26.048094 kubelet[1995]: I0209 08:59:26.047007 1995 state_mem.go:96] "Updated CPUSet assignments" assignments=map[] Feb 9 08:59:26.048094 kubelet[1995]: I0209 08:59:26.047014 1995 policy_none.go:49] "None policy: Start" Feb 9 08:59:26.048500 kubelet[1995]: I0209 08:59:26.048128 1995 memory_manager.go:169] "Starting memorymanager" policy="None" Feb 9 08:59:26.048500 kubelet[1995]: I0209 08:59:26.048161 1995 state_mem.go:35] "Initializing new in-memory state store" Feb 9 08:59:26.048500 kubelet[1995]: I0209 08:59:26.048320 1995 state_mem.go:75] "Updated machine memory state" Feb 9 08:59:26.072866 kubelet[1995]: I0209 08:59:26.072125 1995 manager.go:455] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 9 08:59:26.072866 kubelet[1995]: I0209 08:59:26.072462 1995 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 9 08:59:26.079075 kubelet[1995]: I0209 08:59:26.077221 1995 kubelet_node_status.go:108] "Node was previously registered" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.079075 kubelet[1995]: I0209 08:59:26.077285 1995 kubelet_node_status.go:73] "Successfully registered node" node="ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.126888 kubelet[1995]: I0209 08:59:26.126737 1995 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Feb 9 08:59:26.127153 kubelet[1995]: I0209 08:59:26.127131 1995 status_manager.go:176] "Starting to sync pod status with apiserver" Feb 9 08:59:26.127299 kubelet[1995]: I0209 08:59:26.127284 1995 kubelet.go:2113] "Starting kubelet main sync loop" Feb 9 08:59:26.127511 kubelet[1995]: E0209 08:59:26.127496 1995 kubelet.go:2137] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Feb 9 08:59:26.228283 kubelet[1995]: I0209 08:59:26.228240 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:26.228826 kubelet[1995]: I0209 08:59:26.228796 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:26.229408 kubelet[1995]: I0209 08:59:26.229387 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:26.298076 kubelet[1995]: E0209 08:59:26.295319 1995 kubelet.go:1802] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-3510.3.2-5-6e53ffd492\" already exists" pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.317385 kubelet[1995]: I0209 08:59:26.317345 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/4c1af3803e6ef0d2c5bf3b768d7044e7-k8s-certs\") pod \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" (UID: \"4c1af3803e6ef0d2c5bf3b768d7044e7\") " pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.317674 kubelet[1995]: I0209 08:59:26.317642 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-ca-certs\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.317857 kubelet[1995]: I0209 08:59:26.317844 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.318013 kubelet[1995]: I0209 08:59:26.318001 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.318165 kubelet[1995]: I0209 08:59:26.318153 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/4c1af3803e6ef0d2c5bf3b768d7044e7-ca-certs\") pod \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" (UID: \"4c1af3803e6ef0d2c5bf3b768d7044e7\") " pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.318313 kubelet[1995]: I0209 08:59:26.318299 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/4c1af3803e6ef0d2c5bf3b768d7044e7-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" (UID: \"4c1af3803e6ef0d2c5bf3b768d7044e7\") " pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.318514 kubelet[1995]: I0209 08:59:26.318501 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.318697 kubelet[1995]: I0209 08:59:26.318668 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/b9bac6a0ccc5042be8f5d46f6272401c-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" (UID: \"b9bac6a0ccc5042be8f5d46f6272401c\") " pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.318815 kubelet[1995]: I0209 08:59:26.318805 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/297285cbca3cdb75be03345ce23f82d0-kubeconfig\") pod \"kube-scheduler-ci-3510.3.2-5-6e53ffd492\" (UID: \"297285cbca3cdb75be03345ce23f82d0\") " pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:26.540736 kubelet[1995]: E0209 08:59:26.540693 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:26.546818 kubelet[1995]: E0209 08:59:26.546781 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:26.596331 kubelet[1995]: E0209 08:59:26.596299 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:26.670244 sudo[2008]: pam_unix(sudo:session): session closed for user root Feb 9 08:59:26.885430 kubelet[1995]: I0209 08:59:26.885262 1995 apiserver.go:52] "Watching apiserver" Feb 9 08:59:26.907973 kubelet[1995]: I0209 08:59:26.907917 1995 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Feb 9 08:59:26.922551 kubelet[1995]: I0209 08:59:26.922492 1995 reconciler.go:41] "Reconciler: start to sync state" Feb 9 08:59:27.315583 kubelet[1995]: E0209 08:59:27.315537 1995 kubelet.go:1802] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-3510.3.2-5-6e53ffd492\" already exists" pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:27.316451 kubelet[1995]: E0209 08:59:27.316423 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:27.499628 kubelet[1995]: E0209 08:59:27.499581 1995 kubelet.go:1802] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-3510.3.2-5-6e53ffd492\" already exists" pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:27.500464 kubelet[1995]: E0209 08:59:27.500443 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:27.692326 kubelet[1995]: E0209 08:59:27.692287 1995 kubelet.go:1802] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-3510.3.2-5-6e53ffd492\" already exists" pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" Feb 9 08:59:27.692662 kubelet[1995]: E0209 08:59:27.692637 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:28.145407 kubelet[1995]: E0209 08:59:28.145278 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:28.146333 kubelet[1995]: E0209 08:59:28.146312 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:28.147156 kubelet[1995]: E0209 08:59:28.146384 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:28.288224 kubelet[1995]: I0209 08:59:28.288186 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510.3.2-5-6e53ffd492" podStartSLOduration=4.287611085 pod.CreationTimestamp="2024-02-09 08:59:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 08:59:27.96669674 +0000 UTC m=+2.220921501" watchObservedRunningTime="2024-02-09 08:59:28.287611085 +0000 UTC m=+2.541835844" Feb 9 08:59:28.687113 kubelet[1995]: I0209 08:59:28.687077 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510.3.2-5-6e53ffd492" podStartSLOduration=2.687025113 pod.CreationTimestamp="2024-02-09 08:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 08:59:28.288562487 +0000 UTC m=+2.542787247" watchObservedRunningTime="2024-02-09 08:59:28.687025113 +0000 UTC m=+2.941249878" Feb 9 08:59:29.029020 sudo[1212]: pam_unix(sudo:session): session closed for user root Feb 9 08:59:29.035720 sshd[1209]: pam_unix(sshd:session): session closed for user core Feb 9 08:59:29.039851 systemd-logind[1095]: Session 5 logged out. Waiting for processes to exit. Feb 9 08:59:29.040000 systemd[1]: sshd@4-64.23.148.25:22-139.178.89.65:59024.service: Deactivated successfully. Feb 9 08:59:29.040960 systemd[1]: session-5.scope: Deactivated successfully. Feb 9 08:59:29.041133 systemd[1]: session-5.scope: Consumed 7.078s CPU time. Feb 9 08:59:29.042563 systemd-logind[1095]: Removed session 5. Feb 9 08:59:32.821441 kubelet[1995]: E0209 08:59:32.821392 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:32.846306 kubelet[1995]: I0209 08:59:32.846246 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510.3.2-5-6e53ffd492" podStartSLOduration=6.846182 pod.CreationTimestamp="2024-02-09 08:59:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 08:59:28.688185411 +0000 UTC m=+2.942410180" watchObservedRunningTime="2024-02-09 08:59:32.846182 +0000 UTC m=+7.100406769" Feb 9 08:59:33.112408 kubelet[1995]: E0209 08:59:33.112288 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:33.156394 kubelet[1995]: E0209 08:59:33.156351 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:33.156959 kubelet[1995]: E0209 08:59:33.156929 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:33.698138 update_engine[1096]: I0209 08:59:33.697824 1096 update_attempter.cc:509] Updating boot flags... Feb 9 08:59:35.004513 kubelet[1995]: E0209 08:59:35.004468 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:35.160917 kubelet[1995]: E0209 08:59:35.160420 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:36.163377 kubelet[1995]: E0209 08:59:36.163340 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:38.210408 kubelet[1995]: I0209 08:59:38.210371 1995 kuberuntime_manager.go:1114] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Feb 9 08:59:38.211490 env[1102]: time="2024-02-09T08:59:38.211414210Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Feb 9 08:59:38.211847 kubelet[1995]: I0209 08:59:38.211822 1995 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Feb 9 08:59:38.820568 kubelet[1995]: I0209 08:59:38.820518 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:38.824211 kubelet[1995]: I0209 08:59:38.824168 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:38.829506 systemd[1]: Created slice kubepods-besteffort-pod0038c420_36a5_4d8b_9432_8f8035dd8e59.slice. Feb 9 08:59:38.841435 systemd[1]: Created slice kubepods-burstable-podf0d9ccfc_d25a_418f_88fb_c69046864e01.slice. Feb 9 08:59:38.906496 kubelet[1995]: I0209 08:59:38.906441 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-etc-cni-netd\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.906899 kubelet[1995]: I0209 08:59:38.906877 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-bpf-maps\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907129 kubelet[1995]: I0209 08:59:38.907115 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-hostproc\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907291 kubelet[1995]: I0209 08:59:38.907280 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-net\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907483 kubelet[1995]: I0209 08:59:38.907471 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-xtables-lock\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907586 kubelet[1995]: I0209 08:59:38.907577 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-kernel\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907672 kubelet[1995]: I0209 08:59:38.907664 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/0038c420-36a5-4d8b-9432-8f8035dd8e59-xtables-lock\") pod \"kube-proxy-48r64\" (UID: \"0038c420-36a5-4d8b-9432-8f8035dd8e59\") " pod="kube-system/kube-proxy-48r64" Feb 9 08:59:38.907754 kubelet[1995]: I0209 08:59:38.907746 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-lib-modules\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907875 kubelet[1995]: I0209 08:59:38.907866 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqbgd\" (UniqueName: \"kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-kube-api-access-bqbgd\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.907978 kubelet[1995]: I0209 08:59:38.907970 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0038c420-36a5-4d8b-9432-8f8035dd8e59-lib-modules\") pod \"kube-proxy-48r64\" (UID: \"0038c420-36a5-4d8b-9432-8f8035dd8e59\") " pod="kube-system/kube-proxy-48r64" Feb 9 08:59:38.908083 kubelet[1995]: I0209 08:59:38.908074 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-hubble-tls\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.908184 kubelet[1995]: I0209 08:59:38.908176 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-cgroup\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.908269 kubelet[1995]: I0209 08:59:38.908261 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/0038c420-36a5-4d8b-9432-8f8035dd8e59-kube-proxy\") pod \"kube-proxy-48r64\" (UID: \"0038c420-36a5-4d8b-9432-8f8035dd8e59\") " pod="kube-system/kube-proxy-48r64" Feb 9 08:59:38.908377 kubelet[1995]: I0209 08:59:38.908368 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-config-path\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.908479 kubelet[1995]: I0209 08:59:38.908470 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhbxq\" (UniqueName: \"kubernetes.io/projected/0038c420-36a5-4d8b-9432-8f8035dd8e59-kube-api-access-rhbxq\") pod \"kube-proxy-48r64\" (UID: \"0038c420-36a5-4d8b-9432-8f8035dd8e59\") " pod="kube-system/kube-proxy-48r64" Feb 9 08:59:38.908572 kubelet[1995]: I0209 08:59:38.908564 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cni-path\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.908680 kubelet[1995]: I0209 08:59:38.908671 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f0d9ccfc-d25a-418f-88fb-c69046864e01-clustermesh-secrets\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:38.908789 kubelet[1995]: I0209 08:59:38.908780 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-run\") pod \"cilium-rlbhz\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " pod="kube-system/cilium-rlbhz" Feb 9 08:59:39.105717 kubelet[1995]: I0209 08:59:39.105573 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:39.115108 systemd[1]: Created slice kubepods-besteffort-poda9dc981a_ccb2_436a_8813_7b093c787c35.slice. Feb 9 08:59:39.139722 kubelet[1995]: E0209 08:59:39.139678 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:39.141687 env[1102]: time="2024-02-09T08:59:39.141078657Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-48r64,Uid:0038c420-36a5-4d8b-9432-8f8035dd8e59,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:39.144384 kubelet[1995]: E0209 08:59:39.144349 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:39.144914 env[1102]: time="2024-02-09T08:59:39.144867223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-rlbhz,Uid:f0d9ccfc-d25a-418f-88fb-c69046864e01,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:39.177570 env[1102]: time="2024-02-09T08:59:39.177469283Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 08:59:39.178690 env[1102]: time="2024-02-09T08:59:39.177797926Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 08:59:39.178690 env[1102]: time="2024-02-09T08:59:39.177846298Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 08:59:39.178690 env[1102]: time="2024-02-09T08:59:39.178115466Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/773ad6c40801c22ec999084412149db02c7f9473b15e119f33062bacf0a60a31 pid=2115 runtime=io.containerd.runc.v2 Feb 9 08:59:39.189280 env[1102]: time="2024-02-09T08:59:39.189165733Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 08:59:39.189597 env[1102]: time="2024-02-09T08:59:39.189297691Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 08:59:39.189597 env[1102]: time="2024-02-09T08:59:39.189337812Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 08:59:39.189667 env[1102]: time="2024-02-09T08:59:39.189567749Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d pid=2121 runtime=io.containerd.runc.v2 Feb 9 08:59:39.200002 systemd[1]: Started cri-containerd-773ad6c40801c22ec999084412149db02c7f9473b15e119f33062bacf0a60a31.scope. Feb 9 08:59:39.211378 kubelet[1995]: I0209 08:59:39.211336 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a9dc981a-ccb2-436a-8813-7b093c787c35-cilium-config-path\") pod \"cilium-operator-f59cbd8c6-62cjw\" (UID: \"a9dc981a-ccb2-436a-8813-7b093c787c35\") " pod="kube-system/cilium-operator-f59cbd8c6-62cjw" Feb 9 08:59:39.211799 kubelet[1995]: I0209 08:59:39.211399 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zzg\" (UniqueName: \"kubernetes.io/projected/a9dc981a-ccb2-436a-8813-7b093c787c35-kube-api-access-d9zzg\") pod \"cilium-operator-f59cbd8c6-62cjw\" (UID: \"a9dc981a-ccb2-436a-8813-7b093c787c35\") " pod="kube-system/cilium-operator-f59cbd8c6-62cjw" Feb 9 08:59:39.227215 systemd[1]: Started cri-containerd-fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d.scope. Feb 9 08:59:39.280121 env[1102]: time="2024-02-09T08:59:39.280076220Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-48r64,Uid:0038c420-36a5-4d8b-9432-8f8035dd8e59,Namespace:kube-system,Attempt:0,} returns sandbox id \"773ad6c40801c22ec999084412149db02c7f9473b15e119f33062bacf0a60a31\"" Feb 9 08:59:39.283305 kubelet[1995]: E0209 08:59:39.281517 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:39.289231 env[1102]: time="2024-02-09T08:59:39.288438398Z" level=info msg="CreateContainer within sandbox \"773ad6c40801c22ec999084412149db02c7f9473b15e119f33062bacf0a60a31\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Feb 9 08:59:39.316468 env[1102]: time="2024-02-09T08:59:39.316392908Z" level=info msg="CreateContainer within sandbox \"773ad6c40801c22ec999084412149db02c7f9473b15e119f33062bacf0a60a31\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"aedbbcc907d8d175dab3a41eafb318173f113c993b49e5b67a60565c4431568c\"" Feb 9 08:59:39.320970 env[1102]: time="2024-02-09T08:59:39.320871772Z" level=info msg="StartContainer for \"aedbbcc907d8d175dab3a41eafb318173f113c993b49e5b67a60565c4431568c\"" Feb 9 08:59:39.322274 env[1102]: time="2024-02-09T08:59:39.322184267Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-rlbhz,Uid:f0d9ccfc-d25a-418f-88fb-c69046864e01,Namespace:kube-system,Attempt:0,} returns sandbox id \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\"" Feb 9 08:59:39.323447 kubelet[1995]: E0209 08:59:39.323201 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:39.326373 env[1102]: time="2024-02-09T08:59:39.326320671Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Feb 9 08:59:39.359200 systemd[1]: Started cri-containerd-aedbbcc907d8d175dab3a41eafb318173f113c993b49e5b67a60565c4431568c.scope. Feb 9 08:59:39.422278 env[1102]: time="2024-02-09T08:59:39.422216094Z" level=info msg="StartContainer for \"aedbbcc907d8d175dab3a41eafb318173f113c993b49e5b67a60565c4431568c\" returns successfully" Feb 9 08:59:39.718999 kubelet[1995]: E0209 08:59:39.718954 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:39.720004 env[1102]: time="2024-02-09T08:59:39.719949684Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-f59cbd8c6-62cjw,Uid:a9dc981a-ccb2-436a-8813-7b093c787c35,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:39.741868 env[1102]: time="2024-02-09T08:59:39.741561337Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 08:59:39.741868 env[1102]: time="2024-02-09T08:59:39.741628906Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 08:59:39.741868 env[1102]: time="2024-02-09T08:59:39.741646740Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 08:59:39.742384 env[1102]: time="2024-02-09T08:59:39.742339898Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940 pid=2254 runtime=io.containerd.runc.v2 Feb 9 08:59:39.774023 systemd[1]: Started cri-containerd-ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940.scope. Feb 9 08:59:39.840914 env[1102]: time="2024-02-09T08:59:39.840856992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-f59cbd8c6-62cjw,Uid:a9dc981a-ccb2-436a-8813-7b093c787c35,Namespace:kube-system,Attempt:0,} returns sandbox id \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\"" Feb 9 08:59:39.842484 kubelet[1995]: E0209 08:59:39.842243 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:40.178670 kubelet[1995]: E0209 08:59:40.175635 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:40.230637 kubelet[1995]: I0209 08:59:40.230567 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-48r64" podStartSLOduration=2.230517364 pod.CreationTimestamp="2024-02-09 08:59:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 08:59:40.228997406 +0000 UTC m=+14.483222176" watchObservedRunningTime="2024-02-09 08:59:40.230517364 +0000 UTC m=+14.484742130" Feb 9 08:59:41.179676 kubelet[1995]: E0209 08:59:41.177676 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:45.270398 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3432591536.mount: Deactivated successfully. Feb 9 08:59:48.615076 env[1102]: time="2024-02-09T08:59:48.614981892Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:48.617579 env[1102]: time="2024-02-09T08:59:48.617534825Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:48.619213 env[1102]: time="2024-02-09T08:59:48.619171750Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:48.620311 env[1102]: time="2024-02-09T08:59:48.620264379Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Feb 9 08:59:48.624426 env[1102]: time="2024-02-09T08:59:48.624393508Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Feb 9 08:59:48.627036 env[1102]: time="2024-02-09T08:59:48.626977395Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 9 08:59:48.637733 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1443605220.mount: Deactivated successfully. Feb 9 08:59:48.644626 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2796093648.mount: Deactivated successfully. Feb 9 08:59:48.652627 env[1102]: time="2024-02-09T08:59:48.652524282Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\"" Feb 9 08:59:48.654205 env[1102]: time="2024-02-09T08:59:48.653698442Z" level=info msg="StartContainer for \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\"" Feb 9 08:59:48.699820 systemd[1]: Started cri-containerd-06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29.scope. Feb 9 08:59:48.751490 env[1102]: time="2024-02-09T08:59:48.751431547Z" level=info msg="StartContainer for \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\" returns successfully" Feb 9 08:59:48.758330 systemd[1]: cri-containerd-06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29.scope: Deactivated successfully. Feb 9 08:59:48.787276 env[1102]: time="2024-02-09T08:59:48.787213417Z" level=info msg="shim disconnected" id=06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29 Feb 9 08:59:48.787276 env[1102]: time="2024-02-09T08:59:48.787269914Z" level=warning msg="cleaning up after shim disconnected" id=06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29 namespace=k8s.io Feb 9 08:59:48.787276 env[1102]: time="2024-02-09T08:59:48.787284441Z" level=info msg="cleaning up dead shim" Feb 9 08:59:48.797670 env[1102]: time="2024-02-09T08:59:48.797483541Z" level=warning msg="cleanup warnings time=\"2024-02-09T08:59:48Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2422 runtime=io.containerd.runc.v2\n" Feb 9 08:59:49.194967 kubelet[1995]: E0209 08:59:49.194932 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:49.205065 env[1102]: time="2024-02-09T08:59:49.202580856Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Feb 9 08:59:49.218758 env[1102]: time="2024-02-09T08:59:49.218681434Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\"" Feb 9 08:59:49.222080 env[1102]: time="2024-02-09T08:59:49.221980639Z" level=info msg="StartContainer for \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\"" Feb 9 08:59:49.247958 systemd[1]: Started cri-containerd-61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d.scope. Feb 9 08:59:49.301128 env[1102]: time="2024-02-09T08:59:49.301037798Z" level=info msg="StartContainer for \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\" returns successfully" Feb 9 08:59:49.320523 systemd[1]: systemd-sysctl.service: Deactivated successfully. Feb 9 08:59:49.320919 systemd[1]: Stopped systemd-sysctl.service. Feb 9 08:59:49.321187 systemd[1]: Stopping systemd-sysctl.service... Feb 9 08:59:49.325455 systemd[1]: Starting systemd-sysctl.service... Feb 9 08:59:49.335740 systemd[1]: cri-containerd-61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d.scope: Deactivated successfully. Feb 9 08:59:49.352363 systemd[1]: Finished systemd-sysctl.service. Feb 9 08:59:49.368776 env[1102]: time="2024-02-09T08:59:49.368711313Z" level=info msg="shim disconnected" id=61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d Feb 9 08:59:49.368776 env[1102]: time="2024-02-09T08:59:49.368776603Z" level=warning msg="cleaning up after shim disconnected" id=61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d namespace=k8s.io Feb 9 08:59:49.369225 env[1102]: time="2024-02-09T08:59:49.368792743Z" level=info msg="cleaning up dead shim" Feb 9 08:59:49.381543 env[1102]: time="2024-02-09T08:59:49.381486803Z" level=warning msg="cleanup warnings time=\"2024-02-09T08:59:49Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2485 runtime=io.containerd.runc.v2\n" Feb 9 08:59:49.635567 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29-rootfs.mount: Deactivated successfully. Feb 9 08:59:49.883244 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3982870562.mount: Deactivated successfully. Feb 9 08:59:50.198385 kubelet[1995]: E0209 08:59:50.197936 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:50.214236 env[1102]: time="2024-02-09T08:59:50.214179397Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Feb 9 08:59:50.247742 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2523527491.mount: Deactivated successfully. Feb 9 08:59:50.251055 env[1102]: time="2024-02-09T08:59:50.250989310Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\"" Feb 9 08:59:50.251919 env[1102]: time="2024-02-09T08:59:50.251881532Z" level=info msg="StartContainer for \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\"" Feb 9 08:59:50.290776 systemd[1]: Started cri-containerd-b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4.scope. Feb 9 08:59:50.353823 systemd[1]: cri-containerd-b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4.scope: Deactivated successfully. Feb 9 08:59:50.357597 env[1102]: time="2024-02-09T08:59:50.356933501Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0d9ccfc_d25a_418f_88fb_c69046864e01.slice/cri-containerd-b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4.scope/memory.events\": no such file or directory" Feb 9 08:59:50.359329 env[1102]: time="2024-02-09T08:59:50.359278402Z" level=info msg="StartContainer for \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\" returns successfully" Feb 9 08:59:50.411286 env[1102]: time="2024-02-09T08:59:50.411218043Z" level=info msg="shim disconnected" id=b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4 Feb 9 08:59:50.411286 env[1102]: time="2024-02-09T08:59:50.411272404Z" level=warning msg="cleaning up after shim disconnected" id=b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4 namespace=k8s.io Feb 9 08:59:50.411286 env[1102]: time="2024-02-09T08:59:50.411282653Z" level=info msg="cleaning up dead shim" Feb 9 08:59:50.437191 env[1102]: time="2024-02-09T08:59:50.437123749Z" level=warning msg="cleanup warnings time=\"2024-02-09T08:59:50Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2543 runtime=io.containerd.runc.v2\n" Feb 9 08:59:50.877814 env[1102]: time="2024-02-09T08:59:50.877744811Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:50.878963 env[1102]: time="2024-02-09T08:59:50.878795371Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:50.880532 env[1102]: time="2024-02-09T08:59:50.880482447Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Feb 9 08:59:50.881429 env[1102]: time="2024-02-09T08:59:50.881378379Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Feb 9 08:59:50.885826 env[1102]: time="2024-02-09T08:59:50.885775747Z" level=info msg="CreateContainer within sandbox \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Feb 9 08:59:50.897196 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2611125958.mount: Deactivated successfully. Feb 9 08:59:50.903109 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4190269779.mount: Deactivated successfully. Feb 9 08:59:50.909924 env[1102]: time="2024-02-09T08:59:50.909868650Z" level=info msg="CreateContainer within sandbox \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\"" Feb 9 08:59:50.911224 env[1102]: time="2024-02-09T08:59:50.911157309Z" level=info msg="StartContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\"" Feb 9 08:59:50.939282 systemd[1]: Started cri-containerd-2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded.scope. Feb 9 08:59:51.001386 env[1102]: time="2024-02-09T08:59:51.001335454Z" level=info msg="StartContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" returns successfully" Feb 9 08:59:51.201784 kubelet[1995]: E0209 08:59:51.201735 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:51.205531 kubelet[1995]: E0209 08:59:51.205498 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:51.207776 env[1102]: time="2024-02-09T08:59:51.207733404Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Feb 9 08:59:51.219943 env[1102]: time="2024-02-09T08:59:51.219849570Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\"" Feb 9 08:59:51.220536 env[1102]: time="2024-02-09T08:59:51.220497005Z" level=info msg="StartContainer for \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\"" Feb 9 08:59:51.263483 systemd[1]: Started cri-containerd-57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9.scope. Feb 9 08:59:51.341900 env[1102]: time="2024-02-09T08:59:51.341791670Z" level=info msg="StartContainer for \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\" returns successfully" Feb 9 08:59:51.347556 systemd[1]: cri-containerd-57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9.scope: Deactivated successfully. Feb 9 08:59:51.393884 env[1102]: time="2024-02-09T08:59:51.393796727Z" level=info msg="shim disconnected" id=57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9 Feb 9 08:59:51.393884 env[1102]: time="2024-02-09T08:59:51.393878428Z" level=warning msg="cleaning up after shim disconnected" id=57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9 namespace=k8s.io Feb 9 08:59:51.394174 env[1102]: time="2024-02-09T08:59:51.393894063Z" level=info msg="cleaning up dead shim" Feb 9 08:59:51.410662 env[1102]: time="2024-02-09T08:59:51.410583547Z" level=warning msg="cleanup warnings time=\"2024-02-09T08:59:51Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2636 runtime=io.containerd.runc.v2\n" Feb 9 08:59:51.512596 kubelet[1995]: I0209 08:59:51.512462 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-f59cbd8c6-62cjw" podStartSLOduration=-9.223372024342388e+09 pod.CreationTimestamp="2024-02-09 08:59:39 +0000 UTC" firstStartedPulling="2024-02-09 08:59:39.843422019 +0000 UTC m=+14.097646780" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 08:59:51.383228166 +0000 UTC m=+25.637452934" watchObservedRunningTime="2024-02-09 08:59:51.51238755 +0000 UTC m=+25.766612319" Feb 9 08:59:52.209872 kubelet[1995]: E0209 08:59:52.209842 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:52.211034 kubelet[1995]: E0209 08:59:52.211011 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:52.217137 env[1102]: time="2024-02-09T08:59:52.215781783Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Feb 9 08:59:52.231172 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2148280752.mount: Deactivated successfully. Feb 9 08:59:52.237753 env[1102]: time="2024-02-09T08:59:52.237702534Z" level=info msg="CreateContainer within sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\"" Feb 9 08:59:52.239055 env[1102]: time="2024-02-09T08:59:52.239008839Z" level=info msg="StartContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\"" Feb 9 08:59:52.271712 systemd[1]: Started cri-containerd-6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034.scope. Feb 9 08:59:52.314748 env[1102]: time="2024-02-09T08:59:52.314688393Z" level=info msg="StartContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" returns successfully" Feb 9 08:59:52.463121 kubelet[1995]: I0209 08:59:52.462983 1995 kubelet_node_status.go:493] "Fast updating node status as it just became ready" Feb 9 08:59:52.605222 kubelet[1995]: I0209 08:59:52.605170 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:52.607153 kubelet[1995]: I0209 08:59:52.607122 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 08:59:52.612553 systemd[1]: Created slice kubepods-burstable-pod0533e78c_b54a_4fc9_b3ce_d3f5cbfefeef.slice. Feb 9 08:59:52.616741 systemd[1]: Created slice kubepods-burstable-pod73f46ed6_a3c1_4bb4_b08e_867472213cd1.slice. Feb 9 08:59:52.760751 kubelet[1995]: I0209 08:59:52.760616 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbwp2\" (UniqueName: \"kubernetes.io/projected/73f46ed6-a3c1-4bb4-b08e-867472213cd1-kube-api-access-zbwp2\") pod \"coredns-787d4945fb-ft7bf\" (UID: \"73f46ed6-a3c1-4bb4-b08e-867472213cd1\") " pod="kube-system/coredns-787d4945fb-ft7bf" Feb 9 08:59:52.761113 kubelet[1995]: I0209 08:59:52.761093 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0533e78c-b54a-4fc9-b3ce-d3f5cbfefeef-config-volume\") pod \"coredns-787d4945fb-c7vd8\" (UID: \"0533e78c-b54a-4fc9-b3ce-d3f5cbfefeef\") " pod="kube-system/coredns-787d4945fb-c7vd8" Feb 9 08:59:52.761255 kubelet[1995]: I0209 08:59:52.761243 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nr7v\" (UniqueName: \"kubernetes.io/projected/0533e78c-b54a-4fc9-b3ce-d3f5cbfefeef-kube-api-access-5nr7v\") pod \"coredns-787d4945fb-c7vd8\" (UID: \"0533e78c-b54a-4fc9-b3ce-d3f5cbfefeef\") " pod="kube-system/coredns-787d4945fb-c7vd8" Feb 9 08:59:52.761397 kubelet[1995]: I0209 08:59:52.761383 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/73f46ed6-a3c1-4bb4-b08e-867472213cd1-config-volume\") pod \"coredns-787d4945fb-ft7bf\" (UID: \"73f46ed6-a3c1-4bb4-b08e-867472213cd1\") " pod="kube-system/coredns-787d4945fb-ft7bf" Feb 9 08:59:52.916626 kubelet[1995]: E0209 08:59:52.916597 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:52.917663 env[1102]: time="2024-02-09T08:59:52.917614917Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-787d4945fb-c7vd8,Uid:0533e78c-b54a-4fc9-b3ce-d3f5cbfefeef,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:52.919485 kubelet[1995]: E0209 08:59:52.919461 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:52.920598 env[1102]: time="2024-02-09T08:59:52.920563624Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-787d4945fb-ft7bf,Uid:73f46ed6-a3c1-4bb4-b08e-867472213cd1,Namespace:kube-system,Attempt:0,}" Feb 9 08:59:53.215886 kubelet[1995]: E0209 08:59:53.215847 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:53.251483 kubelet[1995]: I0209 08:59:53.251444 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-rlbhz" podStartSLOduration=-9.223372021603392e+09 pod.CreationTimestamp="2024-02-09 08:59:38 +0000 UTC" firstStartedPulling="2024-02-09 08:59:39.324131065 +0000 UTC m=+13.578355827" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 08:59:53.248461861 +0000 UTC m=+27.502686635" watchObservedRunningTime="2024-02-09 08:59:53.251384899 +0000 UTC m=+27.505609668" Feb 9 08:59:54.217715 kubelet[1995]: E0209 08:59:54.217667 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:54.936927 systemd-networkd[997]: cilium_host: Link UP Feb 9 08:59:54.937124 systemd-networkd[997]: cilium_net: Link UP Feb 9 08:59:54.938140 systemd-networkd[997]: cilium_net: Gained carrier Feb 9 08:59:54.938512 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Feb 9 08:59:54.938573 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Feb 9 08:59:54.938732 systemd-networkd[997]: cilium_host: Gained carrier Feb 9 08:59:54.955174 systemd-networkd[997]: cilium_net: Gained IPv6LL Feb 9 08:59:55.122375 systemd-networkd[997]: cilium_vxlan: Link UP Feb 9 08:59:55.122387 systemd-networkd[997]: cilium_vxlan: Gained carrier Feb 9 08:59:55.220293 kubelet[1995]: E0209 08:59:55.220139 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:55.561113 kernel: NET: Registered PF_ALG protocol family Feb 9 08:59:55.879214 systemd-networkd[997]: cilium_host: Gained IPv6LL Feb 9 08:59:56.443218 systemd-networkd[997]: lxc_health: Link UP Feb 9 08:59:56.462245 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Feb 9 08:59:56.463365 systemd-networkd[997]: lxc_health: Gained carrier Feb 9 08:59:56.977267 systemd-networkd[997]: lxca6ccc8381934: Link UP Feb 9 08:59:56.985069 kernel: eth0: renamed from tmpf49fe Feb 9 08:59:56.997593 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxca6ccc8381934: link becomes ready Feb 9 08:59:56.995983 systemd-networkd[997]: lxca6ccc8381934: Gained carrier Feb 9 08:59:57.005615 systemd-networkd[997]: lxcab030fd31b57: Link UP Feb 9 08:59:57.013159 kernel: eth0: renamed from tmp96e0c Feb 9 08:59:57.018349 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcab030fd31b57: link becomes ready Feb 9 08:59:57.017793 systemd-networkd[997]: lxcab030fd31b57: Gained carrier Feb 9 08:59:57.095225 systemd-networkd[997]: cilium_vxlan: Gained IPv6LL Feb 9 08:59:57.161236 kubelet[1995]: E0209 08:59:57.161068 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:57.671253 systemd-networkd[997]: lxc_health: Gained IPv6LL Feb 9 08:59:58.311277 systemd-networkd[997]: lxcab030fd31b57: Gained IPv6LL Feb 9 08:59:58.504607 kubelet[1995]: I0209 08:59:58.504571 1995 prober_manager.go:287] "Failed to trigger a manual run" probe="Readiness" Feb 9 08:59:58.506633 kubelet[1995]: E0209 08:59:58.506610 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 08:59:58.568935 systemd-networkd[997]: lxca6ccc8381934: Gained IPv6LL Feb 9 08:59:59.227440 kubelet[1995]: E0209 08:59:59.227402 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:01.606342 env[1102]: time="2024-02-09T09:00:01.606180255Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:00:01.606342 env[1102]: time="2024-02-09T09:00:01.606271370Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:00:01.607306 env[1102]: time="2024-02-09T09:00:01.606318761Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:00:01.607787 env[1102]: time="2024-02-09T09:00:01.607702493Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/f49feaadefc718e7d9a5f4351a383ae9194176eb6855de936020f53da7a59b40 pid=3181 runtime=io.containerd.runc.v2 Feb 9 09:00:01.658776 systemd[1]: Started cri-containerd-f49feaadefc718e7d9a5f4351a383ae9194176eb6855de936020f53da7a59b40.scope. Feb 9 09:00:01.690163 systemd[1]: run-containerd-runc-k8s.io-f49feaadefc718e7d9a5f4351a383ae9194176eb6855de936020f53da7a59b40-runc.afNCTB.mount: Deactivated successfully. Feb 9 09:00:01.837224 env[1102]: time="2024-02-09T09:00:01.837149747Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-787d4945fb-ft7bf,Uid:73f46ed6-a3c1-4bb4-b08e-867472213cd1,Namespace:kube-system,Attempt:0,} returns sandbox id \"f49feaadefc718e7d9a5f4351a383ae9194176eb6855de936020f53da7a59b40\"" Feb 9 09:00:01.838299 kubelet[1995]: E0209 09:00:01.838270 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:01.844088 env[1102]: time="2024-02-09T09:00:01.844011806Z" level=info msg="CreateContainer within sandbox \"f49feaadefc718e7d9a5f4351a383ae9194176eb6855de936020f53da7a59b40\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 9 09:00:01.875963 env[1102]: time="2024-02-09T09:00:01.875767696Z" level=info msg="CreateContainer within sandbox \"f49feaadefc718e7d9a5f4351a383ae9194176eb6855de936020f53da7a59b40\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"e0f16d0b629f441d02ce95692913079939285aaaf0bac5ddec5fe4b785c1d326\"" Feb 9 09:00:01.877195 env[1102]: time="2024-02-09T09:00:01.877143057Z" level=info msg="StartContainer for \"e0f16d0b629f441d02ce95692913079939285aaaf0bac5ddec5fe4b785c1d326\"" Feb 9 09:00:01.905489 systemd[1]: Started cri-containerd-e0f16d0b629f441d02ce95692913079939285aaaf0bac5ddec5fe4b785c1d326.scope. Feb 9 09:00:01.973308 env[1102]: time="2024-02-09T09:00:01.973242408Z" level=info msg="StartContainer for \"e0f16d0b629f441d02ce95692913079939285aaaf0bac5ddec5fe4b785c1d326\" returns successfully" Feb 9 09:00:02.123887 env[1102]: time="2024-02-09T09:00:02.123342476Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:00:02.123887 env[1102]: time="2024-02-09T09:00:02.123414754Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:00:02.123887 env[1102]: time="2024-02-09T09:00:02.123452834Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:00:02.123887 env[1102]: time="2024-02-09T09:00:02.123729368Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/96e0c870b73a5376fe89c61602209df856fb03cb701c151a15c02c1f55324f16 pid=3259 runtime=io.containerd.runc.v2 Feb 9 09:00:02.163654 systemd[1]: Started cri-containerd-96e0c870b73a5376fe89c61602209df856fb03cb701c151a15c02c1f55324f16.scope. Feb 9 09:00:02.238751 kubelet[1995]: E0209 09:00:02.237893 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:02.279594 env[1102]: time="2024-02-09T09:00:02.279545960Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-787d4945fb-c7vd8,Uid:0533e78c-b54a-4fc9-b3ce-d3f5cbfefeef,Namespace:kube-system,Attempt:0,} returns sandbox id \"96e0c870b73a5376fe89c61602209df856fb03cb701c151a15c02c1f55324f16\"" Feb 9 09:00:02.281225 kubelet[1995]: E0209 09:00:02.281181 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:02.284393 env[1102]: time="2024-02-09T09:00:02.284334153Z" level=info msg="CreateContainer within sandbox \"96e0c870b73a5376fe89c61602209df856fb03cb701c151a15c02c1f55324f16\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Feb 9 09:00:02.326326 env[1102]: time="2024-02-09T09:00:02.326186976Z" level=info msg="CreateContainer within sandbox \"96e0c870b73a5376fe89c61602209df856fb03cb701c151a15c02c1f55324f16\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"95b63ce97898d5f73acd1615edd35e6286d8569041a857c53445e690b97c6b0d\"" Feb 9 09:00:02.327418 env[1102]: time="2024-02-09T09:00:02.327333884Z" level=info msg="StartContainer for \"95b63ce97898d5f73acd1615edd35e6286d8569041a857c53445e690b97c6b0d\"" Feb 9 09:00:02.385002 systemd[1]: Started cri-containerd-95b63ce97898d5f73acd1615edd35e6286d8569041a857c53445e690b97c6b0d.scope. Feb 9 09:00:02.460010 env[1102]: time="2024-02-09T09:00:02.459764257Z" level=info msg="StartContainer for \"95b63ce97898d5f73acd1615edd35e6286d8569041a857c53445e690b97c6b0d\" returns successfully" Feb 9 09:00:02.622340 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount327851161.mount: Deactivated successfully. Feb 9 09:00:02.944416 kubelet[1995]: I0209 09:00:02.944363 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-787d4945fb-ft7bf" podStartSLOduration=23.944281729 pod.CreationTimestamp="2024-02-09 08:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 09:00:02.291367579 +0000 UTC m=+36.545592351" watchObservedRunningTime="2024-02-09 09:00:02.944281729 +0000 UTC m=+37.198506501" Feb 9 09:00:03.244402 kubelet[1995]: E0209 09:00:03.243750 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:03.244402 kubelet[1995]: E0209 09:00:03.243752 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:03.308563 kubelet[1995]: I0209 09:00:03.308507 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-787d4945fb-c7vd8" podStartSLOduration=24.308418296 pod.CreationTimestamp="2024-02-09 08:59:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 09:00:03.27691489 +0000 UTC m=+37.531139665" watchObservedRunningTime="2024-02-09 09:00:03.308418296 +0000 UTC m=+37.562643065" Feb 9 09:00:04.246607 kubelet[1995]: E0209 09:00:04.246450 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:04.247371 kubelet[1995]: E0209 09:00:04.247348 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:05.248873 kubelet[1995]: E0209 09:00:05.248794 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:40.352091 systemd[1]: Started sshd@6-64.23.148.25:22-139.178.89.65:51772.service. Feb 9 09:00:40.428237 sshd[3424]: Accepted publickey for core from 139.178.89.65 port 51772 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:00:40.431181 sshd[3424]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:00:40.440183 systemd[1]: Started session-6.scope. Feb 9 09:00:40.440702 systemd-logind[1095]: New session 6 of user core. Feb 9 09:00:40.752154 sshd[3424]: pam_unix(sshd:session): session closed for user core Feb 9 09:00:40.757888 systemd[1]: sshd@6-64.23.148.25:22-139.178.89.65:51772.service: Deactivated successfully. Feb 9 09:00:40.758895 systemd[1]: session-6.scope: Deactivated successfully. Feb 9 09:00:40.759491 systemd-logind[1095]: Session 6 logged out. Waiting for processes to exit. Feb 9 09:00:40.760701 systemd-logind[1095]: Removed session 6. Feb 9 09:00:41.129112 kubelet[1995]: E0209 09:00:41.128434 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:43.128597 kubelet[1995]: E0209 09:00:43.128552 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:45.128695 kubelet[1995]: E0209 09:00:45.128656 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:45.760927 systemd[1]: Started sshd@7-64.23.148.25:22-139.178.89.65:51782.service. Feb 9 09:00:45.813243 sshd[3436]: Accepted publickey for core from 139.178.89.65 port 51782 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:00:45.815820 sshd[3436]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:00:45.822118 systemd-logind[1095]: New session 7 of user core. Feb 9 09:00:45.822602 systemd[1]: Started session-7.scope. Feb 9 09:00:45.981565 sshd[3436]: pam_unix(sshd:session): session closed for user core Feb 9 09:00:45.984608 systemd-logind[1095]: Session 7 logged out. Waiting for processes to exit. Feb 9 09:00:45.985010 systemd[1]: sshd@7-64.23.148.25:22-139.178.89.65:51782.service: Deactivated successfully. Feb 9 09:00:45.986086 systemd[1]: session-7.scope: Deactivated successfully. Feb 9 09:00:45.987255 systemd-logind[1095]: Removed session 7. Feb 9 09:00:50.129394 kubelet[1995]: E0209 09:00:50.129349 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:00:50.989539 systemd[1]: Started sshd@8-64.23.148.25:22-139.178.89.65:42324.service. Feb 9 09:00:51.044815 sshd[3449]: Accepted publickey for core from 139.178.89.65 port 42324 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:00:51.047840 sshd[3449]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:00:51.054961 systemd-logind[1095]: New session 8 of user core. Feb 9 09:00:51.055954 systemd[1]: Started session-8.scope. Feb 9 09:00:51.205257 sshd[3449]: pam_unix(sshd:session): session closed for user core Feb 9 09:00:51.209025 systemd[1]: sshd@8-64.23.148.25:22-139.178.89.65:42324.service: Deactivated successfully. Feb 9 09:00:51.209960 systemd[1]: session-8.scope: Deactivated successfully. Feb 9 09:00:51.210697 systemd-logind[1095]: Session 8 logged out. Waiting for processes to exit. Feb 9 09:00:51.211919 systemd-logind[1095]: Removed session 8. Feb 9 09:00:56.210543 systemd[1]: Started sshd@9-64.23.148.25:22-139.178.89.65:42328.service. Feb 9 09:00:56.261471 sshd[3462]: Accepted publickey for core from 139.178.89.65 port 42328 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:00:56.263554 sshd[3462]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:00:56.269121 systemd-logind[1095]: New session 9 of user core. Feb 9 09:00:56.269993 systemd[1]: Started session-9.scope. Feb 9 09:00:56.416264 sshd[3462]: pam_unix(sshd:session): session closed for user core Feb 9 09:00:56.420595 systemd[1]: sshd@9-64.23.148.25:22-139.178.89.65:42328.service: Deactivated successfully. Feb 9 09:00:56.421682 systemd[1]: session-9.scope: Deactivated successfully. Feb 9 09:00:56.425247 systemd-logind[1095]: Session 9 logged out. Waiting for processes to exit. Feb 9 09:00:56.430387 systemd-logind[1095]: Removed session 9. Feb 9 09:01:00.129420 kubelet[1995]: E0209 09:01:00.129379 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:01.423290 systemd[1]: Started sshd@10-64.23.148.25:22-139.178.89.65:51232.service. Feb 9 09:01:01.487011 sshd[3475]: Accepted publickey for core from 139.178.89.65 port 51232 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:01.489693 sshd[3475]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:01.497001 systemd[1]: Started session-10.scope. Feb 9 09:01:01.498378 systemd-logind[1095]: New session 10 of user core. Feb 9 09:01:01.685757 sshd[3475]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:01.694871 systemd[1]: Started sshd@11-64.23.148.25:22-139.178.89.65:51246.service. Feb 9 09:01:01.696290 systemd[1]: sshd@10-64.23.148.25:22-139.178.89.65:51232.service: Deactivated successfully. Feb 9 09:01:01.698697 systemd[1]: session-10.scope: Deactivated successfully. Feb 9 09:01:01.703315 systemd-logind[1095]: Session 10 logged out. Waiting for processes to exit. Feb 9 09:01:01.705468 systemd-logind[1095]: Removed session 10. Feb 9 09:01:01.763788 sshd[3486]: Accepted publickey for core from 139.178.89.65 port 51246 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:01.766714 sshd[3486]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:01.775324 systemd[1]: Started session-11.scope. Feb 9 09:01:01.776293 systemd-logind[1095]: New session 11 of user core. Feb 9 09:01:03.265234 sshd[3486]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:03.276837 systemd[1]: Started sshd@12-64.23.148.25:22-139.178.89.65:51256.service. Feb 9 09:01:03.287470 systemd[1]: sshd@11-64.23.148.25:22-139.178.89.65:51246.service: Deactivated successfully. Feb 9 09:01:03.288745 systemd[1]: session-11.scope: Deactivated successfully. Feb 9 09:01:03.293449 systemd-logind[1095]: Session 11 logged out. Waiting for processes to exit. Feb 9 09:01:03.296266 systemd-logind[1095]: Removed session 11. Feb 9 09:01:03.357738 sshd[3497]: Accepted publickey for core from 139.178.89.65 port 51256 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:03.361108 sshd[3497]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:03.371377 systemd-logind[1095]: New session 12 of user core. Feb 9 09:01:03.372531 systemd[1]: Started session-12.scope. Feb 9 09:01:03.587483 sshd[3497]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:03.593681 systemd[1]: sshd@12-64.23.148.25:22-139.178.89.65:51256.service: Deactivated successfully. Feb 9 09:01:03.594865 systemd[1]: session-12.scope: Deactivated successfully. Feb 9 09:01:03.595909 systemd-logind[1095]: Session 12 logged out. Waiting for processes to exit. Feb 9 09:01:03.597469 systemd-logind[1095]: Removed session 12. Feb 9 09:01:08.595645 systemd[1]: Started sshd@13-64.23.148.25:22-139.178.89.65:42638.service. Feb 9 09:01:08.647156 sshd[3512]: Accepted publickey for core from 139.178.89.65 port 42638 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:08.649642 sshd[3512]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:08.655268 systemd-logind[1095]: New session 13 of user core. Feb 9 09:01:08.655917 systemd[1]: Started session-13.scope. Feb 9 09:01:08.793593 sshd[3512]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:08.797336 systemd[1]: sshd@13-64.23.148.25:22-139.178.89.65:42638.service: Deactivated successfully. Feb 9 09:01:08.798251 systemd[1]: session-13.scope: Deactivated successfully. Feb 9 09:01:08.799375 systemd-logind[1095]: Session 13 logged out. Waiting for processes to exit. Feb 9 09:01:08.800983 systemd-logind[1095]: Removed session 13. Feb 9 09:01:13.128881 kubelet[1995]: E0209 09:01:13.128841 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:13.806757 systemd[1]: Started sshd@14-64.23.148.25:22-139.178.89.65:42650.service. Feb 9 09:01:13.864916 sshd[3526]: Accepted publickey for core from 139.178.89.65 port 42650 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:13.866857 sshd[3526]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:13.873657 systemd-logind[1095]: New session 14 of user core. Feb 9 09:01:13.873752 systemd[1]: Started session-14.scope. Feb 9 09:01:14.031572 sshd[3526]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:14.036180 systemd-logind[1095]: Session 14 logged out. Waiting for processes to exit. Feb 9 09:01:14.036350 systemd[1]: sshd@14-64.23.148.25:22-139.178.89.65:42650.service: Deactivated successfully. Feb 9 09:01:14.037421 systemd[1]: session-14.scope: Deactivated successfully. Feb 9 09:01:14.038737 systemd-logind[1095]: Removed session 14. Feb 9 09:01:19.040175 systemd[1]: Started sshd@15-64.23.148.25:22-139.178.89.65:50132.service. Feb 9 09:01:19.095096 sshd[3537]: Accepted publickey for core from 139.178.89.65 port 50132 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:19.096469 sshd[3537]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:19.102334 systemd-logind[1095]: New session 15 of user core. Feb 9 09:01:19.103648 systemd[1]: Started session-15.scope. Feb 9 09:01:19.130079 kubelet[1995]: E0209 09:01:19.129760 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:19.266399 sshd[3537]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:19.273420 systemd[1]: Started sshd@16-64.23.148.25:22-139.178.89.65:50138.service. Feb 9 09:01:19.274340 systemd[1]: sshd@15-64.23.148.25:22-139.178.89.65:50132.service: Deactivated successfully. Feb 9 09:01:19.275541 systemd[1]: session-15.scope: Deactivated successfully. Feb 9 09:01:19.281309 systemd-logind[1095]: Session 15 logged out. Waiting for processes to exit. Feb 9 09:01:19.282824 systemd-logind[1095]: Removed session 15. Feb 9 09:01:19.331892 sshd[3548]: Accepted publickey for core from 139.178.89.65 port 50138 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:19.334811 sshd[3548]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:19.342614 systemd[1]: Started session-16.scope. Feb 9 09:01:19.343564 systemd-logind[1095]: New session 16 of user core. Feb 9 09:01:19.704705 sshd[3548]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:19.711500 systemd[1]: sshd@16-64.23.148.25:22-139.178.89.65:50138.service: Deactivated successfully. Feb 9 09:01:19.713314 systemd[1]: session-16.scope: Deactivated successfully. Feb 9 09:01:19.714580 systemd-logind[1095]: Session 16 logged out. Waiting for processes to exit. Feb 9 09:01:19.718542 systemd[1]: Started sshd@17-64.23.148.25:22-139.178.89.65:50140.service. Feb 9 09:01:19.723964 systemd-logind[1095]: Removed session 16. Feb 9 09:01:19.781998 sshd[3558]: Accepted publickey for core from 139.178.89.65 port 50140 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:19.783991 sshd[3558]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:19.790130 systemd-logind[1095]: New session 17 of user core. Feb 9 09:01:19.791231 systemd[1]: Started session-17.scope. Feb 9 09:01:21.282860 sshd[3558]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:21.292384 systemd[1]: Started sshd@18-64.23.148.25:22-139.178.89.65:50154.service. Feb 9 09:01:21.295578 systemd[1]: sshd@17-64.23.148.25:22-139.178.89.65:50140.service: Deactivated successfully. Feb 9 09:01:21.296785 systemd[1]: session-17.scope: Deactivated successfully. Feb 9 09:01:21.299036 systemd-logind[1095]: Session 17 logged out. Waiting for processes to exit. Feb 9 09:01:21.302836 systemd-logind[1095]: Removed session 17. Feb 9 09:01:21.366310 sshd[3581]: Accepted publickey for core from 139.178.89.65 port 50154 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:21.367889 sshd[3581]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:21.373514 systemd-logind[1095]: New session 18 of user core. Feb 9 09:01:21.374149 systemd[1]: Started session-18.scope. Feb 9 09:01:21.785197 sshd[3581]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:21.791415 systemd[1]: sshd@18-64.23.148.25:22-139.178.89.65:50154.service: Deactivated successfully. Feb 9 09:01:21.792452 systemd[1]: session-18.scope: Deactivated successfully. Feb 9 09:01:21.793994 systemd-logind[1095]: Session 18 logged out. Waiting for processes to exit. Feb 9 09:01:21.795887 systemd[1]: Started sshd@19-64.23.148.25:22-139.178.89.65:50162.service. Feb 9 09:01:21.798432 systemd-logind[1095]: Removed session 18. Feb 9 09:01:21.855782 sshd[3635]: Accepted publickey for core from 139.178.89.65 port 50162 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:21.856613 sshd[3635]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:21.862602 systemd[1]: Started session-19.scope. Feb 9 09:01:21.863171 systemd-logind[1095]: New session 19 of user core. Feb 9 09:01:22.017885 sshd[3635]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:22.021824 systemd[1]: sshd@19-64.23.148.25:22-139.178.89.65:50162.service: Deactivated successfully. Feb 9 09:01:22.022690 systemd[1]: session-19.scope: Deactivated successfully. Feb 9 09:01:22.024310 systemd-logind[1095]: Session 19 logged out. Waiting for processes to exit. Feb 9 09:01:22.026307 systemd-logind[1095]: Removed session 19. Feb 9 09:01:25.128432 kubelet[1995]: E0209 09:01:25.128393 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:27.025836 systemd[1]: Started sshd@20-64.23.148.25:22-139.178.89.65:50176.service. Feb 9 09:01:27.075074 sshd[3648]: Accepted publickey for core from 139.178.89.65 port 50176 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:27.077905 sshd[3648]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:27.085155 systemd-logind[1095]: New session 20 of user core. Feb 9 09:01:27.085216 systemd[1]: Started session-20.scope. Feb 9 09:01:27.248177 sshd[3648]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:27.251962 systemd-logind[1095]: Session 20 logged out. Waiting for processes to exit. Feb 9 09:01:27.252297 systemd[1]: sshd@20-64.23.148.25:22-139.178.89.65:50176.service: Deactivated successfully. Feb 9 09:01:27.253365 systemd[1]: session-20.scope: Deactivated successfully. Feb 9 09:01:27.254642 systemd-logind[1095]: Removed session 20. Feb 9 09:01:32.254601 systemd[1]: Started sshd@21-64.23.148.25:22-139.178.89.65:53740.service. Feb 9 09:01:32.302553 sshd[3687]: Accepted publickey for core from 139.178.89.65 port 53740 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:32.304413 sshd[3687]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:32.309392 systemd[1]: Started session-21.scope. Feb 9 09:01:32.310845 systemd-logind[1095]: New session 21 of user core. Feb 9 09:01:32.471176 sshd[3687]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:32.474910 systemd[1]: sshd@21-64.23.148.25:22-139.178.89.65:53740.service: Deactivated successfully. Feb 9 09:01:32.475978 systemd[1]: session-21.scope: Deactivated successfully. Feb 9 09:01:32.477316 systemd-logind[1095]: Session 21 logged out. Waiting for processes to exit. Feb 9 09:01:32.478528 systemd-logind[1095]: Removed session 21. Feb 9 09:01:37.479912 systemd[1]: Started sshd@22-64.23.148.25:22-139.178.89.65:53756.service. Feb 9 09:01:37.532180 sshd[3699]: Accepted publickey for core from 139.178.89.65 port 53756 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:37.533291 sshd[3699]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:37.539091 systemd-logind[1095]: New session 22 of user core. Feb 9 09:01:37.539842 systemd[1]: Started session-22.scope. Feb 9 09:01:37.677672 sshd[3699]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:37.681621 systemd[1]: sshd@22-64.23.148.25:22-139.178.89.65:53756.service: Deactivated successfully. Feb 9 09:01:37.682477 systemd[1]: session-22.scope: Deactivated successfully. Feb 9 09:01:37.683296 systemd-logind[1095]: Session 22 logged out. Waiting for processes to exit. Feb 9 09:01:37.684410 systemd-logind[1095]: Removed session 22. Feb 9 09:01:42.686654 systemd[1]: Started sshd@23-64.23.148.25:22-139.178.89.65:52166.service. Feb 9 09:01:42.736251 sshd[3712]: Accepted publickey for core from 139.178.89.65 port 52166 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:42.738685 sshd[3712]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:42.746960 systemd[1]: Started session-23.scope. Feb 9 09:01:42.748447 systemd-logind[1095]: New session 23 of user core. Feb 9 09:01:42.898328 sshd[3712]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:42.901704 systemd[1]: sshd@23-64.23.148.25:22-139.178.89.65:52166.service: Deactivated successfully. Feb 9 09:01:42.902882 systemd[1]: session-23.scope: Deactivated successfully. Feb 9 09:01:42.903784 systemd-logind[1095]: Session 23 logged out. Waiting for processes to exit. Feb 9 09:01:42.904884 systemd-logind[1095]: Removed session 23. Feb 9 09:01:44.129259 kubelet[1995]: E0209 09:01:44.129216 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:47.907032 systemd[1]: Started sshd@24-64.23.148.25:22-139.178.89.65:52180.service. Feb 9 09:01:47.965229 sshd[3724]: Accepted publickey for core from 139.178.89.65 port 52180 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:47.966756 sshd[3724]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:47.973741 systemd[1]: Started session-24.scope. Feb 9 09:01:47.974558 systemd-logind[1095]: New session 24 of user core. Feb 9 09:01:48.134444 sshd[3724]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:48.138368 systemd[1]: sshd@24-64.23.148.25:22-139.178.89.65:52180.service: Deactivated successfully. Feb 9 09:01:48.139402 systemd[1]: session-24.scope: Deactivated successfully. Feb 9 09:01:48.141114 systemd-logind[1095]: Session 24 logged out. Waiting for processes to exit. Feb 9 09:01:48.142937 systemd-logind[1095]: Removed session 24. Feb 9 09:01:50.129271 kubelet[1995]: E0209 09:01:50.129226 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:53.146080 systemd[1]: Started sshd@25-64.23.148.25:22-139.178.89.65:58004.service. Feb 9 09:01:53.202969 sshd[3736]: Accepted publickey for core from 139.178.89.65 port 58004 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:53.205584 sshd[3736]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:53.215400 systemd[1]: Started session-25.scope. Feb 9 09:01:53.218105 systemd-logind[1095]: New session 25 of user core. Feb 9 09:01:53.359640 sshd[3736]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:53.365739 systemd[1]: Started sshd@26-64.23.148.25:22-139.178.89.65:58016.service. Feb 9 09:01:53.371024 systemd-logind[1095]: Session 25 logged out. Waiting for processes to exit. Feb 9 09:01:53.372853 systemd[1]: sshd@25-64.23.148.25:22-139.178.89.65:58004.service: Deactivated successfully. Feb 9 09:01:53.373751 systemd[1]: session-25.scope: Deactivated successfully. Feb 9 09:01:53.375256 systemd-logind[1095]: Removed session 25. Feb 9 09:01:53.421668 sshd[3748]: Accepted publickey for core from 139.178.89.65 port 58016 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:53.423984 sshd[3748]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:53.430296 systemd[1]: Started session-26.scope. Feb 9 09:01:53.431105 systemd-logind[1095]: New session 26 of user core. Feb 9 09:01:57.247929 systemd[1]: run-containerd-runc-k8s.io-6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034-runc.Bnu5DC.mount: Deactivated successfully. Feb 9 09:01:57.287946 env[1102]: time="2024-02-09T09:01:57.287882489Z" level=info msg="StopContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" with timeout 30 (s)" Feb 9 09:01:57.288879 env[1102]: time="2024-02-09T09:01:57.288839404Z" level=info msg="Stop container \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" with signal terminated" Feb 9 09:01:57.296009 env[1102]: time="2024-02-09T09:01:57.295926369Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Feb 9 09:01:57.304677 env[1102]: time="2024-02-09T09:01:57.304636119Z" level=info msg="StopContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" with timeout 1 (s)" Feb 9 09:01:57.305094 systemd[1]: cri-containerd-2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded.scope: Deactivated successfully. Feb 9 09:01:57.307383 env[1102]: time="2024-02-09T09:01:57.307016203Z" level=info msg="Stop container \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" with signal terminated" Feb 9 09:01:57.334464 systemd-networkd[997]: lxc_health: Link DOWN Feb 9 09:01:57.334476 systemd-networkd[997]: lxc_health: Lost carrier Feb 9 09:01:57.362341 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded-rootfs.mount: Deactivated successfully. Feb 9 09:01:57.370085 env[1102]: time="2024-02-09T09:01:57.369896426Z" level=info msg="shim disconnected" id=2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded Feb 9 09:01:57.370085 env[1102]: time="2024-02-09T09:01:57.369958985Z" level=warning msg="cleaning up after shim disconnected" id=2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded namespace=k8s.io Feb 9 09:01:57.370085 env[1102]: time="2024-02-09T09:01:57.369971584Z" level=info msg="cleaning up dead shim" Feb 9 09:01:57.378398 systemd[1]: cri-containerd-6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034.scope: Deactivated successfully. Feb 9 09:01:57.378715 systemd[1]: cri-containerd-6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034.scope: Consumed 9.336s CPU time. Feb 9 09:01:57.386568 env[1102]: time="2024-02-09T09:01:57.386516942Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:01:57Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3805 runtime=io.containerd.runc.v2\n" Feb 9 09:01:57.389324 env[1102]: time="2024-02-09T09:01:57.389275428Z" level=info msg="StopContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" returns successfully" Feb 9 09:01:57.390416 env[1102]: time="2024-02-09T09:01:57.390369992Z" level=info msg="StopPodSandbox for \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\"" Feb 9 09:01:57.397022 env[1102]: time="2024-02-09T09:01:57.390460919Z" level=info msg="Container to stop \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:01:57.393055 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940-shm.mount: Deactivated successfully. Feb 9 09:01:57.407908 systemd[1]: cri-containerd-ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940.scope: Deactivated successfully. Feb 9 09:01:57.426661 env[1102]: time="2024-02-09T09:01:57.426601688Z" level=info msg="shim disconnected" id=6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034 Feb 9 09:01:57.427056 env[1102]: time="2024-02-09T09:01:57.427012440Z" level=warning msg="cleaning up after shim disconnected" id=6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034 namespace=k8s.io Feb 9 09:01:57.427178 env[1102]: time="2024-02-09T09:01:57.427158239Z" level=info msg="cleaning up dead shim" Feb 9 09:01:57.440298 env[1102]: time="2024-02-09T09:01:57.440232020Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:01:57Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3843 runtime=io.containerd.runc.v2\n" Feb 9 09:01:57.443869 env[1102]: time="2024-02-09T09:01:57.443820562Z" level=info msg="StopContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" returns successfully" Feb 9 09:01:57.448475 env[1102]: time="2024-02-09T09:01:57.445953722Z" level=info msg="StopPodSandbox for \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\"" Feb 9 09:01:57.448475 env[1102]: time="2024-02-09T09:01:57.446126242Z" level=info msg="Container to stop \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:01:57.448475 env[1102]: time="2024-02-09T09:01:57.446182004Z" level=info msg="Container to stop \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:01:57.448475 env[1102]: time="2024-02-09T09:01:57.446194173Z" level=info msg="Container to stop \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:01:57.448475 env[1102]: time="2024-02-09T09:01:57.446260000Z" level=info msg="Container to stop \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:01:57.448475 env[1102]: time="2024-02-09T09:01:57.446271819Z" level=info msg="Container to stop \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:01:57.450912 env[1102]: time="2024-02-09T09:01:57.450859497Z" level=info msg="shim disconnected" id=ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940 Feb 9 09:01:57.450912 env[1102]: time="2024-02-09T09:01:57.450912823Z" level=warning msg="cleaning up after shim disconnected" id=ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940 namespace=k8s.io Feb 9 09:01:57.450912 env[1102]: time="2024-02-09T09:01:57.450922864Z" level=info msg="cleaning up dead shim" Feb 9 09:01:57.458592 systemd[1]: cri-containerd-fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d.scope: Deactivated successfully. Feb 9 09:01:57.480626 env[1102]: time="2024-02-09T09:01:57.480569410Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:01:57Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3863 runtime=io.containerd.runc.v2\n" Feb 9 09:01:57.481331 env[1102]: time="2024-02-09T09:01:57.481293072Z" level=info msg="TearDown network for sandbox \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\" successfully" Feb 9 09:01:57.481531 env[1102]: time="2024-02-09T09:01:57.481506743Z" level=info msg="StopPodSandbox for \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\" returns successfully" Feb 9 09:01:57.503975 env[1102]: time="2024-02-09T09:01:57.502306234Z" level=info msg="shim disconnected" id=fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d Feb 9 09:01:57.503975 env[1102]: time="2024-02-09T09:01:57.502360694Z" level=warning msg="cleaning up after shim disconnected" id=fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d namespace=k8s.io Feb 9 09:01:57.503975 env[1102]: time="2024-02-09T09:01:57.502371205Z" level=info msg="cleaning up dead shim" Feb 9 09:01:57.517125 env[1102]: time="2024-02-09T09:01:57.517069018Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:01:57Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3892 runtime=io.containerd.runc.v2\n" Feb 9 09:01:57.517442 env[1102]: time="2024-02-09T09:01:57.517407772Z" level=info msg="TearDown network for sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" successfully" Feb 9 09:01:57.517442 env[1102]: time="2024-02-09T09:01:57.517435757Z" level=info msg="StopPodSandbox for \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" returns successfully" Feb 9 09:01:57.530742 kubelet[1995]: I0209 09:01:57.530700 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a9dc981a-ccb2-436a-8813-7b093c787c35-cilium-config-path\") pod \"a9dc981a-ccb2-436a-8813-7b093c787c35\" (UID: \"a9dc981a-ccb2-436a-8813-7b093c787c35\") " Feb 9 09:01:57.531321 kubelet[1995]: I0209 09:01:57.531301 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9zzg\" (UniqueName: \"kubernetes.io/projected/a9dc981a-ccb2-436a-8813-7b093c787c35-kube-api-access-d9zzg\") pod \"a9dc981a-ccb2-436a-8813-7b093c787c35\" (UID: \"a9dc981a-ccb2-436a-8813-7b093c787c35\") " Feb 9 09:01:57.531510 kubelet[1995]: W0209 09:01:57.531458 1995 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/a9dc981a-ccb2-436a-8813-7b093c787c35/volumes/kubernetes.io~configmap/cilium-config-path: clearQuota called, but quotas disabled Feb 9 09:01:57.534201 kubelet[1995]: I0209 09:01:57.533623 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9dc981a-ccb2-436a-8813-7b093c787c35-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "a9dc981a-ccb2-436a-8813-7b093c787c35" (UID: "a9dc981a-ccb2-436a-8813-7b093c787c35"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 9 09:01:57.538076 kubelet[1995]: I0209 09:01:57.537990 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9dc981a-ccb2-436a-8813-7b093c787c35-kube-api-access-d9zzg" (OuterVolumeSpecName: "kube-api-access-d9zzg") pod "a9dc981a-ccb2-436a-8813-7b093c787c35" (UID: "a9dc981a-ccb2-436a-8813-7b093c787c35"). InnerVolumeSpecName "kube-api-access-d9zzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:01:57.545872 kubelet[1995]: I0209 09:01:57.545823 1995 scope.go:115] "RemoveContainer" containerID="2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded" Feb 9 09:01:57.552584 env[1102]: time="2024-02-09T09:01:57.552535836Z" level=info msg="RemoveContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\"" Feb 9 09:01:57.555307 systemd[1]: Removed slice kubepods-besteffort-poda9dc981a_ccb2_436a_8813_7b093c787c35.slice. Feb 9 09:01:57.560084 env[1102]: time="2024-02-09T09:01:57.559996019Z" level=info msg="RemoveContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" returns successfully" Feb 9 09:01:57.560994 kubelet[1995]: I0209 09:01:57.560956 1995 scope.go:115] "RemoveContainer" containerID="2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded" Feb 9 09:01:57.562360 env[1102]: time="2024-02-09T09:01:57.562285013Z" level=error msg="ContainerStatus for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\": not found" Feb 9 09:01:57.563083 kubelet[1995]: E0209 09:01:57.562925 1995 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\": not found" containerID="2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded" Feb 9 09:01:57.563685 kubelet[1995]: I0209 09:01:57.563654 1995 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded} err="failed to get container status \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\": rpc error: code = NotFound desc = an error occurred when try to find container \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\": not found" Feb 9 09:01:57.565875 kubelet[1995]: I0209 09:01:57.565858 1995 scope.go:115] "RemoveContainer" containerID="6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034" Feb 9 09:01:57.567099 env[1102]: time="2024-02-09T09:01:57.567015152Z" level=info msg="RemoveContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\"" Feb 9 09:01:57.569902 env[1102]: time="2024-02-09T09:01:57.569869861Z" level=info msg="RemoveContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" returns successfully" Feb 9 09:01:57.570252 kubelet[1995]: I0209 09:01:57.570234 1995 scope.go:115] "RemoveContainer" containerID="57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9" Feb 9 09:01:57.571792 env[1102]: time="2024-02-09T09:01:57.571754328Z" level=info msg="RemoveContainer for \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\"" Feb 9 09:01:57.574392 env[1102]: time="2024-02-09T09:01:57.574350581Z" level=info msg="RemoveContainer for \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\" returns successfully" Feb 9 09:01:57.574762 kubelet[1995]: I0209 09:01:57.574729 1995 scope.go:115] "RemoveContainer" containerID="b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4" Feb 9 09:01:57.576317 env[1102]: time="2024-02-09T09:01:57.576233025Z" level=info msg="RemoveContainer for \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\"" Feb 9 09:01:57.578276 env[1102]: time="2024-02-09T09:01:57.578246181Z" level=info msg="RemoveContainer for \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\" returns successfully" Feb 9 09:01:57.578426 kubelet[1995]: I0209 09:01:57.578406 1995 scope.go:115] "RemoveContainer" containerID="61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d" Feb 9 09:01:57.579625 env[1102]: time="2024-02-09T09:01:57.579595372Z" level=info msg="RemoveContainer for \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\"" Feb 9 09:01:57.583185 env[1102]: time="2024-02-09T09:01:57.583149283Z" level=info msg="RemoveContainer for \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\" returns successfully" Feb 9 09:01:57.583379 kubelet[1995]: I0209 09:01:57.583354 1995 scope.go:115] "RemoveContainer" containerID="06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29" Feb 9 09:01:57.585844 env[1102]: time="2024-02-09T09:01:57.585814256Z" level=info msg="RemoveContainer for \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\"" Feb 9 09:01:57.590390 env[1102]: time="2024-02-09T09:01:57.590354832Z" level=info msg="RemoveContainer for \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\" returns successfully" Feb 9 09:01:57.590880 kubelet[1995]: I0209 09:01:57.590848 1995 scope.go:115] "RemoveContainer" containerID="6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034" Feb 9 09:01:57.591180 env[1102]: time="2024-02-09T09:01:57.591118777Z" level=error msg="ContainerStatus for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\": not found" Feb 9 09:01:57.591329 kubelet[1995]: E0209 09:01:57.591308 1995 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\": not found" containerID="6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034" Feb 9 09:01:57.591425 kubelet[1995]: I0209 09:01:57.591348 1995 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034} err="failed to get container status \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\": rpc error: code = NotFound desc = an error occurred when try to find container \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\": not found" Feb 9 09:01:57.591425 kubelet[1995]: I0209 09:01:57.591361 1995 scope.go:115] "RemoveContainer" containerID="57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9" Feb 9 09:01:57.591892 env[1102]: time="2024-02-09T09:01:57.591722315Z" level=error msg="ContainerStatus for \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\": not found" Feb 9 09:01:57.592361 kubelet[1995]: E0209 09:01:57.592340 1995 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\": not found" containerID="57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9" Feb 9 09:01:57.592478 kubelet[1995]: I0209 09:01:57.592371 1995 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9} err="failed to get container status \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\": rpc error: code = NotFound desc = an error occurred when try to find container \"57b02b91e6855de7415dda38998cc43fcf693355464dbf9ea4db3f23ba08cdc9\": not found" Feb 9 09:01:57.592478 kubelet[1995]: I0209 09:01:57.592383 1995 scope.go:115] "RemoveContainer" containerID="b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4" Feb 9 09:01:57.592639 env[1102]: time="2024-02-09T09:01:57.592574005Z" level=error msg="ContainerStatus for \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\": not found" Feb 9 09:01:57.592745 kubelet[1995]: E0209 09:01:57.592728 1995 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\": not found" containerID="b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4" Feb 9 09:01:57.592821 kubelet[1995]: I0209 09:01:57.592757 1995 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4} err="failed to get container status \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\": rpc error: code = NotFound desc = an error occurred when try to find container \"b789501b4713fcc8ac2a4ff541283a192446ccf84a79c588aed494e5c7420cd4\": not found" Feb 9 09:01:57.592821 kubelet[1995]: I0209 09:01:57.592768 1995 scope.go:115] "RemoveContainer" containerID="61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d" Feb 9 09:01:57.592973 env[1102]: time="2024-02-09T09:01:57.592930570Z" level=error msg="ContainerStatus for \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\": not found" Feb 9 09:01:57.593088 kubelet[1995]: E0209 09:01:57.593080 1995 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\": not found" containerID="61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d" Feb 9 09:01:57.593159 kubelet[1995]: I0209 09:01:57.593102 1995 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d} err="failed to get container status \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\": rpc error: code = NotFound desc = an error occurred when try to find container \"61dcc87313d2398e4273bf5a6c373696b719dc68cc96106b4a4071dab859aa6d\": not found" Feb 9 09:01:57.593159 kubelet[1995]: I0209 09:01:57.593112 1995 scope.go:115] "RemoveContainer" containerID="06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29" Feb 9 09:01:57.593301 env[1102]: time="2024-02-09T09:01:57.593258728Z" level=error msg="ContainerStatus for \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\": not found" Feb 9 09:01:57.593565 kubelet[1995]: E0209 09:01:57.593538 1995 remote_runtime.go:415] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\": not found" containerID="06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29" Feb 9 09:01:57.593674 kubelet[1995]: I0209 09:01:57.593574 1995 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={Type:containerd ID:06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29} err="failed to get container status \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\": rpc error: code = NotFound desc = an error occurred when try to find container \"06116306c71143bdd3e32fba2acede90d5431479a5f1139b4ef0978bd4147a29\": not found" Feb 9 09:01:57.631849 kubelet[1995]: I0209 09:01:57.631806 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-hubble-tls\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632021 kubelet[1995]: I0209 09:01:57.631867 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-hostproc\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632021 kubelet[1995]: I0209 09:01:57.631894 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-etc-cni-netd\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632413 kubelet[1995]: I0209 09:01:57.632375 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-net\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632500 kubelet[1995]: I0209 09:01:57.632433 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cni-path\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632500 kubelet[1995]: I0209 09:01:57.632479 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f0d9ccfc-d25a-418f-88fb-c69046864e01-clustermesh-secrets\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632658 kubelet[1995]: I0209 09:01:57.632507 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-bpf-maps\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632658 kubelet[1995]: I0209 09:01:57.632532 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqbgd\" (UniqueName: \"kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-kube-api-access-bqbgd\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632658 kubelet[1995]: I0209 09:01:57.632559 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-xtables-lock\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632658 kubelet[1995]: I0209 09:01:57.632584 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-kernel\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632658 kubelet[1995]: I0209 09:01:57.632626 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-lib-modules\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632658 kubelet[1995]: I0209 09:01:57.632658 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-cgroup\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632862 kubelet[1995]: I0209 09:01:57.632682 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-config-path\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.632862 kubelet[1995]: I0209 09:01:57.632707 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-run\") pod \"f0d9ccfc-d25a-418f-88fb-c69046864e01\" (UID: \"f0d9ccfc-d25a-418f-88fb-c69046864e01\") " Feb 9 09:01:57.633200 kubelet[1995]: I0209 09:01:57.633138 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/a9dc981a-ccb2-436a-8813-7b093c787c35-cilium-config-path\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.633200 kubelet[1995]: I0209 09:01:57.633163 1995 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-d9zzg\" (UniqueName: \"kubernetes.io/projected/a9dc981a-ccb2-436a-8813-7b093c787c35-kube-api-access-d9zzg\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.633307 kubelet[1995]: I0209 09:01:57.633214 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.633307 kubelet[1995]: I0209 09:01:57.633250 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-hostproc" (OuterVolumeSpecName: "hostproc") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.633307 kubelet[1995]: I0209 09:01:57.633264 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.633307 kubelet[1995]: I0209 09:01:57.633304 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.633422 kubelet[1995]: I0209 09:01:57.633322 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cni-path" (OuterVolumeSpecName: "cni-path") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.636072 kubelet[1995]: I0209 09:01:57.633662 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.636072 kubelet[1995]: I0209 09:01:57.633707 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.636072 kubelet[1995]: I0209 09:01:57.633747 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.636072 kubelet[1995]: I0209 09:01:57.633771 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.636072 kubelet[1995]: W0209 09:01:57.633974 1995 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/f0d9ccfc-d25a-418f-88fb-c69046864e01/volumes/kubernetes.io~configmap/cilium-config-path: clearQuota called, but quotas disabled Feb 9 09:01:57.636444 kubelet[1995]: I0209 09:01:57.634573 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:01:57.636487 kubelet[1995]: I0209 09:01:57.636446 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 9 09:01:57.643703 kubelet[1995]: I0209 09:01:57.643587 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0d9ccfc-d25a-418f-88fb-c69046864e01-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 9 09:01:57.643953 kubelet[1995]: I0209 09:01:57.643924 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:01:57.644212 kubelet[1995]: I0209 09:01:57.644159 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-kube-api-access-bqbgd" (OuterVolumeSpecName: "kube-api-access-bqbgd") pod "f0d9ccfc-d25a-418f-88fb-c69046864e01" (UID: "f0d9ccfc-d25a-418f-88fb-c69046864e01"). InnerVolumeSpecName "kube-api-access-bqbgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:01:57.733660 kubelet[1995]: I0209 09:01:57.733600 1995 reconciler_common.go:295] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-kernel\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733660 kubelet[1995]: I0209 09:01:57.733653 1995 reconciler_common.go:295] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-lib-modules\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733660 kubelet[1995]: I0209 09:01:57.733666 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-cgroup\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733683 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-config-path\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733697 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cilium-run\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733707 1995 reconciler_common.go:295] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-hubble-tls\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733717 1995 reconciler_common.go:295] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-hostproc\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733728 1995 reconciler_common.go:295] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-etc-cni-netd\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733738 1995 reconciler_common.go:295] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-host-proc-sys-net\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733748 1995 reconciler_common.go:295] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-bpf-maps\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.733910 kubelet[1995]: I0209 09:01:57.733758 1995 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-bqbgd\" (UniqueName: \"kubernetes.io/projected/f0d9ccfc-d25a-418f-88fb-c69046864e01-kube-api-access-bqbgd\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.734176 kubelet[1995]: I0209 09:01:57.733770 1995 reconciler_common.go:295] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-cni-path\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.734176 kubelet[1995]: I0209 09:01:57.733780 1995 reconciler_common.go:295] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/f0d9ccfc-d25a-418f-88fb-c69046864e01-clustermesh-secrets\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.734176 kubelet[1995]: I0209 09:01:57.733790 1995 reconciler_common.go:295] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/f0d9ccfc-d25a-418f-88fb-c69046864e01-xtables-lock\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:01:57.871032 systemd[1]: Removed slice kubepods-burstable-podf0d9ccfc_d25a_418f_88fb_c69046864e01.slice. Feb 9 09:01:57.871148 systemd[1]: kubepods-burstable-podf0d9ccfc_d25a_418f_88fb_c69046864e01.slice: Consumed 9.472s CPU time. Feb 9 09:01:58.132186 env[1102]: time="2024-02-09T09:01:58.131793146Z" level=info msg="StopContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" with timeout 1 (s)" Feb 9 09:01:58.132186 env[1102]: time="2024-02-09T09:01:58.131914913Z" level=error msg="StopContainer for \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\": not found" Feb 9 09:01:58.132861 kubelet[1995]: E0209 09:01:58.132832 1995 remote_runtime.go:349] "StopContainer from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded\": not found" containerID="2c65bb74994afd3a23f3c58c77ced29f88fa74f04eee9deaef4d2398981f5ded" Feb 9 09:01:58.133568 env[1102]: time="2024-02-09T09:01:58.133343105Z" level=info msg="StopPodSandbox for \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\"" Feb 9 09:01:58.133568 env[1102]: time="2024-02-09T09:01:58.133385732Z" level=info msg="StopContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" with timeout 1 (s)" Feb 9 09:01:58.133568 env[1102]: time="2024-02-09T09:01:58.133429098Z" level=error msg="StopContainer for \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\": not found" Feb 9 09:01:58.133568 env[1102]: time="2024-02-09T09:01:58.133483111Z" level=info msg="TearDown network for sandbox \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\" successfully" Feb 9 09:01:58.133568 env[1102]: time="2024-02-09T09:01:58.133538898Z" level=info msg="StopPodSandbox for \"ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940\" returns successfully" Feb 9 09:01:58.135684 kubelet[1995]: E0209 09:01:58.135657 1995 remote_runtime.go:349] "StopContainer from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034\": not found" containerID="6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034" Feb 9 09:01:58.138014 env[1102]: time="2024-02-09T09:01:58.137329839Z" level=info msg="StopPodSandbox for \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\"" Feb 9 09:01:58.138014 env[1102]: time="2024-02-09T09:01:58.137518703Z" level=info msg="TearDown network for sandbox \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" successfully" Feb 9 09:01:58.138014 env[1102]: time="2024-02-09T09:01:58.137592589Z" level=info msg="StopPodSandbox for \"fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d\" returns successfully" Feb 9 09:01:58.139864 kubelet[1995]: I0209 09:01:58.139837 1995 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=a9dc981a-ccb2-436a-8813-7b093c787c35 path="/var/lib/kubelet/pods/a9dc981a-ccb2-436a-8813-7b093c787c35/volumes" Feb 9 09:01:58.140787 kubelet[1995]: I0209 09:01:58.140761 1995 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=f0d9ccfc-d25a-418f-88fb-c69046864e01 path="/var/lib/kubelet/pods/f0d9ccfc-d25a-418f-88fb-c69046864e01/volumes" Feb 9 09:01:58.233371 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6de41acc741b02bfe9a520ebc2aa66295ac31e4e430d1812b069c1d49d083034-rootfs.mount: Deactivated successfully. Feb 9 09:01:58.233549 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ba3dd5d9da20cc85d3a1a696c7127f53a91294c885e1f27d9a63b9d513abf940-rootfs.mount: Deactivated successfully. Feb 9 09:01:58.233621 systemd[1]: var-lib-kubelet-pods-a9dc981a\x2dccb2\x2d436a\x2d8813\x2d7b093c787c35-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dd9zzg.mount: Deactivated successfully. Feb 9 09:01:58.233695 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d-rootfs.mount: Deactivated successfully. Feb 9 09:01:58.233780 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-fd16cbc5120e383e2a05717df18ae205909c981d27989fc924985a79eae7134d-shm.mount: Deactivated successfully. Feb 9 09:01:58.233849 systemd[1]: var-lib-kubelet-pods-f0d9ccfc\x2dd25a\x2d418f\x2d88fb\x2dc69046864e01-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dbqbgd.mount: Deactivated successfully. Feb 9 09:01:58.233920 systemd[1]: var-lib-kubelet-pods-f0d9ccfc\x2dd25a\x2d418f\x2d88fb\x2dc69046864e01-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Feb 9 09:01:58.233983 systemd[1]: var-lib-kubelet-pods-f0d9ccfc\x2dd25a\x2d418f\x2d88fb\x2dc69046864e01-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Feb 9 09:01:58.868847 sshd[3748]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:58.873991 systemd[1]: sshd@26-64.23.148.25:22-139.178.89.65:58016.service: Deactivated successfully. Feb 9 09:01:58.874990 systemd[1]: session-26.scope: Deactivated successfully. Feb 9 09:01:58.875215 systemd[1]: session-26.scope: Consumed 2.693s CPU time. Feb 9 09:01:58.875784 systemd-logind[1095]: Session 26 logged out. Waiting for processes to exit. Feb 9 09:01:58.878362 systemd[1]: Started sshd@27-64.23.148.25:22-139.178.89.65:53798.service. Feb 9 09:01:58.879985 systemd-logind[1095]: Removed session 26. Feb 9 09:01:58.935407 sshd[3912]: Accepted publickey for core from 139.178.89.65 port 53798 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:58.937292 sshd[3912]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:58.944730 systemd[1]: Started session-27.scope. Feb 9 09:01:58.945580 systemd-logind[1095]: New session 27 of user core. Feb 9 09:01:59.129026 kubelet[1995]: E0209 09:01:59.128900 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:01:59.914557 sshd[3912]: pam_unix(sshd:session): session closed for user core Feb 9 09:01:59.921952 systemd[1]: sshd@27-64.23.148.25:22-139.178.89.65:53798.service: Deactivated successfully. Feb 9 09:01:59.922675 systemd[1]: session-27.scope: Deactivated successfully. Feb 9 09:01:59.923482 systemd-logind[1095]: Session 27 logged out. Waiting for processes to exit. Feb 9 09:01:59.925141 systemd[1]: Started sshd@28-64.23.148.25:22-139.178.89.65:53810.service. Feb 9 09:01:59.928569 systemd-logind[1095]: Removed session 27. Feb 9 09:01:59.983704 sshd[3923]: Accepted publickey for core from 139.178.89.65 port 53810 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:01:59.985376 sshd[3923]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:01:59.991123 systemd[1]: Started session-28.scope. Feb 9 09:01:59.991949 systemd-logind[1095]: New session 28 of user core. Feb 9 09:02:00.013026 kubelet[1995]: I0209 09:02:00.012859 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 09:02:00.013687 kubelet[1995]: E0209 09:02:00.013369 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0d9ccfc-d25a-418f-88fb-c69046864e01" containerName="mount-cgroup" Feb 9 09:02:00.013687 kubelet[1995]: E0209 09:02:00.013404 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0d9ccfc-d25a-418f-88fb-c69046864e01" containerName="clean-cilium-state" Feb 9 09:02:00.013687 kubelet[1995]: E0209 09:02:00.013415 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0d9ccfc-d25a-418f-88fb-c69046864e01" containerName="apply-sysctl-overwrites" Feb 9 09:02:00.013687 kubelet[1995]: E0209 09:02:00.013425 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0d9ccfc-d25a-418f-88fb-c69046864e01" containerName="mount-bpf-fs" Feb 9 09:02:00.013687 kubelet[1995]: E0209 09:02:00.013437 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="a9dc981a-ccb2-436a-8813-7b093c787c35" containerName="cilium-operator" Feb 9 09:02:00.013687 kubelet[1995]: E0209 09:02:00.013446 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="f0d9ccfc-d25a-418f-88fb-c69046864e01" containerName="cilium-agent" Feb 9 09:02:00.013687 kubelet[1995]: I0209 09:02:00.013485 1995 memory_manager.go:346] "RemoveStaleState removing state" podUID="f0d9ccfc-d25a-418f-88fb-c69046864e01" containerName="cilium-agent" Feb 9 09:02:00.013687 kubelet[1995]: I0209 09:02:00.013498 1995 memory_manager.go:346] "RemoveStaleState removing state" podUID="a9dc981a-ccb2-436a-8813-7b093c787c35" containerName="cilium-operator" Feb 9 09:02:00.026484 systemd[1]: Created slice kubepods-burstable-pod1ffc38a0_137c_41be_be44_a523cd97b424.slice. Feb 9 09:02:00.150586 kubelet[1995]: I0209 09:02:00.150532 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-xtables-lock\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151014 kubelet[1995]: I0209 09:02:00.150604 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-clustermesh-secrets\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151014 kubelet[1995]: I0209 09:02:00.150633 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-config-path\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151014 kubelet[1995]: I0209 09:02:00.150653 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-kernel\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151014 kubelet[1995]: I0209 09:02:00.150676 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-etc-cni-netd\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151014 kubelet[1995]: I0209 09:02:00.150700 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-ipsec-secrets\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151259 kubelet[1995]: I0209 09:02:00.150720 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-cgroup\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151259 kubelet[1995]: I0209 09:02:00.150742 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cni-path\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151259 kubelet[1995]: I0209 09:02:00.150763 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-hubble-tls\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151259 kubelet[1995]: I0209 09:02:00.150783 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-bpf-maps\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151259 kubelet[1995]: I0209 09:02:00.150805 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-lib-modules\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151259 kubelet[1995]: I0209 09:02:00.150826 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-run\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151574 kubelet[1995]: I0209 09:02:00.150846 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-net\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151574 kubelet[1995]: I0209 09:02:00.150866 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-hostproc\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.151574 kubelet[1995]: I0209 09:02:00.150886 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxj5l\" (UniqueName: \"kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-kube-api-access-nxj5l\") pod \"cilium-4fzs7\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " pod="kube-system/cilium-4fzs7" Feb 9 09:02:00.367021 sshd[3923]: pam_unix(sshd:session): session closed for user core Feb 9 09:02:00.371833 systemd[1]: sshd@28-64.23.148.25:22-139.178.89.65:53810.service: Deactivated successfully. Feb 9 09:02:00.372709 systemd[1]: session-28.scope: Deactivated successfully. Feb 9 09:02:00.373548 systemd-logind[1095]: Session 28 logged out. Waiting for processes to exit. Feb 9 09:02:00.378624 systemd[1]: Started sshd@29-64.23.148.25:22-139.178.89.65:53820.service. Feb 9 09:02:00.380236 systemd-logind[1095]: Removed session 28. Feb 9 09:02:00.427167 sshd[3940]: Accepted publickey for core from 139.178.89.65 port 53820 ssh2: RSA SHA256:zxCjWE6I1sqRNr8f+A5DoPj4YLVmU7ObDiNpO/GSq00 Feb 9 09:02:00.429401 sshd[3940]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Feb 9 09:02:00.435325 systemd[1]: Started session-29.scope. Feb 9 09:02:00.437274 systemd-logind[1095]: New session 29 of user core. Feb 9 09:02:00.632117 kubelet[1995]: E0209 09:02:00.631941 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:00.635132 env[1102]: time="2024-02-09T09:02:00.634687475Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4fzs7,Uid:1ffc38a0-137c-41be-be44-a523cd97b424,Namespace:kube-system,Attempt:0,}" Feb 9 09:02:00.684845 env[1102]: time="2024-02-09T09:02:00.684725229Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:02:00.685207 env[1102]: time="2024-02-09T09:02:00.685146150Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:02:00.685437 env[1102]: time="2024-02-09T09:02:00.685394260Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:02:00.685899 env[1102]: time="2024-02-09T09:02:00.685847055Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07 pid=3955 runtime=io.containerd.runc.v2 Feb 9 09:02:00.719155 systemd[1]: Started cri-containerd-13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07.scope. Feb 9 09:02:00.763878 env[1102]: time="2024-02-09T09:02:00.763809360Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4fzs7,Uid:1ffc38a0-137c-41be-be44-a523cd97b424,Namespace:kube-system,Attempt:0,} returns sandbox id \"13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07\"" Feb 9 09:02:00.765089 kubelet[1995]: E0209 09:02:00.764863 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:00.771300 env[1102]: time="2024-02-09T09:02:00.771242218Z" level=info msg="CreateContainer within sandbox \"13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 9 09:02:00.785509 env[1102]: time="2024-02-09T09:02:00.785431387Z" level=info msg="CreateContainer within sandbox \"13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\"" Feb 9 09:02:00.787142 env[1102]: time="2024-02-09T09:02:00.786260085Z" level=info msg="StartContainer for \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\"" Feb 9 09:02:00.811898 systemd[1]: Started cri-containerd-f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78.scope. Feb 9 09:02:00.833351 systemd[1]: cri-containerd-f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78.scope: Deactivated successfully. Feb 9 09:02:00.864200 env[1102]: time="2024-02-09T09:02:00.864131367Z" level=info msg="shim disconnected" id=f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78 Feb 9 09:02:00.864682 env[1102]: time="2024-02-09T09:02:00.864654644Z" level=warning msg="cleaning up after shim disconnected" id=f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78 namespace=k8s.io Feb 9 09:02:00.864823 env[1102]: time="2024-02-09T09:02:00.864799366Z" level=info msg="cleaning up dead shim" Feb 9 09:02:00.882367 env[1102]: time="2024-02-09T09:02:00.882220597Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:02:00Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4015 runtime=io.containerd.runc.v2\ntime=\"2024-02-09T09:02:00Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Feb 9 09:02:00.883644 env[1102]: time="2024-02-09T09:02:00.883508695Z" level=error msg="copy shim log" error="read /proc/self/fd/41: file already closed" Feb 9 09:02:00.884177 env[1102]: time="2024-02-09T09:02:00.884130441Z" level=error msg="Failed to pipe stderr of container \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\"" error="reading from a closed fifo" Feb 9 09:02:00.884591 env[1102]: time="2024-02-09T09:02:00.884554050Z" level=error msg="Failed to pipe stdout of container \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\"" error="reading from a closed fifo" Feb 9 09:02:00.887853 env[1102]: time="2024-02-09T09:02:00.887774419Z" level=error msg="StartContainer for \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Feb 9 09:02:00.888515 kubelet[1995]: E0209 09:02:00.888484 1995 remote_runtime.go:326] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78" Feb 9 09:02:00.891493 kubelet[1995]: E0209 09:02:00.891446 1995 kuberuntime_manager.go:872] init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Feb 9 09:02:00.891493 kubelet[1995]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Feb 9 09:02:00.891493 kubelet[1995]: rm /hostbin/cilium-mount Feb 9 09:02:00.891493 kubelet[1995]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-nxj5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,} start failed in pod cilium-4fzs7_kube-system(1ffc38a0-137c-41be-be44-a523cd97b424): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Feb 9 09:02:00.892083 kubelet[1995]: E0209 09:02:00.891730 1995 pod_workers.go:965] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-4fzs7" podUID=1ffc38a0-137c-41be-be44-a523cd97b424 Feb 9 09:02:01.143692 kubelet[1995]: E0209 09:02:01.143529 1995 kubelet.go:2475] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Feb 9 09:02:01.580113 env[1102]: time="2024-02-09T09:02:01.580039606Z" level=info msg="StopPodSandbox for \"13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07\"" Feb 9 09:02:01.580486 env[1102]: time="2024-02-09T09:02:01.580447816Z" level=info msg="Container to stop \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Feb 9 09:02:01.583570 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07-shm.mount: Deactivated successfully. Feb 9 09:02:01.596877 systemd[1]: cri-containerd-13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07.scope: Deactivated successfully. Feb 9 09:02:01.640428 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07-rootfs.mount: Deactivated successfully. Feb 9 09:02:01.648725 env[1102]: time="2024-02-09T09:02:01.648638398Z" level=info msg="shim disconnected" id=13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07 Feb 9 09:02:01.650328 env[1102]: time="2024-02-09T09:02:01.650279188Z" level=warning msg="cleaning up after shim disconnected" id=13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07 namespace=k8s.io Feb 9 09:02:01.650551 env[1102]: time="2024-02-09T09:02:01.650521146Z" level=info msg="cleaning up dead shim" Feb 9 09:02:01.668984 env[1102]: time="2024-02-09T09:02:01.668907114Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:02:01Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4045 runtime=io.containerd.runc.v2\n" Feb 9 09:02:01.669480 env[1102]: time="2024-02-09T09:02:01.669435588Z" level=info msg="TearDown network for sandbox \"13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07\" successfully" Feb 9 09:02:01.669537 env[1102]: time="2024-02-09T09:02:01.669483036Z" level=info msg="StopPodSandbox for \"13240660d8396a87e3e4a2416bf9ca4705f0e8799879fd2560aff3be3ca51c07\" returns successfully" Feb 9 09:02:01.764521 kubelet[1995]: I0209 09:02:01.764455 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-xtables-lock\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.764521 kubelet[1995]: I0209 09:02:01.764513 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-etc-cni-netd\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765195 kubelet[1995]: I0209 09:02:01.764578 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxj5l\" (UniqueName: \"kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-kube-api-access-nxj5l\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765195 kubelet[1995]: I0209 09:02:01.764630 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-config-path\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765195 kubelet[1995]: I0209 09:02:01.764658 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-net\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765195 kubelet[1995]: I0209 09:02:01.764702 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-cgroup\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765195 kubelet[1995]: I0209 09:02:01.764731 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cni-path\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765195 kubelet[1995]: I0209 09:02:01.764762 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-clustermesh-secrets\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765513 kubelet[1995]: I0209 09:02:01.764802 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-hubble-tls\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765513 kubelet[1995]: I0209 09:02:01.764830 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-lib-modules\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765513 kubelet[1995]: I0209 09:02:01.764886 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-run\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765513 kubelet[1995]: I0209 09:02:01.764918 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-hostproc\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765513 kubelet[1995]: I0209 09:02:01.764961 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-kernel\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765513 kubelet[1995]: I0209 09:02:01.765087 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-ipsec-secrets\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765799 kubelet[1995]: I0209 09:02:01.765138 1995 reconciler_common.go:169] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-bpf-maps\") pod \"1ffc38a0-137c-41be-be44-a523cd97b424\" (UID: \"1ffc38a0-137c-41be-be44-a523cd97b424\") " Feb 9 09:02:01.765799 kubelet[1995]: I0209 09:02:01.765271 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.765799 kubelet[1995]: I0209 09:02:01.765306 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.765799 kubelet[1995]: I0209 09:02:01.765322 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.766013 kubelet[1995]: W0209 09:02:01.765977 1995 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/1ffc38a0-137c-41be-be44-a523cd97b424/volumes/kubernetes.io~configmap/cilium-config-path: clearQuota called, but quotas disabled Feb 9 09:02:01.768596 kubelet[1995]: I0209 09:02:01.768519 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 9 09:02:01.768800 kubelet[1995]: I0209 09:02:01.768615 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.768800 kubelet[1995]: I0209 09:02:01.768645 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.768800 kubelet[1995]: I0209 09:02:01.768668 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cni-path" (OuterVolumeSpecName: "cni-path") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.768965 kubelet[1995]: I0209 09:02:01.768929 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-hostproc" (OuterVolumeSpecName: "hostproc") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.768965 kubelet[1995]: I0209 09:02:01.768959 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.769118 kubelet[1995]: I0209 09:02:01.768975 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.769118 kubelet[1995]: I0209 09:02:01.768991 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 9 09:02:01.775273 systemd[1]: var-lib-kubelet-pods-1ffc38a0\x2d137c\x2d41be\x2dbe44\x2da523cd97b424-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dnxj5l.mount: Deactivated successfully. Feb 9 09:02:01.777060 kubelet[1995]: I0209 09:02:01.777011 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-kube-api-access-nxj5l" (OuterVolumeSpecName: "kube-api-access-nxj5l") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "kube-api-access-nxj5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:02:01.781149 systemd[1]: var-lib-kubelet-pods-1ffc38a0\x2d137c\x2d41be\x2dbe44\x2da523cd97b424-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Feb 9 09:02:01.783299 kubelet[1995]: I0209 09:02:01.783260 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 9 09:02:01.785982 kubelet[1995]: I0209 09:02:01.785943 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 9 09:02:01.786466 kubelet[1995]: I0209 09:02:01.786442 1995 operation_generator.go:900] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "1ffc38a0-137c-41be-be44-a523cd97b424" (UID: "1ffc38a0-137c-41be-be44-a523cd97b424"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.866161 1995 reconciler_common.go:295] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-lib-modules\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867021 1995 reconciler_common.go:295] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-kernel\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867083 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-run\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867103 1995 reconciler_common.go:295] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-hostproc\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867121 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-ipsec-secrets\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867138 1995 reconciler_common.go:295] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-bpf-maps\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867158 1995 reconciler_common.go:295] "Volume detached for volume \"kube-api-access-nxj5l\" (UniqueName: \"kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-kube-api-access-nxj5l\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867372 kubelet[1995]: I0209 09:02:01.867177 1995 reconciler_common.go:295] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-xtables-lock\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867194 1995 reconciler_common.go:295] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-etc-cni-netd\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867211 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-config-path\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867231 1995 reconciler_common.go:295] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cilium-cgroup\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867247 1995 reconciler_common.go:295] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-cni-path\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867266 1995 reconciler_common.go:295] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1ffc38a0-137c-41be-be44-a523cd97b424-host-proc-sys-net\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867284 1995 reconciler_common.go:295] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1ffc38a0-137c-41be-be44-a523cd97b424-clustermesh-secrets\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:01.867971 kubelet[1995]: I0209 09:02:01.867302 1995 reconciler_common.go:295] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1ffc38a0-137c-41be-be44-a523cd97b424-hubble-tls\") on node \"ci-3510.3.2-5-6e53ffd492\" DevicePath \"\"" Feb 9 09:02:02.135792 systemd[1]: Removed slice kubepods-burstable-pod1ffc38a0_137c_41be_be44_a523cd97b424.slice. Feb 9 09:02:02.261832 systemd[1]: var-lib-kubelet-pods-1ffc38a0\x2d137c\x2d41be\x2dbe44\x2da523cd97b424-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Feb 9 09:02:02.262310 systemd[1]: var-lib-kubelet-pods-1ffc38a0\x2d137c\x2d41be\x2dbe44\x2da523cd97b424-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Feb 9 09:02:02.584092 kubelet[1995]: I0209 09:02:02.584034 1995 scope.go:115] "RemoveContainer" containerID="f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78" Feb 9 09:02:02.587576 env[1102]: time="2024-02-09T09:02:02.587132788Z" level=info msg="RemoveContainer for \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\"" Feb 9 09:02:02.593065 env[1102]: time="2024-02-09T09:02:02.592984826Z" level=info msg="RemoveContainer for \"f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78\" returns successfully" Feb 9 09:02:02.673914 kubelet[1995]: I0209 09:02:02.673872 1995 topology_manager.go:210] "Topology Admit Handler" Feb 9 09:02:02.674151 kubelet[1995]: E0209 09:02:02.673939 1995 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1ffc38a0-137c-41be-be44-a523cd97b424" containerName="mount-cgroup" Feb 9 09:02:02.674151 kubelet[1995]: I0209 09:02:02.673973 1995 memory_manager.go:346] "RemoveStaleState removing state" podUID="1ffc38a0-137c-41be-be44-a523cd97b424" containerName="mount-cgroup" Feb 9 09:02:02.680946 systemd[1]: Created slice kubepods-burstable-pod6865b303_2d00_42ea_957c_535e307c4ba9.slice. Feb 9 09:02:02.773609 kubelet[1995]: I0209 09:02:02.773559 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-cilium-cgroup\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774061 kubelet[1995]: I0209 09:02:02.773630 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-lib-modules\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774061 kubelet[1995]: I0209 09:02:02.773667 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-cilium-run\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774061 kubelet[1995]: I0209 09:02:02.773699 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-hostproc\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774061 kubelet[1995]: I0209 09:02:02.773734 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-etc-cni-netd\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774061 kubelet[1995]: I0209 09:02:02.773769 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-cni-path\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774061 kubelet[1995]: I0209 09:02:02.773804 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-host-proc-sys-net\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774288 kubelet[1995]: I0209 09:02:02.773839 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-host-proc-sys-kernel\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774288 kubelet[1995]: I0209 09:02:02.773897 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6865b303-2d00-42ea-957c-535e307c4ba9-cilium-config-path\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774288 kubelet[1995]: I0209 09:02:02.773929 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6865b303-2d00-42ea-957c-535e307c4ba9-hubble-tls\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774288 kubelet[1995]: I0209 09:02:02.773966 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v9bc\" (UniqueName: \"kubernetes.io/projected/6865b303-2d00-42ea-957c-535e307c4ba9-kube-api-access-4v9bc\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774288 kubelet[1995]: I0209 09:02:02.774001 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-xtables-lock\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774433 kubelet[1995]: I0209 09:02:02.774064 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6865b303-2d00-42ea-957c-535e307c4ba9-clustermesh-secrets\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774433 kubelet[1995]: I0209 09:02:02.774100 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6865b303-2d00-42ea-957c-535e307c4ba9-bpf-maps\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.774433 kubelet[1995]: I0209 09:02:02.774135 1995 reconciler_common.go:253] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6865b303-2d00-42ea-957c-535e307c4ba9-cilium-ipsec-secrets\") pod \"cilium-vfz5g\" (UID: \"6865b303-2d00-42ea-957c-535e307c4ba9\") " pod="kube-system/cilium-vfz5g" Feb 9 09:02:02.983719 kubelet[1995]: E0209 09:02:02.983649 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:02.985642 env[1102]: time="2024-02-09T09:02:02.984427461Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vfz5g,Uid:6865b303-2d00-42ea-957c-535e307c4ba9,Namespace:kube-system,Attempt:0,}" Feb 9 09:02:03.001513 env[1102]: time="2024-02-09T09:02:03.001216532Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Feb 9 09:02:03.001513 env[1102]: time="2024-02-09T09:02:03.001279145Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Feb 9 09:02:03.001513 env[1102]: time="2024-02-09T09:02:03.001296546Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Feb 9 09:02:03.001822 env[1102]: time="2024-02-09T09:02:03.001574696Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4 pid=4074 runtime=io.containerd.runc.v2 Feb 9 09:02:03.022812 systemd[1]: Started cri-containerd-077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4.scope. Feb 9 09:02:03.072183 env[1102]: time="2024-02-09T09:02:03.072031966Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-vfz5g,Uid:6865b303-2d00-42ea-957c-535e307c4ba9,Namespace:kube-system,Attempt:0,} returns sandbox id \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\"" Feb 9 09:02:03.073833 kubelet[1995]: E0209 09:02:03.073799 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:03.077413 env[1102]: time="2024-02-09T09:02:03.077354204Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Feb 9 09:02:03.088390 env[1102]: time="2024-02-09T09:02:03.088299222Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4\"" Feb 9 09:02:03.089613 env[1102]: time="2024-02-09T09:02:03.089573239Z" level=info msg="StartContainer for \"5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4\"" Feb 9 09:02:03.121744 systemd[1]: Started cri-containerd-5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4.scope. Feb 9 09:02:03.172981 env[1102]: time="2024-02-09T09:02:03.172906254Z" level=info msg="StartContainer for \"5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4\" returns successfully" Feb 9 09:02:03.186649 systemd[1]: cri-containerd-5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4.scope: Deactivated successfully. Feb 9 09:02:03.223446 env[1102]: time="2024-02-09T09:02:03.223381020Z" level=info msg="shim disconnected" id=5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4 Feb 9 09:02:03.223981 env[1102]: time="2024-02-09T09:02:03.223932858Z" level=warning msg="cleaning up after shim disconnected" id=5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4 namespace=k8s.io Feb 9 09:02:03.224229 env[1102]: time="2024-02-09T09:02:03.224199447Z" level=info msg="cleaning up dead shim" Feb 9 09:02:03.242713 env[1102]: time="2024-02-09T09:02:03.241268892Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:02:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4157 runtime=io.containerd.runc.v2\n" Feb 9 09:02:03.597460 kubelet[1995]: E0209 09:02:03.594868 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:03.604747 env[1102]: time="2024-02-09T09:02:03.604692926Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Feb 9 09:02:03.623502 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3195637488.mount: Deactivated successfully. Feb 9 09:02:03.633794 env[1102]: time="2024-02-09T09:02:03.633731762Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a\"" Feb 9 09:02:03.634777 env[1102]: time="2024-02-09T09:02:03.634736982Z" level=info msg="StartContainer for \"3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a\"" Feb 9 09:02:03.676444 systemd[1]: Started cri-containerd-3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a.scope. Feb 9 09:02:03.731902 env[1102]: time="2024-02-09T09:02:03.731843411Z" level=info msg="StartContainer for \"3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a\" returns successfully" Feb 9 09:02:03.765587 systemd[1]: cri-containerd-3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a.scope: Deactivated successfully. Feb 9 09:02:03.791645 env[1102]: time="2024-02-09T09:02:03.791588503Z" level=info msg="shim disconnected" id=3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a Feb 9 09:02:03.792009 env[1102]: time="2024-02-09T09:02:03.791974757Z" level=warning msg="cleaning up after shim disconnected" id=3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a namespace=k8s.io Feb 9 09:02:03.792382 env[1102]: time="2024-02-09T09:02:03.792358876Z" level=info msg="cleaning up dead shim" Feb 9 09:02:03.808197 env[1102]: time="2024-02-09T09:02:03.808135452Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:02:03Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4219 runtime=io.containerd.runc.v2\n" Feb 9 09:02:03.981992 kubelet[1995]: W0209 09:02:03.980658 1995 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ffc38a0_137c_41be_be44_a523cd97b424.slice/cri-containerd-f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78.scope WatchSource:0}: container "f3b7848100ad651d7b515ef6a2213eb6663bcbfd2b1afd40234a5f6a75a5df78" in namespace "k8s.io": not found Feb 9 09:02:04.132208 kubelet[1995]: I0209 09:02:04.132179 1995 kubelet_volumes.go:160] "Cleaned up orphaned pod volumes dir" podUID=1ffc38a0-137c-41be-be44-a523cd97b424 path="/var/lib/kubelet/pods/1ffc38a0-137c-41be-be44-a523cd97b424/volumes" Feb 9 09:02:04.612648 kubelet[1995]: E0209 09:02:04.612619 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:04.616629 env[1102]: time="2024-02-09T09:02:04.616398607Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Feb 9 09:02:04.637956 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1976589595.mount: Deactivated successfully. Feb 9 09:02:04.641652 env[1102]: time="2024-02-09T09:02:04.641600462Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9\"" Feb 9 09:02:04.643115 env[1102]: time="2024-02-09T09:02:04.643080252Z" level=info msg="StartContainer for \"cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9\"" Feb 9 09:02:04.697921 systemd[1]: Started cri-containerd-cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9.scope. Feb 9 09:02:04.774442 env[1102]: time="2024-02-09T09:02:04.774396800Z" level=info msg="StartContainer for \"cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9\" returns successfully" Feb 9 09:02:04.783898 systemd[1]: cri-containerd-cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9.scope: Deactivated successfully. Feb 9 09:02:04.831896 env[1102]: time="2024-02-09T09:02:04.831820242Z" level=info msg="shim disconnected" id=cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9 Feb 9 09:02:04.831896 env[1102]: time="2024-02-09T09:02:04.831885831Z" level=warning msg="cleaning up after shim disconnected" id=cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9 namespace=k8s.io Feb 9 09:02:04.831896 env[1102]: time="2024-02-09T09:02:04.831900804Z" level=info msg="cleaning up dead shim" Feb 9 09:02:04.857163 env[1102]: time="2024-02-09T09:02:04.854192795Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:02:04Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4278 runtime=io.containerd.runc.v2\n" Feb 9 09:02:05.262140 systemd[1]: run-containerd-runc-k8s.io-cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9-runc.xujYLh.mount: Deactivated successfully. Feb 9 09:02:05.262628 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9-rootfs.mount: Deactivated successfully. Feb 9 09:02:05.617006 kubelet[1995]: E0209 09:02:05.616677 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:05.619378 env[1102]: time="2024-02-09T09:02:05.619328172Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Feb 9 09:02:05.639843 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount470219521.mount: Deactivated successfully. Feb 9 09:02:05.643838 env[1102]: time="2024-02-09T09:02:05.643763465Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989\"" Feb 9 09:02:05.644634 env[1102]: time="2024-02-09T09:02:05.644585847Z" level=info msg="StartContainer for \"cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989\"" Feb 9 09:02:05.681157 systemd[1]: Started cri-containerd-cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989.scope. Feb 9 09:02:05.721231 systemd[1]: cri-containerd-cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989.scope: Deactivated successfully. Feb 9 09:02:05.724640 env[1102]: time="2024-02-09T09:02:05.724205558Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6865b303_2d00_42ea_957c_535e307c4ba9.slice/cri-containerd-cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989.scope/memory.events\": no such file or directory" Feb 9 09:02:05.726244 env[1102]: time="2024-02-09T09:02:05.726124112Z" level=info msg="StartContainer for \"cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989\" returns successfully" Feb 9 09:02:05.756190 env[1102]: time="2024-02-09T09:02:05.756138529Z" level=info msg="shim disconnected" id=cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989 Feb 9 09:02:05.756809 env[1102]: time="2024-02-09T09:02:05.756775363Z" level=warning msg="cleaning up after shim disconnected" id=cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989 namespace=k8s.io Feb 9 09:02:05.756961 env[1102]: time="2024-02-09T09:02:05.756941810Z" level=info msg="cleaning up dead shim" Feb 9 09:02:05.768429 env[1102]: time="2024-02-09T09:02:05.768374672Z" level=warning msg="cleanup warnings time=\"2024-02-09T09:02:05Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4335 runtime=io.containerd.runc.v2\n" Feb 9 09:02:06.144619 kubelet[1995]: E0209 09:02:06.144587 1995 kubelet.go:2475] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Feb 9 09:02:06.261657 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989-rootfs.mount: Deactivated successfully. Feb 9 09:02:06.622080 kubelet[1995]: E0209 09:02:06.622027 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:06.627113 env[1102]: time="2024-02-09T09:02:06.626569762Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Feb 9 09:02:06.643929 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3041349096.mount: Deactivated successfully. Feb 9 09:02:06.653858 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2050528380.mount: Deactivated successfully. Feb 9 09:02:06.657920 env[1102]: time="2024-02-09T09:02:06.657848384Z" level=info msg="CreateContainer within sandbox \"077f7403538f01595b4c68d243c3efbf8adb153cfc6275547d0edf01cd2a62e4\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"8013dad42ba301d3323e6238468b439ae794acd68acd5e7b27ac16a81adb25a5\"" Feb 9 09:02:06.659760 env[1102]: time="2024-02-09T09:02:06.659709367Z" level=info msg="StartContainer for \"8013dad42ba301d3323e6238468b439ae794acd68acd5e7b27ac16a81adb25a5\"" Feb 9 09:02:06.680749 systemd[1]: Started cri-containerd-8013dad42ba301d3323e6238468b439ae794acd68acd5e7b27ac16a81adb25a5.scope. Feb 9 09:02:06.753308 env[1102]: time="2024-02-09T09:02:06.753257057Z" level=info msg="StartContainer for \"8013dad42ba301d3323e6238468b439ae794acd68acd5e7b27ac16a81adb25a5\" returns successfully" Feb 9 09:02:07.098309 kubelet[1995]: W0209 09:02:07.098263 1995 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6865b303_2d00_42ea_957c_535e307c4ba9.slice/cri-containerd-5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4.scope WatchSource:0}: task 5f8fa58c69b24b5cb6479d1110b580b59fc4c47e4d39c1217574a829e9c456e4 not found: not found Feb 9 09:02:07.281077 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Feb 9 09:02:07.629848 kubelet[1995]: E0209 09:02:07.628809 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:07.678844 kubelet[1995]: I0209 09:02:07.678792 1995 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-vfz5g" podStartSLOduration=5.678752999 pod.CreationTimestamp="2024-02-09 09:02:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-02-09 09:02:07.678405933 +0000 UTC m=+161.932630702" watchObservedRunningTime="2024-02-09 09:02:07.678752999 +0000 UTC m=+161.932977765" Feb 9 09:02:08.631833 kubelet[1995]: E0209 09:02:08.631792 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:09.055778 systemd[1]: run-containerd-runc-k8s.io-8013dad42ba301d3323e6238468b439ae794acd68acd5e7b27ac16a81adb25a5-runc.BS4wsv.mount: Deactivated successfully. Feb 9 09:02:09.106595 kubelet[1995]: I0209 09:02:09.105571 1995 setters.go:548] "Node became not ready" node="ci-3510.3.2-5-6e53ffd492" condition={Type:Ready Status:False LastHeartbeatTime:2024-02-09 09:02:09.104891072 +0000 UTC m=+163.359115834 LastTransitionTime:2024-02-09 09:02:09.104891072 +0000 UTC m=+163.359115834 Reason:KubeletNotReady Message:container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized} Feb 9 09:02:09.633029 kubelet[1995]: E0209 09:02:09.632973 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:10.208751 kubelet[1995]: W0209 09:02:10.208691 1995 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6865b303_2d00_42ea_957c_535e307c4ba9.slice/cri-containerd-3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a.scope WatchSource:0}: task 3e9378305bb9ffb98a25ad2a4d336d2fb90a51138e996b02ad95117da5a38a6a not found: not found Feb 9 09:02:10.558675 systemd-networkd[997]: lxc_health: Link UP Feb 9 09:02:10.564331 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Feb 9 09:02:10.563297 systemd-networkd[997]: lxc_health: Gained carrier Feb 9 09:02:10.986508 kubelet[1995]: E0209 09:02:10.986463 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:11.638555 kubelet[1995]: E0209 09:02:11.638509 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:12.200348 systemd-networkd[997]: lxc_health: Gained IPv6LL Feb 9 09:02:12.641323 kubelet[1995]: E0209 09:02:12.641195 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:13.318199 kubelet[1995]: W0209 09:02:13.318146 1995 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6865b303_2d00_42ea_957c_535e307c4ba9.slice/cri-containerd-cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9.scope WatchSource:0}: task cb534c150678570295d30bdb25c14dd08321bbf48e3cda2126c0022cf0fa4ab9 not found: not found Feb 9 09:02:13.692563 systemd[1]: Started sshd@30-64.23.148.25:22-218.92.0.45:28632.service. Feb 9 09:02:13.884703 sshd[4980]: Unable to negotiate with 218.92.0.45 port 28632: no matching key exchange method found. Their offer: diffie-hellman-group1-sha1,diffie-hellman-group14-sha1,diffie-hellman-group-exchange-sha1 [preauth] Feb 9 09:02:13.886331 systemd[1]: sshd@30-64.23.148.25:22-218.92.0.45:28632.service: Deactivated successfully. Feb 9 09:02:14.128613 kubelet[1995]: E0209 09:02:14.128481 1995 dns.go:156] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 67.207.67.3 67.207.67.2 67.207.67.3" Feb 9 09:02:15.838240 systemd[1]: run-containerd-runc-k8s.io-8013dad42ba301d3323e6238468b439ae794acd68acd5e7b27ac16a81adb25a5-runc.uwTMZ4.mount: Deactivated successfully. Feb 9 09:02:15.914855 sshd[3940]: pam_unix(sshd:session): session closed for user core Feb 9 09:02:15.919749 systemd[1]: sshd@29-64.23.148.25:22-139.178.89.65:53820.service: Deactivated successfully. Feb 9 09:02:15.920830 systemd[1]: session-29.scope: Deactivated successfully. Feb 9 09:02:15.921639 systemd-logind[1095]: Session 29 logged out. Waiting for processes to exit. Feb 9 09:02:15.922946 systemd-logind[1095]: Removed session 29. Feb 9 09:02:16.433874 kubelet[1995]: W0209 09:02:16.433824 1995 manager.go:1174] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6865b303_2d00_42ea_957c_535e307c4ba9.slice/cri-containerd-cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989.scope WatchSource:0}: task cd617e77500ceca58a6ffb81f2f3c27ba9e7f9254e3414dea09e4dff8877f989 not found: not found