Unlocked LM Studio Backends (v1.59.0): AVX1 & More Supported – Testers Wanted

Posted by TheSpicyBoi123@reddit | LocalLLaMA | View on Reddit | 13 comments

Hello everyone!

The latest patched backend versions (1.59.0) are now out, and they bring full support for “unsupported” hardware via a simple patch (see GitHub). Since the last update 3 months ago, these builds have received major refinements in performance, compatibility, and stability via optimized compiler flags and work by llama cpp team.

Here’s the current testing status:

AVX1 CPU builds: working (tested on Ivy Bridge Xeons)
AVX1 Vulkan builds: working (tested on Ivy Bridge Xeons + Tesla K40 GPUs)
AVX1 CUDA builds: untested (no compatible hardware yet)
Non-AVX experimental builds: untested (no compatible hardware yet)

I’m looking for testers to try the newest versions on different hardware, especially non-AVX2 CPUs and newer NVIDIA GPUs, and share performance results. Testers are also wanted for speed comparisons of the new vs old cpu backends.

👉 GitHub link: lmstudio-unlocked-backend

Brief install instructions:
- navigate to backends folder. ex C:\Users\Admin\.lmstudio\extensions\backends
- (recommended for clean install) delete everything except "vendor" folder
- drop contents from compressed backend of your choice

- select it in LM Studio runtimes and enjoy.