Open WebUI Desktop Released!
Posted by My_Unbiased_Opinion@reddit | LocalLLaMA | View on Reddit | 110 comments
Looks like this also includes llama.cpp.
Posted by My_Unbiased_Opinion@reddit | LocalLLaMA | View on Reddit | 110 comments
Looks like this also includes llama.cpp.
nfamerr@reddit
It's broken asf, can't connect to my already established server, even though it's suppposed to be able to with the localurl
deepspace86@reddit
Cool. Did they fix that god awful MCP integration yet?
ClassicMain@reddit
Hi what's so awful about it? Please let us know so we can fix it! We are not aware of any issues at the moment thanks!.
ParthProLegend@reddit
u/Signal_Confusion_644
Signal_Confusion_644@reddit
Im sorry dude/gal, im still processing that i was NOT crazy, cant cope with it yet.
ParthProLegend@reddit
Lol that's a nice reply. Fully inclusive for millennials. Thankfully I am one. 🤣
ClassicMain@reddit
No rush, take your time. Whenever you're ready, please let us know what broke — we'd really like to fix it if it's still an issue, or at least point you at what changed since.
CC u/deepspace86 do you have any issues still?
deepspace86@reddit
Its not that anything is "broken", just that the continued requirement of the MCPO layer is a non-starter for me.
ClassicMain@reddit
its not required you can just use mcp
deepspace86@reddit
Only for streamable http. Stdio doesn't seem to be an option.
ClassicMain@reddit
yes only streamable
jerieljan@reddit
I haven't tried 0.9 yet but if I have to list from my own use of it from before:
tool calls aren't well communicated while it processes them (when I use it, I only see MCP use when the results are out, and not while it's querying, like how OWUI Tools does it (or any other harness that supports MCP properly) This imho is a deal-breaker, since I always observe what MCP calls it's trying in sequence or if it's overdoing connections, etc.
there are occasions when it simply just sees the MCP as unavailable and needs a retry and it works. This seems to be fixed lately but I recall this being an outstanding issue in past versions and was part of the reason I stopped using OWUI for a while.
trying to diagnose connections in general from the UI sucks. Not just an MCP thing, but also for OpenAPI and even adding plain OpenAI API compatible servers. I always have to connect and observe my application logs just to see what's holding up. (Speaking of OpenAPI tool servers, I'm baffled that I can't just upload an openapi.json file myself since not all OpenAPI servers have it on {base_url]/openapi.json)
Puzzleheaded_Mine392@reddit
separate from OWUI itself, for the in-flight tracing piece you can try the mcp-use Inspector. it watches MCP calls live (inputs, outputs, which tool got picked), standalone from any client. https://github.com/mcp-use/mcp-use/tree/main/libraries/typescript/packages/inspector
there's also a free online version: https://inspector.mcp-use.com/
ClassicMain@reddit
1) native tool calling? On? Mcp should work just fine and also show what tools are being called
2) yes mcp should work today
3) hmmmm that's a bit of a bummer. How can this be improved? Of course currently logging is on backend / application since that's what's making the connections to openapi - the frontend doesn't know what the backend does
jerieljan@reddit
I was referring to how it's supposed to emit events while MCPs are being called. Anyway, I dug further on the docs (here and here) and saw that this stuff is configurable. That's on me I guess, but kinda wish that was clarified better since tool call feedback is on by default on other harnesses (e.g., when you configure something like Claude Code to use an MCP and it executes that, you actually see said tool calls being indicated and not just a pulsating circle or hoping that the thinking response is the real deal)
Honestly, I want the error to be more than just the catch-all "Connection failed" toast when you test the connection and imho it'd be better if it showed the actual error that is returned as an exception in the application log like "Could not fetch tool server spec from https://your.openapi.server/openapi.json" )
ClassicMain@reddit
On 2:
I see why you'd wanna have it. Usually the frontend errors are purposefully kept simpler and user friendly (still a few spots where that isn't the case but we're working on it) to not show ugly errors to non-techy end users since those who need to fix it usually have access to the backend anyways
Foreign_Risk_2031@reddit
MCP is a terrible thing to use ANYWAY
demidev@reddit
They support mcp natively now without the use of MCPO as a bridge if that's what your issue was
deepspace86@reddit
That's not what I would call "natively"
simracerman@reddit
That’s only for streamable http. For stdio or see-based servers you must have MCPO as a translation layer.
That’s in their docs.
simracerman@reddit
Wait, how?
Signal_Confusion_644@reddit
Oh... i was NOT crazy.
Cruxicil@reddit
Can someone explain to me what the server is? I would just like to make us of the chat interface and connect the openrouter API. I am using the app on my laptop, so I can't perform inference locally. Currently it is asking me for a server link to connect to before I can use it.
Caffdy@reddit
Friendly reminder that the guy is another tech-bro wannabe, I don't trust him or his product, very obvious he's riding the bandwagon of open source for publicity. he dismisses people issues on github all the time, closing them without recourse. And don't even start talking about the license.
ADHighDef@reddit
LinkedIn ahhh manifesto
walden42@reddit
I went into the article expecting egotistical BS, instead got the opposite, someone who wants to use AI to push technology forward to make everyone's lives better through self-sufficiency. You just sold me more on open webui, sir.
Caffdy@reddit
you just don't know yet how these people express themselves, but welp, to each their own
walden42@reddit
You shared the article, I assumed, to prove that he's untrustworthy. I don't know the guy and so can't comment on whether he can be trusted, but the article itself doesn't at all indicate it.
Tormeister@reddit
I use OpenCode for coding and multi-turn tasks, and OpenWebUI for "chatting" (like general knowledge queries on big models).
Is there any interesting use case for OpenWebUI I'm missing / I should know about?
My_Unbiased_Opinion@reddit (OP)
I use it to serve local LLMs to family and friends. It's very easy to use for the non techy folks.Â
Danmoreng@reddit
Calling it desktop app but it’s a packaged web frontend. 🥲
Exciting_Garden2535@reddit
Do we have a non-web-based markdown engine for any of the non-web desktop frameworks?
Danmoreng@reddit
Just build it from scratch. It’s really not hard with coding agents. My vibe-coded markdown viewer uses Skia for rendering: https://github.com/Danmoreng/simple-markdown-viewer
Exciting_Garden2535@reddit
Yeah, that is cool and a very good start, but adding all features, like code syntax highlighting, mathematical formulas, mermaid schemas (not in LM Studio, sadly, but some UIs have it, and it is so easy to add to a web-based implementation by himself), etc., is a lot of work. And supporting more than one platform with all these features is just a ton of extra headaches.
I love native apps, and I miss the days when a desktop app with a rich UI was only a few dozen kilobytes. But for this particular case, full markdown support across platforms, I'm afraid, it will always be behind web-based apps. Anyway, I wish you luck!
shadiakiki1986@reddit
Good observation. Related discussion:
Why is Claude an Electron App?
If code is free, why aren’t all apps native?
https://www.dbreunig.com/2026/02/21/why-is-claude-an-electron-app.html
tiffanytrashcan@reddit
This is why a few really cool apps are MacOS only. Or the long history of Windows only programs.
The variety of Linux DEs adds to this, the X11/Wayland issues, with an Electron app you can assume the user's clipboard is going to work even on bizarre bespoke distros.
There was a blog post I recently read from a Game Dev about GNOME, and it's special requirements a developer has to code in to make a native app look right on stuff like Ubuntu. A good wrapper has all of those unique quirks already handled.
Zc5Gwu@reddit
Their next product be like "Open WebUI Desktop Mobile TV".
This_Maintenance_834@reddit
does not work on ubuntu 24.04. the server runs, but the desktop shows nothing but a grey screen. also, it overrides the shift-ctrl-c shortcut in ubuntu terminal. did these programmers use linux at all? no one tell them people need to copy stuff from terminal?
jakegh@reddit
The name is kinda contradictory. Unless it's an electron app I guess, which wouldn't be super surprising.
Gallardo994@reddit
Damn I wish they made a version without any bundled inference engines
Top-Rub-4670@reddit
If they're bundling non-cuda llama.cpp we're talking 40MB of disk space here. Might be even less if they built it without its webui.
open-webui itself, with all its dependencies, is over 2GB.
So, yeah, you have a weird gripe.
datbackup@reddit
1000% this. Frontend that is only a frontend is such an underserved niche
But also understandable considering how small the market actually is
ClassicMain@reddit
The inference engine is NOT installed/bundled by default, it asks you if you want to install it and you can decline :)
demidev@reddit
I think this part probably needs to be made clearer, the way it is written now makes it seem like llama cpp has to be installed together as well
Bulb93@reddit
Sounds very similar to just installing llama.cpp server?
maxpayne07@reddit
I just download using their internal download manager unsloth qwen3.6 , but llama.cpp is not detecting the internal model
Corosus@reddit
100% agree the phrasing needs adjusting.
ClassicMain@reddit
nah its optional - you can pick it to install it or not install it :)
Thanks 4 the feedback will make sure Tim adjusts it.
flextrek_whipsnake@reddit
If that's all you want it's easy to vibe code one in an afternoon. That's what I did.
LePfeiff@reddit
It would be trivially easy to vibecode your own web frontend
ClassicMain@reddit
Open an issue on the desktop repository as a feature request otherwise this will get lost
AnonsAnonAnonagain@reddit
I need mah frontend! So I can put it on my head node, and then inference is only a short api call away
Jackw78@reddit
Cherry studio and Aionui are pretty good imo, neither with any inference engine
silenceimpaired@reddit
Every time there is a open webui post someone is praising cherry ...
tiffanytrashcan@reddit
Interesting how they lock the backend behind a paywall.
ClassicMain@reddit
The inference engine is NOT installed/bundled by default, it asks you if you want to install it and you can decline :)
Tig_Old_Bits@reddit
Cherry Studio is perfection!
AntiqueHedgehog8513@reddit
I've just used the app, and it seems like you can opt to not install it if you want and only use your server!
It's honestly very cool.
OldHamburger7923@reddit
can you have it use local model and it access your local file system / git cloned repo for code changes?
GCoderDCoder@reddit
These moments make me think I might be on the spectrum... after you saying you want a ui for your llms without the llm engine bundled? Because that's what open web ui is before this desktop app... I cant tell if yall are being serious and I'm just missing the joke or if I'm misunderstanding something.
If you want the web interface to feel like a desktop app install it as a chrome app and then you get an icon that only goes to that page. That's how i use mine on mobile.
If I just missed the joke then sorry my only friends are Chinese open weight models that I talk too all day. The quants I use can be low on nuance so it may be rubbing off on me lol
Altruistic_Heat_9531@reddit
if i am not mistaken the repo build is bundled with full fat cuda torch
Patient_Tea_401@reddit
Only if you select the option to include llama.cpp in the settings. The ready build macOS frontend is 306,4 MB on disk. It then needs a OWU server to connect and an inference engine of your choice.
Altruistic_Heat_9531@reddit
is it? damn brb, thanks btw
ouzhja@reddit
Can't you just use whatever engine you want? It says in their page you can connect to any server which means you should also be able to connect to any local inference engine?
mtmttuan@reddit
If I'm using my own endpoint then I don't want to bloat my machine with redundant llama cpp installation.
Patient_Tea_401@reddit
Then you should select the option not to install it in the OWU desktop.
ouzhja@reddit
That's the gripe? An installation that is measured in megabytes, not even gigabytes? Do you know how much other stuff in that entire Open WebUI package and most other software you use is just as "bloaty"? They include it for obvious reason, a "no-hassle one click install & go" approach should include a way to run models as that's what a very large amount of users are going to want from such a package. The llama inclusion is minimal and (from what I can tell just looking at the website) you don't even have to use it... Not really sure why it's even being complained about...
redballooon@reddit
Probably confusing llama.CPP itself with a downloaded model.
ouzhja@reddit
If they forced a model download that would be annoying for sure! Unless it was obviously framed as a "starter model" with an easy way to trash it immediately. I just saw another comment on here that the llama installation is optional so this whole thread is likely pointless anyway lol
screenslaver5963@reddit
The web version comes with a few tiny embedding models. Dunno about the desktop one
ouzhja@reddit
I believe the embedding models are necessary for rag
ClassicMain@reddit
The inference engine is NOT installed/bundled by default, it asks you if you want to install it and you can decline :)
Durian881@reddit
Not necessarily a bad idea. If it's built to support CLIs or be an interface for orchestrating coding assistants and agentic harness, it could be the single installation that provides both front and back end.
screenslaver5963@reddit
I know the web version allows you to pass through an open ai api and ollama api through it self
Patient_Tea_401@reddit
It's optional. You can set OWU server as backend with any inference engine and then connect the desktop as more integrated frontend.
ClassicMain@reddit
Open an issue on the desktop repository as a feature request otherwise this will get lost
DominusIniquitatis@reddit
Is "WebUI" in its name even appropriate then?
-Cacique@reddit
tbf it's Electron
KrazyKirby99999@reddit
"Open" isn't appropriate either
ThePixelHunter@reddit
They've added group channels, reminders, the works. I think they're going for "the everything UI", as in the interface where you do all your work.
Reactor-Licker@reddit
Once this goes out of beta, can I ditch the Docker running instance of Open WebUI connecting to LM Studio on my Windows machine?
Apologies in advance for my ignorance, I only got Open WebUI + LM Studio hosted over my network running a few days ago.
turtleisinnocent@reddit
Friends don't let friends run OpenWebUI.
Velocita84@reddit
Real llama.cpp or still just ollama?
__Maximum__@reddit
Stupid question, it's agnostic
Velocita84@reddit
The post says,
I'm assuming by this they mean it comes bundled with llama.cpp, like it does (did?) with ollama
__Maximum__@reddit
Why would they incude that or ollama? I guess they changed it at some point, my bad.
DinoAmino@reddit
Years ago ollama had a ui. The ui was unbundled from ollama and became OWUI
Iory1998@reddit
Now, this is interesting. I prefer a desktop app.
wowsers7@reddit
Can I connect to any remote openAI compatible API endpoint?
aharongrama@reddit
Thanks for the heads up on the 0.9 update! Quick question: does anyone know if the new Desktop app plays nice with Tailscale yet? I had a lot of trouble trying to get remote connections working properly on 0.8 so I've just been sticking to my headless pip setup. Appreciate any insight!
Prestigious-Use5483@reddit
Anyone know how this compares to Unsloth Studio?
BidWestern1056@reddit
or use one that has already been built for desktop from the start lol
https://github.com/npc-wolrdwide/incognide
vk3r@reddit
I'm unable to connect to my instance. It's protected by PocketID with Pangolin, and I'm having trouble logging in. I don't know how to fix this.
caetydid@reddit
i cant really connect my external llama-server instances. there is no error, but it does not show models or give me a chat window
Salt-Willingness-513@reddit
You can connect to external services within the openwebui settings. Set it up locally and add api url.
caetydid@reddit
yes, I figured this out, thank you!
bonobomaster@reddit
You can only import models via JSON file and that blows!
Just let me pick my GGUF, ask what parameters I want, write the JSON and import that shit...
Steus_au@reddit
amazing :) finally they make it as: opencode web --hostname 0.0.0.0
Darth_Candy@reddit
Why package this with llama.cpp, what makes this different than the llama-server web UI?
ClassicMain@reddit
It's not packaged with llama-server
You can decide whether to install it or not during setup
Salt-Willingness-513@reddit
Unfortunately on fedora kde with appimage i get grey screen on my existing openwebui server and cant set it up locally with unknown error
ClassicMain@reddit
Hi please open an issue on the repo so we can look into it
hl2oli@reddit
Anyone know if it is allowed to customize when used internally for our company. We are not selling a service built on OpenWebUI
Boring_Office@reddit
Yes up to 50 users commercially, and in house unlimited i think? Read the user agreement/license
ketosoy@reddit
It says gpl, which is essentially unrestricted commercially. Â Where are you getting 50 users?
screenslaver5963@reddit
They have a modified version of the license
ketosoy@reddit
This looks unmodified to me https://github.com/open-webui/desktop/blob/main/LICENSE
petuman@reddit
That's license for electron wrapper that launches server of original webapp, which has different license:
https://github.com/open-webui/open-webui/blob/main/LICENSE
yrro@reddit
https://github.com/open-webui/open-webui/blob/main/LICENSE is not compliant with the Open Source Definition due the addition of an advertising clause.
hl2oli@reddit
Very nice. I will check it out. Thanks!