your daily driver stack, what's it look like? and why?

Posted by Pyrenaeda@reddit | LocalLLaMA | View on Reddit | 12 comments

What it says in the title, I'm interested in hearing what you all have landed on as a workable / useful stack for you.

Mine looks like this:

    back end inference servers - llama.cpp, vLLM
                       |
                       V
hermes-agent - cron jobs + OpenAI compatible endpoints
                       |
                       V
     home-grown web UI & iOS / Swift client

I landed on this for a couple reasons:

- I have test driven a bunch of the go-to front ends - Open WebUI, LobeHub, Libera Chat etc. Couldn't get behind them. Too many knobs and too many features. I don't mind lots of knobs but I don't want them in my chat UI. For that I'm looking for a slick and simple experience similar to ChatGPT and Claude UI (the chat side, not cowork). Plus I hate that they don't have good native mobile apps with streaming support. A slick mobile friendly experience is a must-have for me, and the solution of just dropping a shortcut to the mobile version of the web UI on my homescreen doesn't quite cut it.

- hermes-agent comes with a very nice and extensive packet of tools right out of the box, which really cuts down on the number of MCPs one needs. And cron jobs for agentic background work are great to have of course. I couldn't get behind using a messenger app as my primary "chat assistant" UI though for one main reason: it doesn't work for me to not be able to have multiple conversations running with an assistant at once and jump around between them.

So, that landed me where I am. couple of hermes-agent instances: one for background agentic work (for which I use one of the messenger apps as a control interface) and one as an AI assistant, that I interface with through my vibe coded POS-but-pretty web UI and iOS client using the hermes OpenAI compatible API.

How bout you all? OWUI + llama? straight hermes-agent / OpenClaw / etc? llama.cpp web UI and done? something more exotic / esoteric? rationale? lemme hear it.