Are Agents even useful with all local models?

Posted by bsawler@reddit | LocalLLaMA | View on Reddit | 21 comments

I've been trying to step up my usage and try out all the new toys over the past weeks. It feels like I've been jumping from thing to thing to thing.

Claude Code (with local LLM), OpenClaw, Hermes, Pi, Paperclip, etc.

Are there ANY of them that actually "just work" with local LLMs? With the exception of Pi, which is super-restrictive by default, all of the rest just tend to be failure after failure after failure for every task I give them that isn't just "write a markdown document" or "write a bit of code in language X".

Claude was able to (extremely slowly, like 1/10th the speed of Pi) generate some python that was passable. But anything beyond simple document reading/writing/editing would fail because it expected Anthropics various services.

OpenClaw failed non-stop at any task I gave it beyond simple chatting (which if I'm just going to chat, I don't need an agentic harness!) unless I go install a bunch of security-risk-ridden software that's going to do god-knows-what on my network.

Hermes would (sometimes) show up in Discord / Slack. But half of their functionality would fail - sure it could generate a document, and even got it to talk to my local ComfyUI to generate a (truly horrible looking) image, but it couldn't actually pin it to Slack or Discord which means I had no way to getting anything from it short of breaking into the docker's storage and doing a manual exfil operation...

And then lastly Paperclip yay my CEO hired a CTO and CMO... and they both immediately failed their tasks and every issue I file against any of my AI "employees" would end up spinning and failing to complete anything.

All of this is across a number of models on my Strix Halo system (so 128gb, 112gb usable as vram): Qwen 3.5, 3.6, Qwen 3 Coder Next, Llama 3.3 70b, GPT-OSS 120b, GLM 4.7 Flash, Gemma 4 31b and e4b.

I'm 100% willing to believe I'm just dumb and missing something... but after weeks of trying different tools and running into similar issues over and over again... is this just where we're at for local AI? We can locally host all the agents but that means nothing if you still have to sign up for countless subscriptions and pass all the data to outside services, which is the entire reason I (and many of you, I suspect) am wasting all this money on local AI hardware to begin with.