Local coding agents. Am I missing something?

Posted by SpaceKuh@reddit | LocalLLaMA | View on Reddit | 25 comments

I'm an experienced software dev that has been using various LLMs and tools to write code in the past few years. My hardware isn't the greatest for AI with a 4070ti and 64gb ddr5 but I can run a few smaller models. I tried out GemmaE4B, Gemma26b and different devstral models.

In the olama chat window, they work great, especially the smaller models that fit into my vram are incredibly fast. Sure the results cannot compete with frontier models like Gemini, Opus and codex but they are alright. All of that completely falls apart when I use them as coding agents though. I tried them with GitHub Copilot and Continue in VScode and more often than not they would just spin in circles, outright fail and throw errors. Is this the state of local AI currently, where the chat is slowly getting alright but agentic coding is still off the table if you don't have a personal Datacenter at home? I know my hardware isn't optimal but I hear of people running these things on laptops and I have no idea how these agents can compete even with the cheapest commercial models right now.

Did I miss a fundamental step in my setup? (I just installed ollama, installed the models, tried them out, maybe adjusted GPU layers to preserve some vram and added them in continue/Copilot)

Or is this the state of local coding agents right now?

thank you!