Gemma 4 26B A4B just doesn't want to finish the job... or is it me?

Posted by boutell@reddit | LocalLLaMA | View on Reddit | 25 comments

I've tried Gemma 4 26B A4B under both OpenCode and Claude Code now, on an M2 Macbook Pro with 32GB RAM. Both times using Ollama 0.20.2, so yes, I have the updates that make Ollama Gemma 4 compatible.

I gave it a meaty job to do, one that Opus 4.6 aced under Claude Code last week. Straightforward adapter pattern — we support database "A," now support database "B" by generating a wrapper that implements a subset of the database "A" API. Piles of unit tests available, tons of examples of usage in the codebase. I mention this because it shows the challenge is both nontrivial and well-suited to AI.

At first, with both Claude Code and OpenCode, Gemma 4 made some progress on planning, wrote a little code, and... just gave up.

It would announce its progress thus far, and then stop. Full stop according to both the CPU and the GPU.

After giving up, I could get it to respond by talking to it, at which point the CPU and GPU would spin for a while to generate a response. But it wouldn't do anything substantive again. I had very silly conversations in which Gemma 4 would insist it was doing work, and I would point out that the CPU and GPU progress meters indicate it isn't, and so on.

Finally this last time in OpenCode I typed:

"No, you're not. You need to start that part of the work now. I can see the CPU and GPU progress meters, so don't make things up."

And now it's grinding away generating code, with reasonably continuous GPU use. Progress seems very slow, but at least it's trying.

For a while I saw code being generated, now I see ">true" once every minute or two. Test runs perhaps.

Is this just life with open models? I'm spoiled, aren't I.