Local AI coding assistant that runs fully offline (Gemma 4, codebase-aware)

Posted by andres_garrido@reddit | LocalLLaMA | View on Reddit | 11 comments

I’ve been experimenting with running a local coding assistant on Gemma 4 26B, focused on understanding full codebases instead of single-file prompts.

Main idea:

- build a project map (files, symbols, structure)

- run a planning step to decide which files matter

- then retrieve full files + semantic chunks before answering

Goal is to avoid the usual “chat with files” limitation and make it reason about structure first.

Runs fully local (llama.cpp, GGUF), no network calls during inference.

Curious if others here are doing something similar or handling codebase reasoning differently with local models.