Best Agentic pure coding llm for 32gb ddr5 ram and 8gb vram?

Posted by LightH12@reddit | LocalLLaMA | View on Reddit | 12 comments

i'm a little lost on what model to use for Pure coding Agent, i am using LM Studio with Continue CL,
i want to move out of using Gemmini CLI, or at least use something local when my tokens run out, so please don't mention anything online
i have an i7 12650H, 32GB DDR5 RAM (Dual channel), 4060 8GB Mobile. i also want to keep using the device when running the llm since i am coding on it (expect it do run a localhost for my website and intellij so nothing major)
i've looked into Omnicoder, qwen 3.5.

i tried gemmaE4B 7b but let's say it's too dumb to even add Hi world! into an html i have in my project

Speed itself isn't an issue i am using it for casual programming, but i'd at least want it to finish a simple basic task in less than 5min (like add hello work to x.html)

so how many Billion params should i aim and what models? please leave your opinion