Best coding/reasoning model for low vram
Posted by XEUIPR@reddit | LocalLLaMA | View on Reddit | 5 comments
I'm trying to train my own llm for specifically and only coding with complex java algorithms. I have already tried qwen2.5 7b, but that was obviously too much. Are there any good model recommendations for my case? The dataset is 7500 rows and I'm training using unsloth.
I'd also be using low context length (1024-2048)
ttkciar@reddit
What is your GPU?
XEUIPR@reddit (OP)
Thanks, im using the rtx 3050. (college laptop)
BitGreen1270@reddit
Dude I can run Qwen3.6-35b and gemma4-26B on my 780m igpu laptop with 32gb ram and get 20 t/s. You should be able to as well. Just use the right quantization level. I would start with Q4 and go higher or lower. Check out unsloth or bartowski modelsĀ
Velocita84@reddit
Just use whatever qwen3.5/3.6 fits your hardware rather than trying to train a model for coding, you can't beat real labs at that
Guilty_Spray_6035@reddit
gemma-4-e4b, qwen-3.5-9b