Any local llm for mid GPU

Posted by kellyjames436@reddit | LocalLLaMA | View on Reddit | 18 comments

Hey, recently tried Gemma4:9b and Qwen3.5:9b running on my RTX 4060 on a laptop with 16GB ram, but it’s so slow and annoying.

Is there any local llm for coding tasks that can work smoothly on my machine?