Qwen3.6 35b-a3b 🤯

Posted by EffectiveMedium2683@reddit | LocalLLaMA | View on Reddit | 116 comments

Originally I was a diehard fan of Gemma4 26b-a4b because it really is a remarkably intelligent llm. Ran qwen3.6 via ollama and found it impressive but still favored Gemma. Ollama did it a disservice at least on my pc.

Ran it straight through llama.cpp and it is much faster than gemma4 26b-a4b, roughly equivalent in general intelligence, better in strict prompt adherence, and it doesn't slow down on long context. Like, I'm back to being a Qwen fan.

Just thought I'd share haha