I wonder how good the Qwen 3.6 4B will be given the insane boost of performance in the 27B and 36B

Posted by exaknight21@reddit | LocalLLaMA | View on Reddit | 13 comments

I personally am a simpleton with crappy hardware. I run the Qwen 3 4B still for my simple tasks for simple RAG. I personally cannot wait for the 4B Instruct model as I believe it’s my go to “ChatGPT” replacement for dumb question via OpenWebUI and vLLM.

I rock an old T5610, DDR 3 - 64 GB Dual Xeon (sadly AVX) slow processors, 256 GB Sata SSD and an Mi50 32 GB

I run dockerized vLLM (nlzy archived so on the sweet mobydick branch), i run my in-home experiments and use 8K contexr, usually cyankiwi’s awq version, it does wonders for me.

I pray the Qwen team releases this soon!