Qwen releases official MLX quants for Qwen3 models in 4 quantization levels: 4bit, 6bit, 8bit, and BF16

Posted by ResearchCrafty1804@reddit | LocalLLaMA | View on Reddit | 45 comments

Qwen releases official MLX quants for Qwen3 models in 4 quantization levels: 4bit, 6bit, 8bit, and BF16

🚀 Excited to launch Qwen3 models in MLX format today!

Now available in 4 quantization levels: 4bit, 6bit, 8bit, and BF16 — Optimized for MLX framework.

👉 Try it now!

X post: https://x.com/alibaba_qwen/status/1934517774635991412?s=46

Hugging Face: https://huggingface.co/collections/Qwen/qwen3-67dd247413f0e2e4f653967f