Is anyone able to successfully run Qwen 30B Coder BF16?

Posted by TokenRingAI@reddit | LocalLLaMA | View on Reddit | 20 comments

With Llama.cpp and the Unsloth GGUFs for Qwen 30B Coder BF16, I am getting frequent crashes on two entirely different systems, a Ryzen AI Max, and a RTX 6000 Blackwell.

Llama.cpp just exits with no error message after a few messages.

VLLM works perfectly on the Blackwell with the official model from Qwen, except tool calling is currently broken, even with the new qwen 3 tool call parser which VLLM added. So the tool call instructions just end up in the chat stream.