Help needed: Ollama > qwen3.6 in OpenCode on 64Gb M4

Posted by Konamicoder@reddit | LocalLLaMA | View on Reddit | 10 comments

Hi Ollama team!

I’d love to get your advice as to why I’m doing wrong. In running Ollama on an M4 MacBook Pro with 64Gb RAM. Am trying to use OpenCode with qwen3.6-35b-a3b-q4_K_M as the selected model. I made a modelfile version of the model with the following parameters:

PARAMETER num_ctx 32768

PARAMETER num_predict 4096

PARAMETER temperature 0.6

PARAMETER top_k 20

PARAMETER top_p 0.95

PARAMETER min_p 0.0

PARAMETER repeat_penalty 1.0

PARAMETER repeat_last_n 64

I figure a context length of 32K should be fine for my system with 64Gb RAM.

But when I launch OpenCode with this command…

ollama launch opencode —model qwen3.6-35b-a3b-q4_K_M

…and issue a simple cd command to focus OpenCode on my project folder, RAM instantly pegs to 100 percent, and the system locks up. Mouse cursor starts stuttering across the screen. Activity monitor shows two instances of Ollama chewing up 30Gb and 15Gb of my available RAM. I have to force quit Ollama for the system to calm down.

Based on the details I have shared, can someone help me detect the root cause of the issue? Even better, suggest a fix?

Thanks in advance!