llama-server + qwen (code) : acknowledges tasks but silently stops working , requiring constant nudging.

Posted by Althar93@reddit | LocalLLaMA | View on Reddit | 9 comments

Hey all,

I am new to the world of LLMs, and specifically local LLMs.

I am currently trying to get a stable setup with & qwen code using my local llama-server as the provider. The model I am using is 'gemma-4-e2b-it-Q8_0', because it is small & seems to work really well overall.

---

My issue is that when using qwen, I will prompt the model to perform a task. It will usually do the initial legwork & confirm the request, but then more often than not it tells me it is working on the task, when in fact it just stops & goes idle.

I am able to get it unstuck by continuously nudging it to 'continue' or 'resume work' but it keeps going idle again and again.

---

Any ideas or hints as to what might be causing this? Should I be looking at the model I use, some server setup, or could this simply be because my hardware is too weak for this kind of work (I have an RX 6700XT)