I've got a feeling that Llamacpp is not the biggest performance bottleneck, but it might be the OpenCode.

Posted by ThingRexCom@reddit | LocalLLaMA | View on Reddit | 32 comments

I've got a feeling that Llamacpp is not the biggest performance bottleneck, but it might be the OpenCode.

It looks as if OpenCode introduces an artificial delay in agentic coding. Have you noticed similar issues?

Could you suggest other solutions that provide better results with the local Llama server?