Reliable Open Source LLM as a Service

Posted by pravictor@reddit | LocalLLaMA | View on Reddit | 9 comments

Has anyone figured out a provider whose open source models (Kimi, Qwen, GLM e.t.c) can be used reliably in production.

I have tested some well known providers and they all suffer from high latency and poor uptime rendering them mostly useless for production implementation.

I am using them for an agentic workflow in production so reliability and low latency are very important for me.

Is there no provider that compares to Gemini / Claude in reliability but with open source models?

So far tested Teogether.ai and Fireworks and Groq looks like it is dying