Tired of unpredictable API bills from agents? Here’s a 0-dep MCP server to estimate costs in real-time.
Posted by Pitiful-Hearing-5352@reddit | LocalLLaMA | View on Reddit | 1 comments
Been running some agent workflows lately and got hit with unexpected API costs.
Tried a few tools but most were either overkill or needed extra setup just to estimate tokens.
So I made a small MCP server that just estimates cost before the call.
No deps, just stdin/stdout.
Example:
gpt-4o (8k in / 1k out) → \~$0.055
Gemini flash → way cheaper
Repo: https://github.com/kaizeldev/mcp-cost-estimator
Curious how others are handling this?
ttkciar@reddit
This is off-topic for LocalLLaMA. You might want to post instead to r/LLM or r/LLMDevs.