Tracked my Ollama Cloud free tier usage to estimate Pro and Max quotas. Anyone else done this?
Posted by nusquama@reddit | LocalLLaMA | View on Reddit | 1 comments
I tried to figure out Ollama Cloud's actual quotas. Here's what I found (take it with a grain of salt)
Ollama Cloud doesn't publish how many tokens you get on each plan. They just say Pro is "50x more" and Max is "5x more than Pro." But 50x of what? No idea.
So I ran a session on the free tier and watched my usage dashboard. Here's what I observed:
- 2M tokens consumed = 100% of my 5h session quota and 40% of my weekly quota
If I take those numbers at face value and apply the official multipliers:
| Plan | Price | 5h session (est.) | Weekly (est.) |
|---|---|---|---|
| Free | $0 | \~2M tokens | \~5M tokens |
| Pro | $20/mo | \~100M tokens | \~250M tokens |
| Max | $100/mo | \~500M tokens | \~1.25B tokens |
Now, big disclaimers:
- This is based on MY usage with ONE model in ONE session. I have no idea if this generalizes.
- Ollama measures GPU time, not tokens. So a bigger model would eat through the same quota faster with fewer tokens.
- I counted tokens on my end, Ollama counts GPU time on theirs. These might not map cleanly.
- I could be completely wrong. This is just what I observed.
If anyone else has tracked their usage, I'd love to compare notes. Would be interesting to see if these numbers hold across different models and session patterns.
Either way, even if my estimates are off by 2x, Pro at $20/mo seems like it would cover most individual usage easily.
Equivalent_Job_2257@reddit
Man, this is localllama, not cloud llama