Open-source RAG routes are splintering — MiniRAG, Agent-UniRAG, SymbioticRAG… which one are you actually using?
Posted by Cheryl_Apple@reddit | LocalLLaMA | View on Reddit | 8 comments
I’ve been poking around the open-source RAG scene and the variety is wild — not just incremental forks, but fundamentally different philosophies.
Quick sketch:
- MiniRAG: ultra-light, pragmatic — built to run cheaply/locally.
- Agent-UniRAG: retrieval + reasoning as one continuous agent pipeline.
- SymbioticRAG: human-in-the-loop + feedback learning; treats users as part of the retrieval model.
- RAGFlow / Verba / LangChain-style stacks: modular toolkits that let you mix & match retrievers, rerankers, and LLMs.
What surprises me is how differently they behave depending on the use case: small internal KBs vs. web-scale corpora, single-turn factual Qs vs. multi-hop reasoning, and latency/infra constraints. Anecdotally I’ve seen MiniRAG beat heavier stacks on latency and robustness for small corpora, while agentic approaches seem stronger on multi-step reasoning — but results vary a lot by dataset and prompt strategy.
There’s a community effort (search for RagView on GitHub or ragview.ai) that aggregates side-by-side comparisons — worth a look if you want apples-to-apples experiments.
So I’m curious from people here who actually run these in research or production:
- Which RAG route gives you the best trade-off between accuracy, speed, and controllability?
- What failure modes surprised you (hallucinations, context loss, latency cliffs)?
- Any practical tips for choosing between a lightweight vs. agentic approach?
Drop your real experiences (not marketing). Concrete numbers, odd bugs, or short config snippets are gold.
donotfire@reddit
Reads like AI
EndlessZone123@reddit
Op's entire post history is to promote ragview.
milksteak11@reddit
Colberrrr
teh_spazz@reddit
This is an ad.
Appropriate-Law8785@reddit
I don't see people talk about text processing and chunking, which I think it's the most important step.
No-Refrigerator-1672@reddit
I am utilizing RAG for scientific paper processing, specifically in physics field. So far I've experimented with two systems: LightRAG and RAG Flow. All of them were deployed with local AI soultions: Colnomic-embed-multimodal-7b for embedding, and Qwen3 32B or Mistral 3.2 for models. LightRAG (graph-based system) is very impressive in structuring the data, but suffers quite hardly from inability to deduplicate entities, and spends ungodly amounts of tokens for paper processing, as well as it's default UI doesn't allow to manage multiple databases, at least it didn't 4 months ago. RAG Flow is better at managing multiple databases, you can seperate knowledge by topics and select which ones to use. It has multiple document parsing models, and, hilariously, the "paper" mode advertised for scientific papers fails miserably at structuring them, but default works just fine. I am yet to test RAG Flows ability to create graph-based knowledge bases, but so far I prefer it over LightRAG due to convinience of usage.
Cheryl_Apple@reddit (OP)
Based on my understanding, RAGFlow should support the LightRAG preprocessing approach.
No-Refrigerator-1672@reddit
In my understanding, for preprocessing they use something else. If they do use LightRAG, then it's for knowledge graph construction. However, I've deployed RAG Flow just last week so I can't comment on graph mode yet.