Local LLM users struggle to build reliable private knowledge bases
A LocalLLaMA thread reveals practitioners want private RAG setups for life docs, not code, but face retrieval trust issues, context limits, and tooling churn that make daily use a maintenance burden.

Practitioners running local LLMs say the gap between experimental setups and production-grade personal workflows remains wide. A LocalLLaMA user this week asked whether anyone is running a local LLM as a daily personal knowledge base — not for coding or chat experiments, but to query their own notes, PDFs, and documents privately.
The poster described hitting a wall: most resources assume a developer audience building something new, or they're two years old and recommend tools that have since changed. The specific pain points named include model choice for RAG on consumer hardware, retrieval trust ("do you double check everything because hallucinations?"), tooling confusion (LlamaIndex vs Ollama vs newer alternatives), and context length management as personal document collections grow.
The question isn't whether local RAG is technically possible — it is. The question is whether anyone has made it work without it becoming "a part time job to maintain." Quantization tradeoffs, retrieval accuracy, and the fast churn of RAG tooling all push against the "set it and forget it" experience that a true daily knowledge base would require.
No clear consensus emerged in the replies. Some users pointed to Obsidian plugins or self-hosted vector databases, others admitted they'd tried and reverted to manual search. The underlying constraint is the same one that affects all local inference: consumer hardware limits both model size and the size of the retrieval corpus you can realistically keep in VRAM or RAM. Context length helps, but a 128k-token window still forces chunking and embedding for anything beyond a few dozen documents.
The thread is a snapshot of where local RAG stands in mid-2025 — technically feasible, operationally fragile, and still waiting for tooling that makes private knowledge bases as reliable as a file browser.