Air-gappable RAG over your docs. The OG offline-RAG project, now mature and team-friendly.
Editorial verdict: “Best when air-gap compliance is the requirement. Less polished than AnythingLLM, more configurable.”
Which runtime + OS combos this app works against. Source of truth for "will it run on my setup?"
PrivateGPT is the OG project for 'chat with your docs, fully offline.' Mature ingestion pipeline, multiple UI options (FastAPI server + Gradio + headless), and a swappable LLM + embedder backend. Choose this when reproducible offline RAG matters more than UI polish.
Document retrieval + chat, fully offline-capable.
Pre-filled with this app's recommended use case + budget tier. Get the full rig + runtime + model picks.
The full directory — filter by category, runtime, OS, privacy posture, or VRAM.
What this app talks to: Ollama, vLLM, llama.cpp, MLX, LM Studio. The upstream layer.
Did this app work for you on a specific rig? Submit the benchmark — it powers the model + hardware pages.