LibreChat
Open-source ChatGPT clone with multi-provider support (OpenAI, Anthropic, local LLMs via OpenAI-compatible APIs). The most popular self-hosted ChatGPT-shaped frontend. Strong multi-user + RAG + plugin support; pairs well with vLLM / Ollama / LM Studio backends.
Overview
What LibreChat actually is
LibreChat is the most production-grade self-hosted ChatGPT-shaped frontend in 2026. It started as a faithful clone of the ChatGPT UI but has evolved into something distinctly more capable: multi-provider routing (OpenAI, Anthropic, Google, AWS Bedrock, OpenRouter, plus any OAI-compatible local backend), first-class multi-user authentication, per-user usage tracking, plugin / tool support, RAG over uploaded files, image generation, and presets that let users bookmark prompt + model + parameter combinations.
If Open WebUI is the homelab default and AnythingLLM is the document-first default, LibreChat is the team-first multi-provider default. It's what you reach for when "we have ten people, some prefer Claude, some prefer Llama 3.1 70B local, and we want one auth-gated UI in front of all of them."
Where it fits in the stack
LibreChat is a frontend with serious operational machinery:
- Frontend: React web UI (Mongo + Meilisearch backed)
- Backend services: Node.js API server + Mongo (chats, users, presets) + Meilisearch (search) + optional Redis
- LLM providers (configurable): Ollama, vLLM, LM Studio, OpenAI, Anthropic, AWS Bedrock, Google Vertex, OpenRouter, Groq, custom OAI-shape
- Auth: local accounts, OAuth (Google, GitHub, Discord, etc.), LDAP, OpenID Connect
- Plugins / tools: native plugin system + MCP support landed in 2025
The Mongo dependency is the biggest difference from Open WebUI's SQLite-default — LibreChat is architected for "real" multi-tenant deployments, not laptop convenience.
Best use cases
- Internal team AI gateway. Auth + per-user usage tracking + multi-provider routing in one deploy.
- Mixed local + cloud teams. Some queries hit local Llama 3.1 70B (privacy / cost), some hit Claude (frontier capability).
- Prompt-engineering workbench. The "presets" system saves model + system prompt + parameter combinations, makes prompt iteration tractable.
- Coding / docs assistants behind SSO. OIDC integration means LibreChat fits a corp SSO setup without bolt-on glue.
- Cost-controlled OpenAI replacement with audit trail. Mongo retains all chat history; per-user spend reports are a built-in feature.
OS support
| OS | Quality | Notes |
|---|---|---|
| Linux (Docker) | excellent | the production path |
| Linux (native Node) | excellent | works for dev |
| macOS | excellent | dev or single-host |
| Windows | excellent | dev or single-host |
| Windows (WSL2) | excellent | matches Linux |
LibreChat is genuinely cross-platform for development; for production almost everyone runs Docker Compose with Mongo + Meilisearch + the LibreChat container.
Hardware / runtime support
Like AnythingLLM, LibreChat doesn't do inference. Hardware support is whatever the backend offers:
- NVIDIA CUDA via Ollama / vLLM / LM Studio
- Apple Silicon Metal via Ollama / LM Studio / MLX-LM
- AMD ROCm via Ollama / llama.cpp
- Intel Arc / NPU laptops via Ollama / OpenVINO
- Cloud APIs if local isn't the goal
The frontend itself runs comfortably on a 2-CPU / 2 GB Linux VM; the heavy resource consumer is the backend Mongo if you have many users.
Model / quant format support
Whatever the backend supports. LibreChat passes the configured model name through the OpenAI-compatible API. Format details are someone else's problem.
For the cross-runtime view see /systems/quantization-formats.
Setup path
Docker Compose is the production path:
git clone https://github.com/danny-avila/LibreChat.git
cd LibreChat
cp .env.example .env
# edit .env to point at backends
docker compose up -d
That brings up Mongo, Meilisearch, the API server, and the React frontend. Browse to http://localhost:3080, register the first user (which becomes admin).
To wire in a local Ollama:
ENDPOINTS=openAI
OPENAI_REVERSE_PROXY=http://host.docker.internal:11434/v1
OPENAI_MODELS=llama3.1:8b,qwen2.5:32b
What breaks first
- Mongo connection drift. A Mongo upgrade or auth-change kills the whole stack; back up the Mongo volume before any version bump.
- Backend URL inside Docker.
localhostfrom inside the LibreChat container is the container, not your host — you needhost.docker.internal(Windows / Mac) or the host network IP (Linux). - OAuth callback URLs. Setting up Google / GitHub OAuth requires the exact callback URL match; misconfigured ones produce silent auth failures.
- Plugin / MCP wiring. Plugins with their own auth (browser, code interpreter) need separate config and break on minor version bumps more often than the core does.
- Meilisearch index drift on big chat history. Search-related crashes after schema bumps; usually a re-index fixes it but it's a real ops chore.
Alternatives by intent
| If you want… | Reach for |
|---|---|
| Lighter, single-host, no Mongo | Open WebUI |
| Document-first, workspace UX | AnythingLLM |
| In-IDE assistant | Continue.dev |
| Pure local desktop app | LM Studio |
| Customizable React app | fork LibreChat itself — it's MIT-licensed |
Best pairings
- Ollama for local + OpenAI/Anthropic for cloud — the canonical hybrid setup
- vLLM + 32B AWQ-INT4 + LibreChat — the team-grade local-only deployment
- OIDC SSO + LibreChat + local backend — the compliance-friendly internal-AI gateway
- Apple M3 Ultra + MLX-LM 70B + LibreChat — the high-VRAM Mac team backend
Who should avoid LibreChat
- Single-user laptops with no team need. Mongo overhead isn't worth it; use Open WebUI.
- Document-RAG-first operators. Use AnythingLLM; LibreChat's RAG is fine but not the focus.
- Teams that want a managed product. Self-hosting LibreChat is real ops work; if that's not viable, a hosted ChatGPT Team / Claude Team plan is honest about the trade.
- Operators on hostile networks where Mongo can't run. Architectural mismatch.
Related
- Stacks: /stacks/local-coding-agent, /stacks/private-rag-laptop
- System guides: /setup, /compatibility
- Tools: Ollama, vLLM, Open WebUI
- Hardware: RTX 4090, Apple M3 Ultra, RTX 3090
Pros
- Multi-provider out of the box — local + cloud APIs in one UI
- First-class multi-user with auth + per-user usage tracking
- Active development + large community
Cons
- Heavier deployment than Open WebUI (Mongo dependency)
- RAG path narrower than AnythingLLM for document-grounding workflows
- Plugin ecosystem trails ChatGPT and OpenAI's mainline
Compatibility
| Operating systems | Windows macOS Linux Docker |
| GPU backends | n/a — frontend |
| License | Open source · free + open-source |
Runtime health
Operator-grade signals on how actively LibreChat is being maintained, how fresh its measurements are, and what failure classes operators have flagged. Every label below is anchored to a real date or count — we never infer maintainer activity we can't show.
Release cadence
Derived from the most recent editorial signal on this row.
6 days since last refresh · source: lastUpdated
Benchmark freshness
How recent the editorial measurements on this runtime are.
No editorial benchmarks for this runtime yet.
Community reproduction
Submissions that match an editorial measurement on similar hardware.
No community reproductions on file yet.
Get LibreChat
Frequently asked
Is LibreChat free?
What operating systems does LibreChat support?
Which GPUs work with LibreChat?
Reviewed by RunLocalAI Editorial. See our editorial policy for how we evaluate tools.
Related — keep moving
Verify LibreChat runs on your specific hardware before committing money.