What plugs into your local AI runtime. 37 curated apps across 12 categories — chat UIs, coding agents, RAG pipelines, voice, image, browser extensions, editor plugins, mobile + desktop, agent frameworks, productivity, SDK wrappers.
Each entry carries an honest editorial verdict — pros, cons, the runtime it works against, the minimum VRAM, and the privacy posture. Filter to your stack, jump to the detail page, ship.
URL updates as you change filters — share or bookmark a result. All filters are server-rendered, so the page works without JS.
5 of 37 apps matching your filters
Terminal coding agent that edits files via your local model. Git-aware, surgical, fast.
“Best terminal-native coding agent for local models. Qwen 2.5 Coder 32B is its sweet spot.”
VS Code extension that runs a full agent loop locally — reads, writes, runs commands, asks first.
“Best IDE-integrated agent that fully respects 'all local' as a first-class option.”
Open-source autocomplete + chat for VS Code and JetBrains. Local-model-first.
“Best Copilot replacement that defaults to local. Configurable; pair with Qwen 2.5 Coder.”
Self-hosted coding agent server with team SSO, audit logs, and dashboards. Enterprise-grade.
“Best self-hosted server for teams. SSO + audit logs make it the IT-friendly pick.”
Free, lightweight VS Code copilot that runs entirely on Ollama. Strong on autocomplete.
“Best minimal-surface Copilot-replacement that's been Ollama-native since day one.”
We curate this directory editorially — same review queue as the benchmarks feed. Open an issue with the project link and your one-line pitch for why it belongs.
Editorial review applies — same standards as the rest of the site. We won't list apps that don't actually work against a local runtime, regardless of marketing claims.
The runtime layer: Ollama, vLLM, llama.cpp, MLX, LM Studio server, ComfyUI. What apps in this directory talk to.
Tell us your use case, get a full rig recipe — runtime + models + the apps from this directory that fit your stack.
Match an app's minimum-VRAM requirement to real hardware with our price/perf comparison.
Real operator submissions on the model × hardware × app combos that work. The proof behind the editorial picks.