RUNLOCALAIv38
→WILL IT RUNBEST GPUCOMPARETROUBLESHOOTSTARTPULSEMODELSHARDWARETOOLSBENCH
RUNLOCALAI

Operator-grade instrument for local-AI hardware intelligence. Hand-written verdicts. Real benchmarks. Reproducible commands.

OP·Fredoline Eruo
DIR
  • Models
  • Hardware
  • Tools
  • Benchmarks
  • Will it run?
GUIDES
  • Best GPU
  • Best laptop
  • Best Mac
  • Best used GPU
  • Best budget GPU
  • Best GPU for Ollama
  • Best GPU for SD
  • AI PC build $2K
  • CUDA vs ROCm
  • 16 vs 24 GB
  • Compare hardware
  • Custom compare
REF
  • Systems
  • Ecosystem maps
  • Pillar guides
  • Methodology
  • Glossary
  • Errors KB
  • Troubleshooting
  • Resources
  • Public API
EDITOR
  • About
  • About the author
  • Changelog
  • Latest
  • Updates
  • Submit benchmark
  • Send feedback
  • Trust
  • Editorial policy
  • How we make money
  • Contact
LEGAL
  • Privacy
  • Terms
  • Sitemap
MAIL · MONTHLY DIGEST
Get monthly local AI changes
Monthly recap. No spam.
DISCLOSURE

Some links on this site are affiliate links (Amazon Associates and other first-class retailers). When you buy through them, we earn a small commission at no extra cost to you. Affiliate links do not influence our verdicts — there are cards we rate highly that we don't have affiliate relationships with, and cards that sell well that we refuse to recommend. Read more →

SYS · ONLINEUPTIME · 100%2026 · operator-owned
RUNLOCALAI · v38
← Home·Runtimes (/tools)

>Apps directory

What plugs into your local AI runtime. 37 curated apps across 12 categories — chat UIs, coding agents, RAG pipelines, voice, image, browser extensions, editor plugins, mobile + desktop, agent frameworks, productivity, SDK wrappers.

Each entry carries an honest editorial verdict — pros, cons, the runtime it works against, the minimum VRAM, and the privacy posture. Filter to your stack, jump to the detail page, ship.

Filter the directory

URL updates as you change filters — share or bookmark a result. All filters are server-rendered, so the page works without JS.

§ Category
§ Sort by

5 of 37 apps matching your filters

Aider

Coding agent
★ 4.7
Fully offline

Terminal coding agent that edits files via your local model. Git-aware, surgical, fast.

“Best terminal-native coding agent for local models. Qwen 2.5 Coder 32B is its sweet spot.”

ollamaopenai-compat
Free· 24GB+ VRAM
★ 27.0k

Cline

Coding agent
★ 4.6
Hybrid (offline or cloud)

VS Code extension that runs a full agent loop locally — reads, writes, runs commands, asks first.

“Best IDE-integrated agent that fully respects 'all local' as a first-class option.”

ollamaopenai-compatanthropicopenai
Free· 24GB+ VRAM
★ 30.0k

Continue

Coding agent
★ 4.5
Hybrid (offline or cloud)

Open-source autocomplete + chat for VS Code and JetBrains. Local-model-first.

“Best Copilot replacement that defaults to local. Configurable; pair with Qwen 2.5 Coder.”

ollamaopenai-compatanthropicopenai
Free· 12GB+ VRAM
★ 21.0k

Tabby

Coding agent
★ 4.4
Fully offline

Self-hosted coding agent server with team SSO, audit logs, and dashboards. Enterprise-grade.

“Best self-hosted server for teams. SSO + audit logs make it the IT-friendly pick.”

llama-cppopenai-compat
Freemium· 16GB+ VRAM
★ 23.0k

Twinny

Coding agent
★ 4.2
Fully offline

Free, lightweight VS Code copilot that runs entirely on Ollama. Strong on autocomplete.

“Best minimal-surface Copilot-replacement that's been Ollama-native since day one.”

ollama
Free· 8GB+ VRAM
★ 3.5k

Missing an app? Suggest one.

We curate this directory editorially — same review queue as the benchmarks feed. Open an issue with the project link and your one-line pitch for why it belongs.

Open a GitHub issue →Or contribute a benchmark →

Editorial review applies — same standards as the rest of the site. We won't list apps that don't actually work against a local runtime, regardless of marketing claims.

Where to go from here

Runtimes (/tools) →

The runtime layer: Ollama, vLLM, llama.cpp, MLX, LM Studio server, ComfyUI. What apps in this directory talk to.

Stack Builder →

Tell us your use case, get a full rig recipe — runtime + models + the apps from this directory that fit your stack.

GPU chooser →

Match an app's minimum-VRAM requirement to real hardware with our price/perf comparison.

Community benchmarks →

Real operator submissions on the model × hardware × app combos that work. The proof behind the editorial picks.