RUNLOCALAIv38
→WILL IT RUNBEST GPUCOMPARETROUBLESHOOTSTARTPULSEMODELSHARDWARETOOLSBENCH
RUNLOCALAI

Operator-grade instrument for local-AI hardware intelligence. Hand-written verdicts. Real benchmarks. Reproducible commands.

OP·Fredoline Eruo
DIR
  • Models
  • Hardware
  • Tools
  • Benchmarks
  • Will it run?
GUIDES
  • Best GPU
  • Best laptop
  • Best Mac
  • Best used GPU
  • Best budget GPU
  • Best GPU for Ollama
  • Best GPU for SD
  • AI PC build $2K
  • CUDA vs ROCm
  • 16 vs 24 GB
  • Compare hardware
  • Custom compare
REF
  • Systems
  • Ecosystem maps
  • Pillar guides
  • Methodology
  • Glossary
  • Errors KB
  • Troubleshooting
  • Resources
  • Public API
EDITOR
  • About
  • About the author
  • Changelog
  • Latest
  • Updates
  • Submit benchmark
  • Send feedback
  • Trust
  • Editorial policy
  • How we make money
  • Contact
LEGAL
  • Privacy
  • Terms
  • Sitemap
MAIL · MONTHLY DIGEST
Get monthly local AI changes
Monthly recap. No spam.
DISCLOSURE

Some links on this site are affiliate links (Amazon Associates and other first-class retailers). When you buy through them, we earn a small commission at no extra cost to you. Affiliate links do not influence our verdicts — there are cards we rate highly that we don't have affiliate relationships with, and cards that sell well that we refuse to recommend. Read more →

SYS · ONLINEUPTIME · 100%2026 · operator-owned
RUNLOCALAI · v38
← Home·/apps·RAG app

Khoj

Hybrid (offline or cloud)

Self-hosted AI assistant for your notes, emails, docs. Web + mobile + desktop, all local-first.

Editorial verdict: “Best 'AI second brain' app. Self-hosted, local-first, works against Obsidian.”

RAG app
Freemium
AGPL-3.0
★ 4.3 / 5
GitHub ★ 18,000
↗ Homepage↗ GitHub↗ Docs

Compatibility at a glance

Which runtime + OS combos this app works against. Source of truth for "will it run on my setup?"

§ Runtimes supported
ollamallama-cppopenai-compatanthropic
§ OS / platform
linuxmacoswindowsiosandroidweb
§ Hardware + model hint
Minimum VRAM
8 GB
Recommended starter model
Llama 3.1 8B Q4_K_M
→ Build the rest of the stack with /stack-builder→ Pick a GPU for this app

What it is

Khoj is a self-hosted AI second-brain. Indexes your notes (Obsidian, Org-mode, Markdown), emails, browser history, PDFs — then chat with it. Local-first: runs against Ollama or llama.cpp by default, falls back to cloud only if you opt in. Cross-platform (web, iOS, Android, desktop). Niche but well-executed.

✓ Strengths

  • +Genuinely cross-platform (mobile apps work)
  • +Strong Obsidian integration
  • +Self-host + cloud sync is well-designed

△ Caveats

  • −Initial index of a large corpus is slow
  • −Some niche features (image generation) require cloud

About the RAG app category

Document retrieval + chat, fully offline-capable.

§ Other rag app apps
PrivateGPT

Best when air-gap compliance is the requirement. Less polished than AnythingLLM, more configurable.

Verba

Best for 'don't make me choose chunking strategy' teams. Opinionated stack works.

Where to go from here

Stack Builder →

Pre-filled with this app's recommended use case + budget tier. Get the full rig + runtime + model picks.

Back to /apps →

The full directory — filter by category, runtime, OS, privacy posture, or VRAM.

Runtimes (/tools) →

What this app talks to: Ollama, vLLM, llama.cpp, MLX, LM Studio. The upstream layer.

Community benchmarks →

Did this app work for you on a specific rig? Submit the benchmark — it powers the model + hardware pages.