Build: NVIDIA RTX 4090 48GB (China-mod) + — + 32 GB RAM (windows)
Ranked by fit for reasoning use case + predicted speed. Click a row for VRAM breakdown.
ollama run deepseek-r1:7bollama run phi4-reasoning:14bollama run deepseek-r1:14bollama run phi4:14bTight VRAM, partial CPU offload, or context-limited.
ollama run deepseek-r1:32bollama run qwq:32bollama run deepseek-r1:70bollama run nemotron3:nanoollama run nemotron:70bollama run qwen3:30bHypothetical scenarios. We re-ran the compatibility engine for each.
~$80–150
Doubles your CPU-offload working set. Helps when models don't quite fit in VRAM.
Unlocks: 36 new comfortable, 28 new tradeoff
see current pricing
80 GB VRAM (vs your 48 GB) plus a bandwidth jump from ~1008 GB/s to ~3350 GB/s.
Unlocks: 59 new comfortable
~$2400
Tensor parallelism splits the model across both cards, effectively doubling VRAM. Bandwidth doesn't double — runs ~1.5× the single-card speed in practice.
Unlocks: 65 new comfortable
Some links above are affiliate links. We may earn a commission at no extra cost to you. How we make money.
Need more memory than you have. Shown for orientation.
Even with CPU offload, needs more memory than your VRAM (48 GB) + 60% of system RAM (19 GB) combined.
Even with CPU offload, needs more memory than your VRAM (48 GB) + 60% of system RAM (19 GB) combined.
Even with CPU offload, needs more memory than your VRAM (48 GB) + 60% of system RAM (19 GB) combined.
Even with CPU offload, needs more memory than your VRAM (48 GB) + 60% of system RAM (19 GB) combined.
Even with CPU offload, needs more memory than your VRAM (48 GB) + 60% of system RAM (19 GB) combined.
Want a specific benchmark we don't have? Email support@runlocalai.co and we'll prioritize it.