RUNLOCALAIv38
→WILL IT RUNBEST GPUCOMPARETROUBLESHOOTSTARTPULSEMODELSHARDWARETOOLSBENCH
RUNLOCALAI

Operator-grade instrument for local-AI hardware intelligence. Hand-written verdicts. Real benchmarks. Reproducible commands.

OP·Fredoline Eruo
DIR
  • Models
  • Hardware
  • Tools
  • Benchmarks
  • Will it run?
GUIDES
  • Best GPU
  • Best laptop
  • Best Mac
  • Best used GPU
  • Best budget GPU
  • Best GPU for Ollama
  • Best GPU for SD
  • AI PC build $2K
  • CUDA vs ROCm
  • 16 vs 24 GB
  • Compare hardware
  • Custom compare
REF
  • Systems
  • Ecosystem maps
  • Pillar guides
  • Methodology
  • Glossary
  • Errors KB
  • Troubleshooting
  • Resources
  • Public API
EDITOR
  • About
  • About the author
  • Changelog
  • Latest
  • Updates
  • Submit benchmark
  • Send feedback
  • Trust
  • Editorial policy
  • How we make money
  • Contact
LEGAL
  • Privacy
  • Terms
  • Sitemap
MAIL · MONTHLY DIGEST
Get monthly local AI changes
Monthly recap. No spam.
DISCLOSURE

Some links on this site are affiliate links (Amazon Associates and other first-class retailers). When you buy through them, we earn a small commission at no extra cost to you. Affiliate links do not influence our verdicts — there are cards we rate highly that we don't have affiliate relationships with, and cards that sell well that we refuse to recommend. Read more →

SYS · ONLINEUPTIME · 100%2026 · operator-owned
RUNLOCALAI · v38
Llama 3.3 70B Instruct / on / NVIDIA GeForce RTX 4090
Won't fit

Running Llama 3.3 70B Instruct on NVIDIA GeForce RTX 4090

Llama 3.3 70B Instruct requires more memory than NVIDIA GeForce RTX 4090 provides (24 GB available).

By Fredoline Eruo·Last verified May 14, 2026

Model size

70B params
Llama 3.3 70B Instruct →

Memory available

24 GB
NVIDIA GeForce RTX 4090 →

Recommended quant

—
Highest quality that fits

Variants and what fits

QuantizationFile sizeVRAM requiredFits on NVIDIA GeForce RTX 4090?
Q4_K_M40.0 GB48 GB
No
Q5_K_M47.0 GB56 GB
No
Q8_070.0 GB80 GB
No

Real benchmarks

ToolQuantContexttok/sVRAM usedSource
OllamaQ4_K_M8,19214.8 tok/s23.4 GB
community
llama.cppQ4_K_M4,0968.0 tok/s—
community

Frequently asked

Can NVIDIA GeForce RTX 4090 run Llama 3.3 70B Instruct?

Llama 3.3 70B Instruct requires more memory than NVIDIA GeForce RTX 4090 provides (24 GB available).

What quantization should I use?

No quantization of Llama 3.3 70B Instruct fits on NVIDIA GeForce RTX 4090. Pick a smaller model.

How fast will it be?

Measured at 14.8 tok/s on this combination (community-sourced).

See also: Llama 3.3 70B Instruct, NVIDIA GeForce RTX 4090, all benchmarks.

Reviewed by RunLocalAI Editorial. See our editorial policy.

Community benchmarks for this exact pair

Submit your own →

Operator-submitted measurements for this specific model + hardware combination. Editorial review required before publication; provenance badge on every row.

No community benchmarks yet for this combination. Submit yours →