Hardware

79 GPUs, SoCs, and laptops covered. What runs on each, with real benchmarks.

GPUs

NVIDIA GB200 NVL72

13824GB

72-GPU Blackwell rack with 36 Grace CPUs. Hyperscale-only — relevant context here for understanding 'what frontier training runs on'.

nvidia
workstation
CUDA

AMD Instinct MI355X

288GB

Latest CDNA 4. 288GB HBM3e — currently the highest VRAM per chip on the market.

amd
workstation
ROCm

AMD Instinct MI325X

256GB

256GB HBM3e — direct competitor to NVIDIA H200 with more memory.

amd
workstation
ROCm

NVIDIA B200

192GB

Datacenter Blackwell. 192GB HBM3e per chip, ~8 TB/s bandwidth. Cloud-tier — you rent these by the hour.

nvidia
workstation
CUDA

AMD Instinct MI300X

192GB

192GB HBM3 datacenter card. Used by Microsoft, Oracle, Meta cloud deployments.

amd
workstation
ROCm

NVIDIA H100 NVL

188GB

Dual-card H100 with 188GB combined memory. Built for LLM serving.

nvidia
workstation
CUDA

NVIDIA H200

141GB

Hopper refresh — 141GB HBM3e at ~4.8 TB/s. Datacenter-class; rentable on RunPod, Lambda, etc.

nvidia
workstation
CUDA

Intel Gaudi 3

128GB

Intel's enterprise AI accelerator. 128GB HBM2e. Habana stack required — limited ecosystem support.

intel
workstation

AMD Instinct MI250X

128GB

Previous-gen CDNA 2. 128GB HBM2e. Powered the Frontier supercomputer.

amd
workstation
ROCm

NVIDIA RTX PRO 6000 Blackwell

96GB

Pro Blackwell — 96GB GDDR7 ECC. The single-card answer to 70B and 100B+ local inference.

nvidia
workstation
CUDA

Intel Gaudi 2

96GB

Previous-gen Habana accelerator. 96GB HBM2e.

intel
workstation

NVIDIA H100 PCIe

80GB

PCIe Hopper. Lower power, lower bandwidth than SXM. Server-tier.

nvidia
workstation
CUDA

NVIDIA H100 SXM

80GB

Hopper SXM5 — 80GB HBM3 at 3.35 TB/s. The original GPU that trained GPT-4. Cloud-rentable.

nvidia
workstation
CUDA

NVIDIA A100 80GB SXM

80GB

Ampere datacenter flagship. 80GB HBM2e at 2 TB/s. Still common at cloud providers.

nvidia
workstation
CUDA

AMD Instinct MI210

64GB

64GB CDNA 2. Lower-power AMD datacenter option.

amd
workstation
ROCm

NVIDIA L40S

48GB

Ada-gen datacenter card. 48GB GDDR6 — popular at cloud GPU rentals as a budget H100 alternative.

nvidia
workstation
CUDA

NVIDIA L40

48GB

Original Ada datacenter. Slower than L40S. 48GB GDDR6.

nvidia
workstation
CUDA

NVIDIA RTX 6000 Ada Generation

48GB

Pro Ada — 48GB ECC. Pre-Blackwell workstation default.

nvidia
workstation
CUDA

NVIDIA A40

48GB

Ampere workstation/datacenter hybrid. 48GB GDDR6.

nvidia
workstation
CUDA

NVIDIA RTX A6000 (Ampere)

48GB

Ampere-gen workstation card with 48GB. Common in AI labs; used market is reasonable for 48GB at this point.

nvidia
workstation
CUDA

NVIDIA A100 40GB

40GB

Original A100. 40GB HBM2 at 1.55 TB/s. Trained the early generation of frontier models.

nvidia
workstation
CUDA

NVIDIA GeForce RTX 5090

32GB

Blackwell flagship. 32GB GDDR7 on a 512-bit bus delivers ~1.79 TB/s memory bandwidth — the new top of consumer hardware for local LLM infere

nvidia
enthusiast
CUDA

NVIDIA RTX 5000 Ada Generation

32GB

32GB workstation Ada. Mid-tier pro card.

nvidia
workstation
CUDA

NVIDIA GeForce RTX 5090 Mobile

24GB

Mobile Blackwell flagship. 24GB GDDR7 in a laptop is the new high-water mark.

nvidia
enthusiast
CUDA

NVIDIA L4

24GB

Inference-focused Ada datacenter card. Low-power 24GB suitable for 7B-14B serving.

nvidia
workstation
CUDA

AMD Radeon RX 7900 XTX

24GB

AMD's 24GB challenger to the 4090. ROCm Linux now solid for llama.cpp and vLLM. Best price-per-VRAM-GB on the new market.

amd
enthusiast
ROCm

NVIDIA GeForce RTX 3090 Ti

24GB

Highest-tier Ampere consumer card. Used market gold for AI: 24GB at sub-$1200 in 2026.

nvidia
enthusiast
CUDA

NVIDIA GeForce RTX 4090

24GB

The community-default high-end local-AI card from 2022 to 2025. 24GB GDDR6X at ~1 TB/s makes 70B Q4 comfortably loadable.

nvidia
enthusiast
CUDA

NVIDIA RTX A5000

24GB

24GB Ampere workstation card. Tighter power envelope than RTX 3090.

nvidia
workstation
CUDA

NVIDIA GeForce RTX 3090

24GB

The original 24GB CUDA value pick. Used market still strong in 2026 — many AI hobbyists run dual 3090 setups for 70B inference.

nvidia
enthusiast
CUDA

AMD Radeon RX 7900 XT

20GB

20GB RDNA 3. Cheaper alternative to XTX.

amd
enthusiast
ROCm

NVIDIA GeForce RTX 5070 Ti

16GB

16GB Blackwell at the upper-mid price tier. Strong 14B–32B model performance.

nvidia
high
CUDA

NVIDIA GeForce RTX 5080

16GB

Second-tier Blackwell. 16GB GDDR7, ~960 GB/s bandwidth. Fastest 16GB consumer card on the market.

nvidia
enthusiast
CUDA

AMD Radeon RX 9070

16GB

16GB RDNA 4 at sub-$600. ROCm + Vulkan supported.

amd
high
ROCm

NVIDIA GeForce RTX 5060 Ti 16GB

16GB

The 16GB sub-$500 sweet spot. Best value for entering local AI seriously.

nvidia
mid
CUDA

AMD Radeon RX 9070 XT

16GB

RDNA 4 flagship. 16GB at $599 — best AMD value for local AI in 2026.

amd
high
ROCm

NVIDIA GeForce RTX 4080 Super

16GB

Refreshed 4080 with 16GB GDDR6X. Slightly behind 5080 but well-supported.

nvidia
high
CUDA

NVIDIA GeForce RTX 4070 Ti Super

16GB

16GB upgrade of the 4070 Ti. Solid mid-high pick for local AI.

nvidia
high
CUDA

AMD Radeon RX 7600 XT

16GB

Sub-$330 16GB AMD. Memory-bandwidth-limited but great VRAM-per-dollar.

amd
mid
ROCm

AMD Radeon RX 7800 XT

16GB

16GB RDNA 3 mid-range.

amd
high
ROCm

NVIDIA GeForce RTX 4090 Mobile

16GB

Mobile Ada flagship. 16GB VRAM in a laptop. Premium gaming and AI laptop default.

nvidia
enthusiast
CUDA

NVIDIA GeForce RTX 4060 Ti 16GB

16GB

The poster child of 'cheap 16GB CUDA card'. Memory bandwidth is mediocre but 16GB at $400-something opens up 14B Q4.

nvidia
mid
CUDA

NVIDIA GeForce RTX 3080 16GB (Mobile)

16GB

Laptop variant of Ampere. 16GB VRAM in a portable form factor was rare and remains a sleeper pick on the used market.

nvidia
high
CUDA

NVIDIA GeForce RTX 4080

16GB

Original 4080. 16GB GDDR6X. Still capable for 14B–32B Q4 work.

nvidia
high
CUDA

Intel Arc A770 16GB

16GB

Alchemist 16GB. Cheapest path to that VRAM tier. Vulkan llama.cpp is the most-tested route.

intel
mid

NVIDIA GeForce RTX 5070

12GB

Mid-range Blackwell with 12GB. 7B-14B Q4 territory.

nvidia
mid
CUDA

NVIDIA GeForce RTX 4070 Super

12GB

Refreshed 4070. Strong mid-range value for 12GB-tier local AI.

nvidia
mid
CUDA

Intel Arc B580

12GB

Battlemage architecture. 12GB at $250 — the budget compute card. IPEX-LLM and Vulkan are usable paths for AI.

intel
mid

NVIDIA GeForce RTX 4070

12GB

Original 4070. 12GB Ada. Now eclipsed by 4070 Super at the same price.

nvidia
mid
CUDA

AMD Radeon RX 7700 XT

12GB

12GB RDNA 3.

amd
mid
ROCm

NVIDIA GeForce RTX 4070 Ti

12GB

12GB Ada — fits 7B–14B Q4 with usable context.

nvidia
high
CUDA

NVIDIA GeForce RTX 3080 12GB

12GB

Mid-life 12GB refresh of the 3080. Decent 7B–14B card on the used market.

nvidia
high
CUDA

NVIDIA GeForce RTX 3060 12GB

12GB

The community pick for 'cheapest CUDA card with serious VRAM'. The value floor for local AI in 2026.

nvidia
mid
CUDA

Intel Arc B570

10GB

10GB Battlemage at sub-$220. Entry budget compute.

intel
mid

NVIDIA GeForce RTX 3080 10GB

10GB

Original 10GB 3080. Tight on VRAM for AI but still capable for 7B work.

nvidia
high
CUDA

NVIDIA GeForce RTX 5060

8GB

Entry Blackwell. 8GB limits to 7B Q4 with limited context.

nvidia
entry
CUDA

NVIDIA GeForce RTX 5060 Ti 8GB

8GB

8GB Blackwell. Capable of 7B Q4 only — go 16GB SKU instead for AI work.

nvidia
mid
CUDA

NVIDIA GeForce RTX 4060

8GB

Entry-level Ada. 8GB limits to 7B Q4.

nvidia
entry
CUDA

NVIDIA GeForce RTX 4060 Ti 8GB

8GB

8GB version — go 16GB SKU for AI work.

nvidia
mid
CUDA

NVIDIA GeForce RTX 3070

8GB

8GB Ampere. Fits 7B Q4 only.

nvidia
mid
CUDA

APUs

Laptops

Pre-built desktops

Apple Silicon / SoCs