Step 1 — Run the bench On the machine with the model loaded. ~2 minutes for 5 runs.
# One-liner, no install:
curl -fsSL https://www.runlocalai.co/bench.mjs | node --model llama3.1:8b
# Or download + customize:
curl -fsSL https://www.runlocalai.co/bench.mjs -o bench.mjs
node bench.mjs --model qwen3:14b --prompt "Explain attention"Want even less friction? Run with --submit and the bench CLI ships directly via API. Use this page when you want to review the values before they hit the queue.
Step 2 — Paste, pick hardware, submit Paste the full terminal output below. Auto-detection lights up the fields. Pick the matching hardware. Hit submit.
§ Paste terminal output ↺ Load sample
Hardware (required) * — pick the hardware you ran on — AMD Instinct MI210 (64 GB) AMD Instinct MI250X (128 GB) AMD Instinct MI300A (APU) (128 GB) AMD Instinct MI300X (192 GB) AMD Instinct MI325X (256 GB) AMD Instinct MI355X (288 GB) AMD Radeon RX 5500 XT 8GB (8 GB) AMD Radeon RX 5600 XT (6 GB) AMD Radeon RX 570 (4 GB) AMD Radeon RX 5700 XT (8 GB) AMD Radeon RX 580 8GB (8 GB) AMD Radeon RX 6600 (8 GB) AMD Radeon RX 6600 XT (8 GB) AMD Radeon RX 6650 XT (8 GB) AMD Radeon RX 6700 XT (12 GB) AMD Radeon RX 6750 XT (12 GB) AMD Radeon RX 6800 (16 GB) AMD Radeon RX 6800 XT (16 GB) AMD Radeon RX 6900 XT (16 GB) AMD Radeon RX 6950 XT (16 GB) AMD Radeon RX 7600 XT (16 GB) AMD Radeon RX 7700 XT (12 GB) AMD Radeon RX 7800 XT (16 GB) AMD Radeon RX 7900 GRE (16 GB) AMD Radeon RX 7900 XT (20 GB) AMD Radeon RX 7900 XTX (24 GB) AMD Radeon RX 9070 (16 GB) AMD Radeon RX 9070 XT (16 GB) ASUS ROG Strix Scar 18 (RTX 5090 Mobile) (24 GB) Framework Laptop 16 (RX 7700S) (8 GB) Intel Arc A770 16GB (16 GB) Intel Arc B570 (10 GB) Intel Arc B580 (12 GB) Intel Gaudi 2 (96 GB) Intel Gaudi 3 (128 GB) Lenovo Legion 5 Pro Gen 7 (RTX 3080 16GB) (16 GB) NVIDIA A100 40GB (40 GB) NVIDIA A100 80GB SXM (80 GB) NVIDIA A40 (48 GB) NVIDIA B200 (192 GB) NVIDIA GB200 NVL72 (13824 GB) NVIDIA GeForce GTX 1050 Ti (4 GB) NVIDIA GeForce GTX 1060 3GB (3 GB) NVIDIA GeForce GTX 1060 6GB (6 GB) NVIDIA GeForce GTX 1070 (8 GB) NVIDIA GeForce GTX 1070 Ti (8 GB) NVIDIA GeForce GTX 1080 (8 GB) NVIDIA GeForce GTX 1080 Ti (11 GB) NVIDIA GeForce GTX 1650 (4 GB) NVIDIA GeForce GTX 1650 Super (4 GB) NVIDIA GeForce GTX 1660 (6 GB) NVIDIA GeForce GTX 1660 Super (6 GB) NVIDIA GeForce GTX 1660 Ti (6 GB) NVIDIA GeForce RTX 2060 (6 GB) NVIDIA GeForce RTX 2060 Super (8 GB) NVIDIA GeForce RTX 2070 (8 GB) NVIDIA GeForce RTX 2070 Super (8 GB) NVIDIA GeForce RTX 2080 Super (8 GB) NVIDIA GeForce RTX 2080 Ti (11 GB) NVIDIA GeForce RTX 3050 (8 GB) NVIDIA GeForce RTX 3050 Ti (Mobile) (4 GB) NVIDIA GeForce RTX 3060 12GB (12 GB) NVIDIA GeForce RTX 3060 Ti (8 GB) NVIDIA GeForce RTX 3070 (8 GB) NVIDIA GeForce RTX 3070 Ti (8 GB) NVIDIA GeForce RTX 3080 10GB (10 GB) NVIDIA GeForce RTX 3080 12GB (12 GB) NVIDIA GeForce RTX 3080 16GB (Mobile) (16 GB) NVIDIA GeForce RTX 3080 Ti (12 GB) NVIDIA GeForce RTX 3090 (24 GB) NVIDIA GeForce RTX 3090 Ti (24 GB) NVIDIA GeForce RTX 4060 (8 GB) NVIDIA GeForce RTX 4060 Ti 16GB (16 GB) NVIDIA GeForce RTX 4060 Ti 8GB (8 GB) NVIDIA GeForce RTX 4070 (12 GB) NVIDIA GeForce RTX 4070 Super (12 GB) NVIDIA GeForce RTX 4070 Ti (12 GB) NVIDIA GeForce RTX 4070 Ti Super (16 GB) NVIDIA GeForce RTX 4080 (16 GB) NVIDIA GeForce RTX 4080 Super (16 GB) NVIDIA GeForce RTX 4090 (24 GB) NVIDIA GeForce RTX 4090 Mobile (16 GB) NVIDIA GeForce RTX 5060 (8 GB) NVIDIA GeForce RTX 5060 Ti 16GB (16 GB) NVIDIA GeForce RTX 5060 Ti 8GB (8 GB) NVIDIA GeForce RTX 5070 (12 GB) NVIDIA GeForce RTX 5070 Ti (16 GB) NVIDIA GeForce RTX 5080 (16 GB) NVIDIA GeForce RTX 5090 (32 GB) NVIDIA GeForce RTX 5090 Mobile (24 GB) NVIDIA H100 NVL (188 GB) NVIDIA H100 PCIe (80 GB) NVIDIA H100 SXM (80 GB) NVIDIA H200 (141 GB) NVIDIA L4 (24 GB) NVIDIA L40 (48 GB) NVIDIA L40S (48 GB) NVIDIA RTX 5000 Ada Generation (32 GB) NVIDIA RTX 6000 Ada Generation (48 GB) NVIDIA RTX A5000 (24 GB) NVIDIA RTX A6000 (Ampere) (48 GB) NVIDIA RTX PRO 6000 Blackwell (96 GB) Razer Blade 16 (2025, RTX 5090 Mobile) (24 GB)
Model — pick the model — Llama 3.2 11B Vision Instruct (11.0B) · llama3.2-vision:11b OLMo 2 13B (13.0B) Llama 4 Scout (109.0B) · llama4:scout Tulu 3 8B (8.0B) DeepSeek V4 Pro (1.6T MoE) (1600.0B) Llama 3.2 90B Vision Instruct (90.0B) · llama3.2-vision:90b Mixtral 8x22B Instruct (141.0B) · mixtral:8x22b Qwen 3 4B (4.0B) · qwen3:4b Hermes 3 Llama 3.1 70B (70.0B) · hermes3:70b Hermes 4 Llama 3.3 70B (70.0B) Qwen 3 235B-A22B (235.0B) · qwen3:235b CodeGemma 7B (7.0B) · codegemma:7b Mistral Large 2 (123B) (123.0B) · mistral-large:123b DeepSeek Coder V3 (33.0B) Llama 3.2 1B Instruct (1.0B) · llama3.2:1b Pixtral 12B (12.0B) · pixtral:12b Llama 3.3 70B Instruct (70.0B) · llama3.3:70b Molmo 7B-D (8.0B) Phi-3.5 Vision (4.2B) Qwen 2.5 32B Instruct (32.0B) · qwen2.5:32b Gemma 3 1B (1.0B) · gemma3:1b Gemma 4 E4B (Effective 4B) (4.0B) · gemma4:e4b Gemma 4 E2B (Effective 2B) (2.0B) · gemma4:e2b Jamba 1.5 Large (398.0B) Gemma 4 26B MoE (26.0B) · gemma4:26b-moe Hunyuan Large 389B MoE (389.0B) DeepSeek R1 Distill Qwen 14B (14.0B) · deepseek-r1:14b Llama 3.1 Nemotron Ultra 253B (253.0B) Llama 3.1 Nemotron Nano 8B (8.0B) Falcon Mamba 7B (7.0B) Granite 3.0 2B Instruct (2.0B) Qwen 3 30B-A3B (30.0B) · qwen3:30b DeepSeek R1 Distill Qwen 7B (7.0B) · deepseek-r1:7b Kimi K2.6 (1000.0B) Mistral 7B Instruct v0.3 (7.0B) · mistral:7b Qwen 2.5 Math 72B (72.0B) Granite 3.0 8B Instruct (8.0B) Qwen 3.5 235B-A17B (MoE) (397.0B) Qwen 2.5 Coder 7B Instruct (7.0B) · qwen2.5-coder:7b DeepSeek V4 Flash (284B MoE) (284.0B) DBRX Base (132.0B) Qwen 2.5 7B Instruct (7.0B) · qwen2.5:7b Qwen 2.5 14B Instruct (14.0B) · qwen2.5:14b Codestral 22B (22.0B) · codestral:22b Phi-3.5 Mini Instruct (3.8B) · phi3.5:3.8b Mistral Small 3 24B (24.0B) · mistral-small:24b InternVL 2.5 78B (78.0B) DeepSeek MoE 16B Base (16.0B) Gemma 3 12B (12.0B) · gemma3:12b DeepSeek R1 Distill Llama 70B (70.0B) · deepseek-r1:70b DeepSeek R1 Distill Mistral 24B (24.0B) Dolphin 3.0 Mistral 24B (24.0B) · dolphin-mistral:24b Llama 3.2 11B Vision (11.0B) DeepSeek R1 Distill Qwen 3 32B (32.0B) Qwen 2.5-VL 7B (7.0B) DeepSeek V4 (745.0B) DeepSeek V3 (671B MoE) (671.0B) · deepseek-v3:671b DeepSeek V3 Lite (16B MoE) (16.0B) DeepSeek R1 Distill Llama 8B (8.0B) Dolphin 3 Llama 3.3 70B (70.0B) DeepSeek R1 Distill Qwen 1.5B (1.5B) DeepSeek Coder V2 Lite (16B) (16.0B) · deepseek-coder-v2:16b EVA Llama 3.3 70B (70.0B) Gemma 2 9B Instruct (9.0B) · gemma2:9b Mistral Medium 3 24B (dense) (24.0B) Stable LM 2 12B (12.0B) Llama 4 70B (70.0B) Codestral Mamba 7B (7.0B) Llama 3.2 3B Instruct (3.0B) · llama3.2:3b Yi Coder 9B (9.0B) Phi-4 Mini 4B (3.8B) Phi-4 Multimodal (14.0B) Mistral Small 3.2 24B (24.0B) Jamba 1.5 Mini (52.0B) Gemma 3 27B (27.0B) · gemma3:27b Mistral Nemo 12B Instruct (12.0B) · mistral-nemo:12b Gemma 4 31B Dense (31.0B) · gemma4:31b Gemma 3 4B (4.0B) · gemma3:4b Qwen 2.5 Coder 3B (3.0B) Qwen 3 7B (7.0B) GLM-5 (200.0B) Qwen 2.5 Math 7B (7.0B) GLM-5 Pro (144.0B) Nemotron 3 Nano 9B (9.0B) Qwen 2.5 Coder 1.5B (1.5B) MiniCPM-V 2.6 8B (8.0B) OpenCoder 8B (8.0B) Granite 3.2 8B (8.0B) Step-3 (1000.0B) DBRX Instruct (132.0B) EXAONE 3.5 8B (7.8B) InternLM 3 8B (8.0B) GLM-4 9B (9.0B) MiniCPM 3 4B (4.0B) Kimi K1.5 (200.0B) Granite 3.3 8B (8.0B) QwQ 32B Preview (32.0B) · qwq:32b Falcon 3 10B (10.0B) Dolphin 3.0 Llama 3.2 3B (3.0B) Hermes 3 Llama 3.2 3B (3.0B) Janus-Pro 7B (7.0B) RWKV 7 'Goose' 1.5B (1.5B) Ministral 8B Instruct (8.0B) Granite 3 MoE (3B active) (16.0B) Mixtral 8x7B Instruct (47.0B) · mixtral:8x7b Llama 3.1 Nemotron 70B Instruct (70.0B) · nemotron:70b MiniCPM-V 3 8B (8.0B) Qwen 2-VL 7B (7.0B) Ministral 3B Instruct (3.0B) Llama 3.2 90B Vision (90.0B) Qwen 3 72B (72.0B) Hermes 3 Llama 3.1 8B (8.0B) · hermes3:8b EXAONE 3.5 32B (32.0B) Qwen 2.5 Coder 14B Instruct (14.0B) Aya 23 8B (8.0B) Devstral Small 2 24B (24.0B) OLMo 2 32B (32.0B) Phi-4 Reasoning Mini 4B (3.8B) Baichuan 4 13B (13.0B) Command R+ (Aug 2024) (104.0B) Moondream 2 (1.9B) SmolLM 3 3B (3.0B) Aya Expanse 32B (32.0B) Phind CodeLlama 34B v2 (34.0B) StarCoder 2 7B (7.0B) Aya 23 35B (35.0B) CodeQwen 1.5 7B (7.0B) Qwen 3 8B (8.0B) · qwen3:8b Llama 3.3 8B Instruct (8.0B) Whisper Large v3 Turbo (0.8B) Whisper Large v3 (1.6B) LLaVA 1.6 Mistral 7B (7.0B) StarCoder 2 3B (3.0B) StarCoder 2 15B (15.0B) Llama 4 405B (405.0B) Qwen 2.5 Coder 32B Instruct (32.0B) · qwen2.5-coder:32b Phi-4 14B (14.0B) · phi4:14b Phi-4 Reasoning 14B (14.0B) · phi4-reasoning:14b LLaVA-OneVision 7B (7.0B) Qwen 3 32B (32.0B) · qwen3:32b InternVL 2.5 26B (26.0B) Qwen 3 14B (14.0B) · qwen3:14b Command R+ 104B (104.0B) · command-r-plus:104b Nemotron 3 Nano (30B-A3B) (30.0B) · nemotron3:nano SmolLM 2 1.7B Instruct (1.7B) Nemotron Mini 4B Instruct (4.0B) GLM-4V 9B (13.9B) PaliGemma 2 3B (3.0B) PaliGemma 2 10B (10.0B) Falcon 3 7B Instruct (7.0B) InternLM 2.5 7B Chat (7.0B) EXAONE 3.5 2.4B (2.4B) NV-Embed v2 (7.8B) BGE M3 (0.6B) SmolLM 2 360M Instruct (0.4B) BGE Reranker v2 M3 (0.6B) DeepSeek R1 (671B reasoning) (671.0B) · deepseek-r1:671b Qwen 2.5 1.5B Instruct (1.5B) Qwen 2.5 3B Instruct (3.0B) Qwen 2.5-VL 3B (3.0B) DeepSeek Coder V2 236B (236.0B) Qwen 3 Embedding 8B (8.0B) DeepSeek V2.5 236B (236.0B) Qwen 2.5 0.5B Instruct (0.5B) Llama 3.1 70B Instruct (70.0B) · llama3.1:70b Magistral 32B (32.0B) MedGemma 27B (27.0B) Llama 3.1 8B Instruct (8.0B) · llama3.1:8b Qwen 2.5 72B Instruct (72.0B) · qwen2.5:72b DeepSeek R1 Distill Qwen 32B (32.0B) · deepseek-r1:32b Mistral Medium 3.5 (675B MoE) (675.0B) Command R 35B (35.0B) · command-r:35b Mistral Saba 24B (24.0B) Molmo 72B (72.0B) Nemotron 3 Super 49B (49.0B) Llama 4 Maverick (400.0B) · llama4:maverick Nemotron 3 Super (120B-A12B) (120.0B) · nemotron3:super OpenBioLLM Llama 3 70B (70.0B) Qwen 2.5-VL 72B (72.0B) Qwen 3 Coder 32B (32.0B) Tulu 3 70B (70.0B) WizardLM-2 8x22B (141.0B) Yi 1.5 34B (34.0B) · yi:34b
Extra notes (optional)
Paste output, pick hardware + model to enable submit.
Submit benchmark
What gets stored Your submission lands in the editorial review queue. Nothing auto-publishes. When approved, it appears on:
The model's detail page (/models/<slug>) as a measured benchmark The hardware's detail page (/hardware/<slug>) and the leaderboard's confidence chip The community benchmarks feed The cost calculator and quant advisor — your measurement replaces our bandwidth-derived estimate for the exact model × hardware × quant combo you ran Your IP is hashed daily (never stored raw). Email is optional and only used to contact you with reproduction questions. See privacy for the full handling policy.