Llama 3.2 90B Vision Instruct
The 90B vision Llama. Best-in-class first-party multimodal open weight at the time of release. Workstation-class only.
Overview
The 90B vision Llama. Best-in-class first-party multimodal open weight at the time of release. Workstation-class only.
Strengths
- Top-tier open-weight vision quality
- 128K context
Weaknesses
- Needs 60GB+ VRAM
- EU restricted
Quantization variants
Each quantization trades model quality for file size and VRAM. Q4_K_M is the most popular starting point.
| Quantization | File size | VRAM required |
|---|---|---|
| Q4_K_M | 51.0 GB | 60 GB |
Get the model
Ollama
One-line install
ollama run llama3.2-vision:90bRead our Ollama review →HuggingFace
Original weights
Source repository — direct quantization required.
Hardware that runs this
Cards with enough VRAM for at least one quantization of Llama 3.2 90B Vision Instruct.
Models worth comparing
Same parameter band, plus what's one tier above and below — so you can decide what actually fits your hardware.
Frequently asked
What's the minimum VRAM to run Llama 3.2 90B Vision Instruct?
Can I use Llama 3.2 90B Vision Instruct commercially?
What's the context length of Llama 3.2 90B Vision Instruct?
How do I install Llama 3.2 90B Vision Instruct with Ollama?
Does Llama 3.2 90B Vision Instruct support images?
Source: huggingface.co/meta-llama/Llama-3.2-90B-Vision-Instruct
Reviewed by RunLocalAI Editorial. See our editorial policy for how we research and verify model claims.