Phi-4 Mini 4B
Microsoft's edge-tier Phi-4 variant. 3.8B params; designed for phone / tablet / Pi deployment. Strong reasoning per parameter — Phi family's traditional advantage carries to the smallest tier.
Overview
Microsoft's edge-tier Phi-4 variant. 3.8B params; designed for phone / tablet / Pi deployment. Strong reasoning per parameter — Phi family's traditional advantage carries to the smallest tier.
Family & lineage
How this model relates to others in its lineage. Family members share architecture and training-data roots; parent / children edges record direct distillation or fine-tune relationships.
Strengths
- Edge-tier deployment (phone, Pi)
- MIT license
- Strong reasoning per parameter
Weaknesses
- Small parameter count limits open-ended generation
Quantization variants
Each quantization trades model quality for file size and VRAM. Q4_K_M is the most popular starting point.
| Quantization | File size | VRAM required |
|---|---|---|
| Q4_K_M | 2.4 GB | 4 GB |
Get the model
HuggingFace
Original weights
Source repository — direct quantization required.
Hardware that runs this
Cards with enough VRAM for at least one quantization of Phi-4 Mini 4B.
Models worth comparing
Same parameter band, plus what's one tier above and below — so you can decide what actually fits your hardware.
Frequently asked
What's the minimum VRAM to run Phi-4 Mini 4B?
Can I use Phi-4 Mini 4B commercially?
What's the context length of Phi-4 Mini 4B?
Source: huggingface.co/microsoft/Phi-4-mini
Reviewed by RunLocalAI Editorial. See our editorial policy for how we research and verify model claims.