Will It Run AI
CalculatorModelsHardwareCompare
Product
  • Calculator
  • Compare
  • Tier List
Browse
  • Models
  • Hardware
  • Docs
About
  • Why It Works
  • What's New
  • Legal Notice
  • Privacy Policy

All estimates are approximations based on mathematical models and public specifications. Actual performance may vary. Do not make purchasing decisions based solely on these estimates.

Data sourced from Hugging Face, Ollama, and official model documentation. Model names and logos are trademarks of their respective owners.

© 2026 Will It Run AI — Fase Consulting Ibiza, S.L. (NIF: B57969656)

Home/Models/Ministral 8B

MistralMistral

Ministral 8B

Current
huggingfaceHuggingFaceollamaOllama
198.1KDownloads575LikesOct 2024Released131K tokensContextMistral ResearchLicense3 EntryQuality

Get started

— copy & paste to run locally
Ollama
ollama run ministral-8b
HuggingFace
huggingface-cli download ministral-8b

Quick specs

Parameters8B
Architecturedense
Context131K tokens
Modalitytext
Min RAM3.1 GB
Rec. RAM4.9 GB (Q4_K_M)
LicenseMistral Research
FamilyMistral
✓ Chat

About this model

We introduce two new state-of-the-art models for local intelligence, on-device computing, and at-the-edge use cases. We call them les Ministraux: Ministral 3B and Ministral 8B.

  • •Released under the Mistral Research License, reach out to us for a commercial license
  • •Trained with a 128k context window with interleaved sliding-window attention
  • •Trained on a large proportion of multilingual and code data
  • •Supports function calling
  • •Vocabulary size of 131k, using the V3-Tekken tokenizer

Related models

Your hardware

Detecting...

Quick picks

Intel
Best budgetC
Intel Arc B580 12GB~$249 — 45 tok/s
NVIDIA
Best overallB
RTX 3080 10GB~$699 — 118 tok/s

Best hardware

Top picks for Ministral 8B

NVIDIA
RTX 3080 10GBB
10 GB
NVIDIA
RTX 2080 Ti 11GBB
11 GB
NVIDIA
RTX 3080 12GBB
12 GB
NVIDIA
RTX 3080 Ti 12GBB
12 GB
NVIDIA
RTX 5070 12GBB
12 GB

Quantization options

VRAM estimates by quant level

No hardware detected — fit column shows raw VRAM estimates

QuantBitsVRAMQualityFit
Q2_K
2
3.1 GB
Low—
Q3_K_S
3
3.9 GB
Low—
NVFP4
4
4.5 GB
Medium—
Q4_K_M
4
4.9 GB
Medium—
Q5_K_M
5
5.8 GB
High—
Q6_K
6
6.6 GB
High—
Q8_0
8
8.6 GB
Very High—
F16
16
16.4 GB
Maximum—

Hardware compatibility

Fit estimates across all hardware

Open calculator

Computing compatibility...

Memory breakdown

Reference: NVIDIA A10 24GB

Weights4.9 GB
KV Cache1.3 GB
Runtime0.9 GB
Headroom2.4 GB