Will It Run AI
CalculatorModelsHardwareCompare
Product
  • Calculator
  • Compare
  • Tier List
Browse
  • Models
  • Hardware
  • Docs
About
  • Why It Works
  • What's New
  • Legal Notice
  • Privacy Policy

All estimates are approximations based on mathematical models and public specifications. Actual performance may vary. Do not make purchasing decisions based solely on these estimates.

Data sourced from Hugging Face, Ollama, and official model documentation. Model names and logos are trademarks of their respective owners.

© 2026 Will It Run AI — Fase Consulting Ibiza, S.L. (NIF: B57969656)

Home/Models/Phi 3 Medium 14B

MicrosoftMicrosoft

Phi 3 Medium 14B

Current
huggingfaceHuggingFaceollamaOllama
98.1KDownloads387LikesMay 2024Released128K tokensContextMITLicense4 EntryQuality

Get started

— copy & paste to run locally
Ollama
ollama run phi-3-medium-14b
HuggingFace
huggingface-cli download phi-3-medium-14b

Quick specs

Parameters14B
Architecturedense
Context128K tokens
Modalitytext
Min RAM5.5 GB
Rec. RAM8.5 GB (Q4_K_M)
LicenseMIT
FamilyPhi
✓ Chat✓ Reasoning

About this model

The Phi-3-Medium-128K-Instruct is a 14B parameters, lightweight, state-of-the-art open model trained with the Phi-3 datasets that includes both synthetic data and the filtered publicly available websites data with a focus on high-quality and reasoning dense properties. The model belongs to the Phi-3 family with the Medium version in two variants 4k and 128K which is the context length (in tokens) that it can support.

  • •Memory/compute constrained environments
  • •Latency bound scenarios
  • •Strong reasoning (especially code, math and logic)

Related models

Your hardware

Detecting...

Quick picks

AMD
Best budgetC
RX 7600 XT 16GB~$329 — 20 tok/s
AMD
Best overallC
RX 7900 XT 20GB~$899 — 56 tok/s

Best hardware

Top picks for Phi 3 Medium 14B

NVIDIA
RTX A4500 20GBC
20 GB
AMD
RX 7900 XT 20GBC
20 GB
NVIDIA
RTX 4090 24GBC
24 GB
NVIDIA
RTX 5090 Laptop 24GBC
24 GB
NVIDIA
NVIDIA A30 24GBC
24 GB

Quantization options

VRAM estimates by quant level

No hardware detected — fit column shows raw VRAM estimates

QuantBitsVRAMQualityFit
Q2_K
2
5.5 GB
Low—
Q3_K_S
3
6.9 GB
Low—
NVFP4
4
7.8 GB
Medium—
Q4_K_M
4
8.5 GB
Medium—
Q5_K_M
5
10.1 GB
High—
Q6_K
6
11.5 GB
High—
Q8_0
8
15.0 GB
Very High—
F16
16
28.7 GB
Maximum—

Hardware compatibility

Fit estimates across all hardware

Open calculator

Computing compatibility...

Memory breakdown

Reference: NVIDIA A10 24GB

Weights8.5 GB
KV Cache2.2 GB
Runtime0.9 GB
Headroom2.4 GB