Will It Run AI
CalculatorModelsHardwareCompare
Product
  • Calculator
  • Compare
  • Tier List
Browse
  • Models
  • Hardware
  • Docs
About
  • Why It Works
  • What's New
  • Legal Notice
  • Privacy Policy

All estimates are approximations based on mathematical models and public specifications. Actual performance may vary. Do not make purchasing decisions based solely on these estimates.

Data sourced from Hugging Face, Ollama, and official model documentation. Model names and logos are trademarks of their respective owners.

© 2026 Will It Run AI — Fase Consulting Ibiza, S.L. (NIF: B57969656)

Browse AI Models

328 models available

/
Status:
Sort:
MradermacherMMradermacherCodestral RAG 19B Pruned i1
19B0K ctx10.6 GB
denseLegacy

 

MradermacherMMradermacherblossom v3 baichuan2 7b i1
7B0K ctx3.9 GB
denseLegacy

 

MradermacherMMradermacherHelply 10.2b chat i1
10.2B0K ctx5.7 GB
denseLegacy

 

MradermacherMMradermacherAI21 Jamba2 3B i1
3B0K ctx1.7 GB
denseLegacy

 

MradermacherMMradermacherblossom v1 baichuan 7b i1
7B0K ctx3.9 GB
denseLegacy

 

MistralMistralMinistral 8B
8B131K ctx4.5 GBcurrent
denseLegacy

We introduce two new state-of-the-art models for local intelligence, on-device computing, and at-the-edge use cases. We call them les Ministraux: Ministral 3B and Ministral 8B.

BigCodeBigCodeStarCoder2 7B
7B16K ctx3.9 GBcurrent
denseLegacy

- Project Website: bigcode-project.org - Paper: Link - Point of Contact: contact@bigcode-project.org - Languages: 17 Programming languages

MradermacherMMradermacherBaichuanMed OCR 72B i1
72B0K ctx40.3 GB
denseLegacy

 

GoogleGoogleGemma 2 9B
9B8K ctx5 GBcurrent
denseLegacy

Gemma 2 9B is Google's mid-size open model built on Gemini research. Features improved reasoning and safety with a novel architecture optimized for efficient inference on consumer hardware.

IBMIBMGranite 3.1 8B
8B128K ctx4.5 GBcurrent
denseLegacy

Model Summary: Granite-3.1-8B-Instruct is a 8B parameter long-context instruct model finetuned from Granite-3.1-8B-Base using a combination of open source instruction datasets with permissive license and internally collected synthetic datasets tailored for solving long context problems. This model is developed using a diverse set of techniques with a structured chat format, including supervised finetuning, model alignment using reinforcement learning, and model merging.

LLaVALLaVALLaVA 1.5 7B
7B4K ctx3.9 GBlegacy
denseLegacy

Model type: LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture.

Cognitive ComputationsCognitive ComputationsSamantha 7B
7B4K ctx3.9 GBlegacy
denseLegacy

Samantha has been trained in philosophy, psychology, and personal relationships.

WizardLMWizardLMWizardMath 7B
7B4K ctx3.9 GBlegacy
denseLegacy

📃 [WizardLM] • 📃 [WizardCoder] • 📃 [WizardMath]

01.AI01.AIYi 1.5 9B
9B4K ctx5 GBcurrent
denseLegacy

🐙 GitHub • 👾 Discord • 🐤 Twitter • 💬 WeChat

CerebrasCerebrasCerebras-GPT 13B
13B131K ctx7.3 GBlegacy
denseLegacy

Check out our Blog Post and arXiv paper!

MistralMistralMinistral 3 3B
3B262K ctx1.7 GBfrontier
multimodalLegacy

The smallest model in the Ministral 3 family, Ministral 3 3B is a powerful, efficient tiny language model with vision capabilities.

MicrosoftMicrosoftPhi 4 Mini 4B
4B128K ctx2.2 GBfrontier
denseLegacy

Phi-4-mini-instruct is a lightweight open model built upon synthetic data and filtered publicly available websites - with a focus on high-quality, reasoning dense data. The model belongs to the Phi-4 model family and supports 128K token context length. The model underwent an enhancement process, incorporating both supervised fine-tuning and direct preference optimization to support precise instruction adherence and robust safety measures.

AlibabaAlibabaQwen 2.5 7B
7B131K ctx3.9 GBcurrent
denseLegacy

Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:

MetaMetaCodeLlama 7B Instruct
7B16K ctx3.9 GBlegacy
denseLegacy

Code Llama is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 34 billion parameters. This is the repository for the 7B instruct-tuned version in the Hugging Face Transformers format. This model is designed for general code synthesis and understanding. Links to other models can be found in the index at the bottom.

MosaicMLMosaicMLMPT-7B-Instruct
7B8K ctx3.9 GBlegacy
denseLegacy

MPT-7B Instruct is MosaicML's instruction-tuned model with a commercially permissive license. Supports 65K context with ALiBi positional encoding for efficient long-document processing.

Stability AIStability AIStableLM 2 12B
12B4K ctx6.7 GBlegacy
denseLegacy

`Stable LM 2 12B Chat` is a 12 billion parameter instruction tuned language model trained on a mix of publicly available datasets and synthetic datasets, utilizing Direct Preference Optimization (DPO).

DeepSeekDeepSeekDeepSeek LLM 7B
7B4K ctx3.9 GBlegacy
denseLegacy

Introducing DeepSeek LLM, an advanced language model comprising 7 billion parameters. It has been trained from scratch on a vast dataset of 2 trillion tokens in both English and Chinese. In order to foster research, we have made DeepSeek LLM 7B/67B Base and DeepSeek LLM 7B/67B Chat open source for the research community.

DeepSeekDeepSeekDeepSeek R1 1.5B
1.5B33K ctx0.8 GBactive
denseLegacy

DeepSeek R1 Distill Qwen 1.5B is a compact reasoning model distilled from DeepSeek-R1, based on Qwen2.5-Math-1.5B. Fine-tuned on 800K curated samples, it achieves 83.9% on MATH-500 and supports chain-of-thought reasoning on resource-constrained devices.

MetaMetaLlama 3.1 8B
8B128K ctx4.5 GBlegacy
denseLegacy

Llama 3.1 8B is Meta's efficient general-purpose model supporting 128K context and multilingual text generation. Optimized for dialogue, summarization, reasoning, and code generation tasks.

PreviousPage 13 of 14Next