Will It Run AI
CalculatorModelsHardwareCompare
Product
  • Calculator
  • Compare
  • Tier List
Browse
  • Models
  • Hardware
  • Docs
About
  • Why It Works
  • What's New
  • Legal Notice
  • Privacy Policy

All estimates are approximations based on mathematical models and public specifications. Actual performance may vary. Do not make purchasing decisions based solely on these estimates.

Data sourced from Hugging Face, Ollama, and official model documentation. Model names and logos are trademarks of their respective owners.

© 2026 Will It Run AI — Fase Consulting Ibiza, S.L. (NIF: B57969656)

Browse AI Models

328 models available

/
Status:
Sort:
DeepSeekDeepSeekDeepSeek R1 671B
671B (37B active)131K ctx375.8 GBfrontier
moeLegacy

We introduce our first-generation reasoning models, DeepSeek-R1-Zero and DeepSeek-R1. DeepSeek-R1-Zero, a model trained via large-scale reinforcement learning (RL) without supervised fine-tuning (SFT) as a preliminary step, demonstrated remarkable performance on reasoning. With RL, DeepSeek-R1-Zero naturally emerged with numerous powerful and interesting reasoning behaviors. However, DeepSeek-R1-Zero encounters challenges such as endless repetition, poor readability, and language mixing.

MistralMistralDevstral 2 123B Instruct
123B256K ctx68.9 GBfrontier
denseLegacy

Devstral is an agentic LLM for software engineering tasks. Devstral 2 excels at using tools to explore codebases, editing multiple files and power software engineering agents. The model achieves remarkable performance on SWE-bench.

Z.aiZ.aiGLM-5
744B (40B active)200K ctx416.6 GBfrontier
moeLegacy

📍 Use GLM-5 API services on Z.ai API Platform.

Moonshot AIMoonshot AIKimi K2.5
1000B (32B active)256K ctx560 GBfrontier
moeLegacy

Kimi K2.5 is Moonshot AI's advanced reasoning model with strong performance in math, coding, and multilingual tasks. Features long-context understanding and agentic capabilities for complex multi-step problem solving.

MistralMistralMistral Large 3
675B (41B active)256K ctx378 GBfrontier
moeLegacy

Mistral-Large-Instruct-2411 is an advanced dense Large Language Model (LLM) of 123B parameters with state-of-the-art reasoning, knowledge and coding capabilities extending Mistral-Large-Instruct-2407 with better Long Context, Function Calling and System Prompt.

MistralMistralMistral Small 4 119B
119B (6.5B active)256K ctx66.6 GBfrontier
moeLegacy

Mistral Small 4 is a powerful hybrid model capable of acting as both a general instruction model and a reasoning model. It unifies the capabilities of three different model families—Instruct, Reasoning (previously called Magistral), and Devstral—into a single, unified model.

AlibabaAlibabaQwen3-Coder 30B A3B Instruct
30.5B (3.3B active)256K ctx17.1 GBfrontier
moeLegacy

Qwen3-Coder is available in multiple sizes. Today, we're excited to introduce Qwen3-Coder-30B-A3B-Instruct. This streamlined model maintains impressive performance and efficiency, featuring the following key enhancements:

AlibabaAlibabaQwen3-Coder 480B A35B Instruct
480B (35B active)256K ctx268.8 GBfrontier
moeLegacy

Today, we're announcing Qwen3-Coder, our most agentic code model to date. Qwen3-Coder is available in multiple sizes, but we're excited to introduce its most powerful variant first: Qwen3-Coder-480B-A35B-Instruct. featuring the following key enhancements:

AlibabaAlibabaQwen3-Coder-Next
80B (3B active)256K ctx44.8 GBfrontier
moeLegacy

Today, we're announcing Qwen3-Coder-Next, an open-weight language model designed specifically for coding agents and local development. It features the following key enhancements:

UnslothUnslothQwen3.5 27B
27B0K ctx15.1 GB
denseLegacy

 

UnslothUnslothQwen3.5 35B A3B
35B0K ctx19.6 GB
denseLegacy

 

UnslothUnslothQwen3.5 9B
9B0K ctx5 GB
denseLegacy

 

GoogleGooglegemma 2b
2B0K ctx1.1 GB
denseLegacy

 

HauhauCSHHauhauCSQwen3.5 9B Uncensored HauhauCS Aggressive
9B0K ctx5 GB
denseLegacy

 

BartowskiBBartowskigemma 2 2b it
2B0K ctx1.1 GB
denseLegacy

 

UnslothUnslothQwen3.5 122B A10B
122B0K ctx68.3 GB
denseLegacy

 

BartowskiBBartowskiMeta Llama 3.1 8B Instruct
8B0K ctx4.5 GB
denseLegacy

 

DeepSeekDeepSeekDeepSeek V3 671B
671B (37B active)131K ctx375.8 GBfrontier
moeLegacy

We present DeepSeek-V3, a strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token. To achieve efficient inference and cost-effective training, DeepSeek-V3 adopts Multi-head Latent Attention (MLA) and DeepSeekMoE architectures, which were thoroughly validated in DeepSeek-V2. Furthermore, DeepSeek-V3 pioneers an auxiliary-loss-free strategy for load balancing and sets a multi-token prediction training objective for stronger performance.

MistralMistralMixtral 8x22B
141B (39B active)66K ctx79 GBcurrent
moeLegacy

from mistral_common.tokens.tokenizers.mistral import MistralTokenizer from mistral_common.protocol.instruct.messages import UserMessage from mistral_common.protocol.instruct.request import ChatCompletionRequest

AlibabaAlibabaQwen 2.5 72B
72B131K ctx40.3 GBcurrent
denseLegacy

Qwen2.5 is the latest series of Qwen large language models. For Qwen2.5, we release a number of base language models and instruction-tuned language models ranging from 0.5 to 72 billion parameters. Qwen2.5 brings the following improvements upon Qwen2:

AlibabaAlibabaQwen 3 235B A22B
235B (22B active)131K ctx131.6 GBfrontier
moeLegacy

We introduce the updated version of the Qwen3-235B-A22B non-thinking mode, named Qwen3-235B-A22B-Instruct-2507, featuring the following key enhancements:

AlibabaAlibabaQwen3-VL 30B A3B Instruct
30B (3B active)256K ctx16.8 GBfrontier
moeLegacy

Meet Qwen3-VL — the most powerful vision-language model in the Qwen series to date.

BartowskiBBartowskiLlama 3.2 3B Instruct
3B0K ctx1.7 GB
denseLegacy

 

UnslothUnslothQwen3.5 4B
4B0K ctx2.2 GB
denseLegacy

 

PreviousPage 1 of 14Next