Skip to main content
Mistral AIReleased July 18, 2024

Mistral Nemo

Open Source#27 Arena Rank12B parameters

Mistral Nemo holds a solid spot in the Arena rankings at #27. Context window: 0.128K tokens.

Context

128K

Input

$0.30

Key Specifications

🏆

Arena Rank

#27

📐

Context Window

128K

📥

Input Price

per 1M tokens

$0.30

📤

Output Price

per 1M tokens

$0.30

🧠

Parameters

12B

🔓

Open Source

Yes

Best For

Lightweight tasksdrop-in replacement

About Mistral Nemo

Mistral Nemo, developed jointly by Mistral AI and NVIDIA, is a compact open-source model with 12 billion parameters designed as a high-performance replacement for smaller models. Despite its size, the model delivers performance significantly above its weight class on coding, reasoning, and multilingual tasks, benefiting from the combined expertise of Mistral's model architecture team and NVIDIA's training infrastructure. Mistral Nemo can run on a single consumer GPU, making it ideal for organizations with limited compute resources or strict data privacy requirements that preclude cloud-based API usage. Its small footprint enables fast inference and low-cost deployment while maintaining the quality standards of the Mistral model family. Free and open-source, the model supports commercial use and fine-tuning. It has become a popular choice for developers seeking capable, self-hosted AI without the hardware demands of larger models.

Pricing per 1M tokens

Input Tokens

$0.30

Output Tokens

$0.30

Frequently Asked Questions

What is Mistral Nemo?
Mistral Nemo, developed jointly by Mistral AI and NVIDIA, is a compact open-source model with 12 billion parameters designed as a high-performance replacement for smaller models. Despite its size, the model delivers performance significantly above its weight class on coding, reasoning, and multilingual tasks, benefiting from the combined expertise of Mistral's model architecture team and NVIDIA's training infrastructure. Mistral Nemo can run on a single consumer GPU, making it ideal for organizations with limited compute resources or strict data privacy requirements that preclude cloud-based API usage. Its small footprint enables fast inference and low-cost deployment while maintaining the quality standards of the Mistral model family. Free and open-source, the model supports commercial use and fine-tuning. It has become a popular choice for developers seeking capable, self-hosted AI without the hardware demands of larger models.
How much does Mistral Nemo cost?
Mistral Nemo costs $0.30 per 1M input tokens and $0.30 per 1M output tokens. You pay only for what you use, which keeps costs predictable.
What is Mistral Nemo's context window?
Mistral Nemo has a context window of 128K tokens. This determines how much text the model can process in a single request — bigger windows mean longer documents and richer conversation history.
Is Mistral Nemo open source?
Yes, Mistral Nemo is open source. The model weights are publicly available, so developers can download, fine-tune, and self-host it. Open-source models give teams more control over data privacy and deployment.
What is Mistral Nemo best for?
Mistral Nemo is best suited for: Lightweight tasks, drop-in replacement. These use cases play to the model's strengths in capability, speed, and cost within Mistral AI's lineup.