Mistral AIReleased December 11, 2023

Mixtral 8x7B

Open Source56B (13B active) parameters

Context

32K

Input

Free (open)

Key Specifications

🏆

Arena Rank

Not disclosed

📐

Context Window

32K

📥

Input Price

per 1M tokens

Free (open)

📤

Output Price

per 1M tokens

Free (open)

🧠

Parameters

56B (13B active)

🔓

Open Source

Yes

Best For

Efficient inferencemultilingualcoding

About Mixtral 8x7B

Mixtral 8x7B is Mistral AI's pioneering mixture-of-experts model that proved sparse architectures could deliver GPT-3.5 level performance while using only 13 billion active parameters per token. Its release via torrent was a landmark moment for open-source AI, demonstrating that a European startup could produce models competitive with Silicon Valley's best.

Pricing per 1M tokens

Input Tokens

Free (open)

Output Tokens

Free (open)

Frequently Asked Questions

What is Mixtral 8x7B?
Mixtral 8x7B is Mistral AI's pioneering mixture-of-experts model that proved sparse architectures could deliver GPT-3.5 level performance while using only 13 billion active parameters per token. Its release via torrent was a landmark moment for open-source AI, demonstrating that a European startup could produce models competitive with Silicon Valley's best.
How much does Mixtral 8x7B cost?
Mixtral 8x7B costs Free (open) per 1 million input tokens and Free (open) per 1 million output tokens. Pricing is based on token usage, making it cost-effective for both small and large-scale applications.
What is Mixtral 8x7B's context window?
Mixtral 8x7B has a context window of 32K tokens. This determines how much text the model can process in a single request — larger context windows allow the model to handle longer documents, maintain more conversation history, and reason over bigger codebases.
Is Mixtral 8x7B open source?
Yes, Mixtral 8x7B is open source. This means the model weights are publicly available, allowing developers and organizations to download, fine-tune, and self-host the model on their own infrastructure. Open-source models offer greater flexibility and data privacy control.
What is Mixtral 8x7B best for?
Mixtral 8x7B is best suited for: Efficient inference, multilingual, coding. These use cases leverage the model's specific strengths in terms of capability, speed, and cost-effectiveness within Mistral AI's model lineup.