Mixtral 8x7B
Mixtral 8x7B is Mistral AI's entry in a crowded field. Context window: 0.032K tokens.
Context
32K
Input
Free (open)
Key Specifications
Arena Rank
Not disclosed
Context Window
32K
Input Price
per 1M tokens
Free (open)
Output Price
per 1M tokens
Free (open)
Parameters
56B (13B active)
Open Source
Best For
About Mixtral 8x7B
Mixtral 8x7B, developed by Mistral AI, is an open-source Mixture-of-Experts model with 56 billion total parameters (13 billion active per token) and a 32K token context window. The model pioneered the practical application of MoE architecture in open-source AI, demonstrating that sparse expert routing could deliver performance comparable to much larger dense models at a fraction of the inference cost. Mixtral 8x7B handles coding, reasoning, and multilingual tasks efficiently, activating only the most relevant experts for each input. Free and fully open-source, it runs on consumer-grade multi-GPU setups and has become a benchmark for efficient model design. Its success influenced subsequent MoE models from DeepSeek, Alibaba, and others. The model remains widely deployed in production for cost-sensitive applications requiring better-than-7B performance.
Pricing per 1M tokens
Input Tokens
Free (open)
Output Tokens
Free (open)
Compare Mixtral 8x7B
See how Mixtral 8x7B stacks up against other leading AI models