Skip to main content
MicrosoftReleased April 15, 2024

WizardLM-2 8x22B

Open Source176B (39B active) parameters

WizardLM-2 8x22B is Microsoft's entry in a crowded field. Context window: 0.064K tokens.

Context

64K

Input

Free (open)

Key Specifications

🏆

Arena Rank

Not disclosed

📐

Context Window

64K

📥

Input Price

per 1M tokens

Free (open)

📤

Output Price

per 1M tokens

Free (open)

🧠

Parameters

176B (39B active)

🔓

Open Source

Yes

Best For

Complex instructionsreasoningcoding

About WizardLM-2 8x22B

WizardLM-2 8x22B, developed by Microsoft, is an instruction-tuned Mixture-of-Experts model with 176 billion total parameters (39 billion active per token) and a 64K token context window. Built upon the Mixtral 8x22B architecture, it applies Microsoft's WizardLM training methodology to enhance complex instruction following, reasoning, and coding capabilities. The model demonstrates substantial improvements over its base on multi-step reasoning, structured output generation, and nuanced writing tasks. WizardLM-2 uses Evol-Instruct, a method that progressively evolves training instructions to increase complexity and diversity. Free and open-source, it can be deployed on enterprise multi-GPU setups. The model represents Microsoft's contribution to the open-source community through instruction-tuning research that advances the capability of existing base models without requiring new pre-training runs.

Pricing per 1M tokens

Input Tokens

Free (open)

Output Tokens

Free (open)

Frequently Asked Questions

What is WizardLM-2 8x22B?
WizardLM-2 8x22B, developed by Microsoft, is an instruction-tuned Mixture-of-Experts model with 176 billion total parameters (39 billion active per token) and a 64K token context window. Built upon the Mixtral 8x22B architecture, it applies Microsoft's WizardLM training methodology to enhance complex instruction following, reasoning, and coding capabilities. The model demonstrates substantial improvements over its base on multi-step reasoning, structured output generation, and nuanced writing tasks. WizardLM-2 uses Evol-Instruct, a method that progressively evolves training instructions to increase complexity and diversity. Free and open-source, it can be deployed on enterprise multi-GPU setups. The model represents Microsoft's contribution to the open-source community through instruction-tuning research that advances the capability of existing base models without requiring new pre-training runs.
How much does WizardLM-2 8x22B cost?
Input pricing for WizardLM-2 8x22B is Free (open) per million tokens; output runs Free (open). Token-based pricing means you can scale up or down without a fixed commitment.
What is WizardLM-2 8x22B's context window?
The context window for WizardLM-2 8x22B is 64K tokens. That's the maximum amount of text you can feed into a single prompt, including system instructions, conversation history, and the actual query.
Is WizardLM-2 8x22B open source?
WizardLM-2 8x22B is fully open source. You can grab the weights, run it on your own hardware, and fine-tune it for specific tasks. That flexibility is a big deal for teams with strict data requirements.
What is WizardLM-2 8x22B best for?
The sweet spot for WizardLM-2 8x22B is: Complex instructions, reasoning, coding. If your workload fits one of these categories, it's worth benchmarking against alternatives.