WizardLM-2 8x22B
WizardLM-2 8x22B is Microsoft's entry in a crowded field. Context window: 0.064K tokens.
Context
64K
Input
Free (open)
Key Specifications
Arena Rank
Not disclosed
Context Window
64K
Input Price
per 1M tokens
Free (open)
Output Price
per 1M tokens
Free (open)
Parameters
176B (39B active)
Open Source
Best For
About WizardLM-2 8x22B
WizardLM-2 8x22B, developed by Microsoft, is an instruction-tuned Mixture-of-Experts model with 176 billion total parameters (39 billion active per token) and a 64K token context window. Built upon the Mixtral 8x22B architecture, it applies Microsoft's WizardLM training methodology to enhance complex instruction following, reasoning, and coding capabilities. The model demonstrates substantial improvements over its base on multi-step reasoning, structured output generation, and nuanced writing tasks. WizardLM-2 uses Evol-Instruct, a method that progressively evolves training instructions to increase complexity and diversity. Free and open-source, it can be deployed on enterprise multi-GPU setups. The model represents Microsoft's contribution to the open-source community through instruction-tuning research that advances the capability of existing base models without requiring new pre-training runs.
Pricing per 1M tokens
Input Tokens
Free (open)
Output Tokens
Free (open)
Compare WizardLM-2 8x22B
See how WizardLM-2 8x22B stacks up against other leading AI models