Mixtral Models by MistralAI
About
The Mixtral family of large language models (LLMs), developed by Mistral AI, offers a groundbreaking approach in open-source AI through a sparse mixture-of-experts (SMoE) architecture. This innovative design allows the models to manage a significant number of parameters while ensuring efficient inference speed by activating only a subset of parameters for each token. Such architecture enables Mixtral models to deliver performance on par with much larger models, standing out in various benchmarks and outperforming competitors like Llama 2, and even equaling the prowess of closed-source models such as GPT-3.5. These models are multilingual, supporting languages such as English, French, Italian, German, and Spanish, and excel in domains like code generation. Instruction-tuned versions like Mixtral-8x7B-Instruct-v0.1 cater to applications requiring robust instruction-following and chat capabilities. The Mixtral family provides versatile models of differing sizes, addressing diverse computational and application requirements.
Specifications(5 models)
| Model | Released | Context | Parameters | Fn Calling |
|---|---|---|---|---|
| Mixtral 8x22B Instruct v0.3 | 2024-07 | 64K | 8x22B | Yes |
| Mixtral 8x22B v0.1 | 2024-04 | 64K | 8x22B | No |
| Mixtral 8x22B Instruct v0.1 | 2024-04 | 64K | 8x22B | No |
| Mixtral 8x7B | 2023-12 | 32K | 8x7B | No |
| Mixtral 8x7B Instruct v0.1 | 2023-12 | 33K | 56B | No |
Available From(22 providers)
Pricing
Frequently Asked Questions
- What is Mixtral used for?
- Mixtral is used for agent workflows and tool use and coding. The family description and listed model capabilities point to those workloads as the best fit.
- How does Mixtral compare to Ministral?
- Mixtral by MistralAI is strongest where you need agent workflows and tool use, while Ministral by MistralAI is the closest related family to check for structured outputs. Mixtral has 5 listed variants and reaches up to 64K context, while Ministral reaches up to 32K context, so compare the specs and pricing tables before choosing a production model.
- Which Mixtral model should I use?
- For the lowest listed input price, start with Mixtral 8x7B through Mistral AI Studio at $0.15/1M input tokens. For the most capable/latest local choice, evaluate Mixtral 8x22B Instruct v0.3 with 64K context and function calling.





