Mistral: Mixtral 8x7B Instruct
Description
Mixtral 8x7B Instruct is a pretrained generative Sparse Mixture of Experts model with 8 experts totaling 47 billion parameters. It has been fine-tuned by Mistral AI specifically for chat and instruction-following applications.
The model uses a Mixture of Experts architecture, which means that while it has 47B total parameters across all experts, only a subset of parameters are activated for each token, making it more computationally efficient than a dense model of equivalent size.
Technical Specifications
Context & Token Limits
| Parameter |
Value |
| Context Window |
32,000 tokens |
| Context Length |
32,768 tokens |
| Max Completion Tokens |
16,384 tokens |
| Total Parameter Count |
47 billion (8 experts) |
Supported Parameters
| Parameter |
Description |
max_tokens |
Maximum number of tokens to generate |
temperature |
Controls randomness (0.0 - 2.0) |
top_p |
Nucleus sampling threshold |
top_k |
Top-k sampling parameter |
stop |
Stop sequences |
frequency_penalty |
Penalize frequent tokens |
presence_penalty |
Penalize tokens based on presence |
repetition_penalty |
Penalize repeated tokens |
seed |
Random seed for reproducibility |
min_p |
Minimum probability threshold |
response_format |
Output format specification |
tools |
Tool/function definitions |
tool_choice |
Tool selection strategy |
Modalities
| Direction |
Supported Types |
| Input |
Text only |
| Output |
Text only |
Pricing
| Type |
Price |
| Input Tokens |
$0.54 per 1M tokens |
| Output Tokens |
$0.54 per 1M tokens |
Cost Examples
| Use Case |
Input Tokens |
Output Tokens |
Total Cost |
| Short query |
100 |
500 |
$0.000324 |
| Medium conversation |
1,000 |
2,000 |
$0.00162 |
| Long document processing |
10,000 |
5,000 |
$0.0081 |
| Full context usage |
32,768 |
16,384 |
$0.0265 |
Capabilities
- Instruction Following: Fine-tuned for following complex instructions
- Conversational AI: Optimized for multi-turn chat interactions
- Tool/Function Calling: Supports tool definitions and function calling via
tools and tool_choice parameters
- Multilingual: Supports multiple languages (French, German, Spanish, Italian, English)
- Code Generation: Capable of generating and understanding code
Limitations
- Does not support reasoning mode
- Text-only (no image or audio processing)
- No native vision capabilities
Use Cases
- Customer Support: Automated response generation and query handling
- Content Generation: Blog posts, articles, creative writing
- Code Assistance: Code generation, debugging, and explanation
- Translation: Multi-language translation tasks
- Summarization: Document and text summarization
- Data Extraction: Structured data extraction from unstructured text
From Mistral AI
| Model |
Description |
mistralai/mistral-7b-instruct |
Smaller 7B parameter instruct model |
mistralai/mistral-medium |
Medium-sized Mistral model |
mistralai/mistral-large |
Larger Mistral model |
mistralai/mixtral-8x22b |
Larger MoE model with 8x22B architecture |
Similar MoE Models
| Model |
Parameters |
Provider |
| Mixtral 8x22B |
141B |
Mistral AI |
| Grok-1 |
314B |
xAI |
Model Identity
| Property |
Value |
| Model Name |
Mistral: Mixtral 8x7B Instruct |
| Model ID |
mistralai/mixtral-8x7b-instruct |
| Creator |
Mistral AI |
| Release Date |
December 10, 2023 |
| Architecture |
Sparse Mixture of Experts (MoE) |
DeepInfra
API Usage
LangMart Endpoint
curl https://api.langmart.ai/v1/chat/completions \
-H "Authorization: Bearer $LANGMART_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "mistralai/mixtral-8x7b-instruct",
"messages": [
{
"role": "user",
"content": "Hello, how are you?"
}
]
}'
curl https://api.langmart.ai/v1/chat/completions \
-H "Authorization: Bearer $LANGMART_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "mistralai/mixtral-8x7b-instruct",
"messages": [
{
"role": "user",
"content": "What is the weather in Paris?"
}
],
"tools": [
{
"type": "function",
"function": {
"name": "get_weather",
"description": "Get current weather for a location",
"parameters": {
"type": "object",
"properties": {
"location": {
"type": "string",
"description": "City name"
}
},
"required": ["location"]
}
}
}
],
"tool_choice": "auto"
}'
Access Points
The Mixtral 8x7B model has demonstrated strong performance across various benchmarks:
- Competitive with or exceeding GPT-3.5 on many tasks
- Strong multilingual capabilities
- Efficient inference due to MoE architecture (only ~13B parameters active per forward pass)
- Good balance of quality and cost-effectiveness
Data sourced from LangMart.ai on December 23, 2025