AllenAI: Olmo 3.1 32B Think (free)
Description
This is a 32-billion parameter model emphasizing reasoning capabilities. The system excels at deep reasoning, complex multi-step logic, and advanced instruction following. Version 3.1 represents improvements over earlier iterations with refined reasoning behavior and stronger performance across demanding evaluations.
The model is designed for complex problem-solving requiring explicit reasoning chains before final responses. All responses include internal thinking between <think> tags (mandatory reasoning mode).
Technical Specifications
| Specification | Value |
|---|---|
| Context Window | 65,536 tokens |
| Context Length | 65,536 tokens |
| Input Modalities | Text only |
| Output Modalities | Text only |
| Quantization | BF16 |
| Parameters | 32 billion |
Pricing
| Type | Price |
|---|---|
| Input | $0 per 1M tokens |
| Output | $0 per 1M tokens |
Note: This model is completely free to use.
Supported Parameters
Generation Parameters
- Temperature
- Top-P
- Top-K
- Min-P
- Seed
- Stop sequences
Penalty Parameters
- Frequency Penalty
- Presence Penalty
- Repetition Penalty
Advanced Features
- Logit bias
- Structured outputs
- Response formatting
- Reasoning (mandatory)
Tool Support
- Tool choice support:
none,auto,required - Function calling capabilities enabled
- Multipart support enabled
Use Cases
- Complex mathematical problem solving
- Multi-step logical reasoning
- Code analysis and debugging
- Detailed explanations and tutorials
- Research and analysis tasks
- Tasks requiring transparent decision-making processes
Key Features
Mandatory Reasoning
All responses include internal thinking between <think> tags. This reasoning chain is visible in the output and allows users to understand the model's thought process before receiving the final answer.
Tool/Function Calling
Full support for tool and function calling capabilities, enabling integration with external systems and APIs.
Open Source
Licensed under Apache 2.0, allowing for commercial use, modification, and distribution.
Limitations
- Text-only input and output (no image/audio support)
- Rate limited to 50 requests per minute on free tier
- All responses include reasoning overhead (increased token usage)
API Usage Example
curl https://api.langmart.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $LANGMART_API_KEY" \
-d '{
"model": "allenai/olmo-3.1-32b-think:free",
"messages": [
{
"role": "user",
"content": "Solve this step by step: What is 15% of 240?"
}
]
}'
Related Models
allenai/olmo-3.1-32b- Base version without mandatory reasoning- Other reasoning-focused models on LangMart
Basic Information
| Property | Value |
|---|---|
| Model ID | allenai/olmo-3.1-32b-think:free |
| Full Name | AllenAI: Olmo 3.1 32B Think (free) |
| Author | AllenAI |
| Created | December 16, 2025 |
| License | Apache 2.0 |
Provider Details
| Property | Value |
|---|---|
| Primary Provider | Parasail |
| Provider Model ID | parasail-olmo-31-32b-think |
| Rate Limits | 50 requests per minute |
| Data Policy | No training usage; no prompt retention |
Usage Statistics (December 23, 2025)
| Metric | Value |
|---|---|
| Total Requests | 9,840 |
| Prompt Tokens | 119.9M |
| Reasoning Tokens | 23.1M |
| Completion Tokens | 23.1M |
Response Format
The model returns responses with reasoning wrapped in <think> tags:
<think>
To find 15% of 240, I need to:
1. Convert 15% to a decimal: 15/100 = 0.15
2. Multiply 240 by 0.15
3. 240 * 0.15 = 36
</think>
15% of 240 is 36.
Source
- LangMart: https://langmart.ai/model-docs