Groq: GPT-OSS 20B
Model Overview
| Property | Value |
|---|---|
| Model ID | groq/openai/gpt-oss-20b |
| Name | GPT-OSS 20B |
| Provider | Groq / OpenAI |
| Parameters | 20B |
Description
A lightweight 20 billion parameter version of OpenAI's open-weight GPT-OSS model. Optimized for fast inference on Groq infrastructure.
Specifications
| Spec | Value |
|---|---|
| Context Window | 131,072 tokens |
| Max Completion | 8,192 tokens |
| Inference Speed | ~1000 tokens/sec |
Pricing
| Type | Price |
|---|---|
| Input | $0.03 per 1M tokens |
| Output | $0.14 per 1M tokens |
Capabilities
- Fast Inference: Yes
- Cost Efficient: Yes
- Streaming: Yes
Use Cases
High-throughput applications, cost-sensitive deployments, general text generation.
Integration with LangMart
Gateway Support: Type 2 (Cloud), Type 3 (Self-hosted)
API Usage:
curl -X POST https://api.langmart.ai/v1/chat/completions \
-H "Authorization: Bearer sk-your-api-key" \
-H "Content-Type: application/json" \
-d '{
"model": "groq/openai/gpt-oss-20b",
"messages": [{"role": "user", "content": "Hello"}],
"max_tokens": 2048
}'
Related Models
- groq/openai/gpt-oss-120b - Full 120B version
- groq/openai/gpt-oss-safeguard-20b - Safety version
Last Updated: December 28, 2025