O

OpenChat 3.5 7B (Free)

Openchat
Reasoning Vision
8K
Context
Free
Input /1M
Free
Output /1M
N/A
Max Output

OpenChat 3.5 7B (Free)

Overview

OpenChat 3.5 7B is a high-performance open-source language model fine-tuned from Mistral-7B-v0.1 using C-RLFT (Conditioned Reinforcement Learning Fine-Tuning). It represents one of the best-performing open-source 7B models, outperforming many larger models on various benchmarks.

Description

OpenChat is a library of open-source language models fine-tuned with C-RLFT, a strategy inspired by offline reinforcement learning. The model is trained on mixed-quality data without preference labels, achieving remarkable performance that rivals or exceeds many proprietary models.

Key innovations:

  • C-RLFT Training: Conditioned Reinforcement Learning Fine-Tuning methodology
  • Mixed-Quality Data: Training without requiring preference labels
  • Research-Backed: Based on research paper arXiv:2309.11235

Technical Specifications

Specification Value
Context Length 8,192 tokens
Input Modalities Text
Output Modalities Text
Tensor Type BF16
Format Safetensors
Base Model Group Mistral
Instruction Type OpenChat
Default Stop Sequences `<
Minimum GPU RAM 24GB

Capabilities

Capability Supported
Chat/Conversation Yes
Code Generation Yes (Strong)
Mathematical Reasoning Yes (Specialized Mode)
General Knowledge Yes
Instruction Following Yes
Multi-turn Conversations Yes
Function Calling No
Vision/Image Input No
Reasoning Traces No

Supported Parameters

Parameter Description Recommended Range
temperature Controls randomness 0.0 - 2.0
top_p Nucleus sampling 0.0 - 1.0
top_k Top-k sampling 1 - 100
max_tokens Maximum output tokens 1 - 8192
stop Stop sequences `["<
frequency_penalty Reduce repetition -2.0 - 2.0
presence_penalty Encourage diversity -2.0 - 2.0
Model Description
openchat/openchat-7b Base OpenChat 7B (fine-tuned on Mistral 7B)
openchat/openchat-8b Newer version fine-tuned on Llama 8B
mistralai/mistral-7b-instruct Base Mistral model

Model Information

Property Value
Model Name OpenChat 3.5 7B
Model ID openchat/openchat-7b:free
HuggingFace Model openchat/openchat-3.5-0106
Author OpenChat
Parameter Count 7 Billion
Base Model Mistral-7B-v0.1
Created November 28, 2023
Last Updated November 10, 2025
License Apache-2.0

Pricing (OpenRouter Free Tier)

Type Price
Input Free ($0/1M tokens)
Output Free ($0/1M tokens)

Note: Free tier models on LangMart may have rate limits and usage restrictions.

Performance Benchmarks

Overall Performance Comparison

Model Avg Score MT-Bench HumanEval BBH MC AGIEval TruthfulQA MMLU GSM8K BBH CoT
OpenChat-3.5-0106 64.5 7.8 71.3 51.5 49.1 61.0 65.8 77.4 62.2
ChatGPT (March) 61.5 7.94 48.1 47.6 47.1 57.7 67.3 74.9 70.1
OpenChat-3.5-1210 63.8 7.76 68.9 49.5 48.0 61.8 65.3 77.3 61.8

HumanEval+ (Coding Benchmark)

Model Pass@1 Parameters
OpenChat-3.5-0106 65.9% 7B
ChatGPT (Dec 2023) 64.6% Unknown
WizardCoder-Python-34B 64.6% 34B

Key Performance Highlights

  • Outperforms Grok-0 (33B) on all 4 tested benchmarks
  • Outperforms Grok-1 on average and 3 out of 4 benchmarks
  • 15-point improvement in coding over previous OpenChat-3.5 version
  • Best performing open-source 7B model in its class

Operating Modes

1. Default Mode (GPT4 Correct)

Best for coding, chat, and general tasks.

Conversation Template:

GPT4 Correct User: [user_message]<|end_of_turn|>GPT4 Correct Assistant: [assistant_response]<|end_of_turn|>

2. Mathematical Reasoning Mode

Tailored for solving math problems.

Conversation Template:

Math Correct User: [math_problem]<|end_of_turn|>Math Correct Assistant: [solution]<|end_of_turn|>

Usage Examples

LangMart API Request

curl https://api.langmart.ai/v1/chat/completions \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer $LANGMART_API_KEY" \
  -d '{
    "model": "openchat/openchat-7b:free",
    "messages": [
      {"role": "user", "content": "Explain the concept of machine learning in simple terms."}
    ]
  }'

Python with OpenAI SDK

from openai import OpenAI

client = OpenAI(
    base_url="https://api.langmart.ai/v1",
    api_key="your-langmart-api-key"
)

response = client.chat.completions.create(
    model="openchat/openchat-7b:free",
    messages=[
        {"role": "user", "content": "Write a Python function to calculate fibonacci numbers."}
    ]
)

print(response.choices[0].message.content)

Self-Hosted Deployment

python -m ochat.serving.openai_api_server \
  --model openchat/openchat-3.5-0106 \
  --engine-use-ray \
  --worker-use-ray

Server runs at localhost:18888 with OpenAI-compatible API.

Known Limitations

  • Complex multi-step reasoning tasks may be challenging
  • May hallucinate non-existent information
  • Potential for generating harmful or biased outputs
  • Some mathematical and arithmetic problems may produce incorrect results
  • Limited context window (8K tokens) compared to newer models

Citation

@article{wang2023openchat,
  title={OpenChat: Advancing Open-source Language Models with Mixed-Quality Data},
  author={Wang, Guan and Cheng, Sijie and Zhan, Xianyuan and Li, Xiangang and Song, Sen and Liu, Yang},
  journal={arXiv preprint arXiv:2309.11235},
  year={2023}
}

Resources


Last Updated: December 2025