O

LangMart: Mistral: Mistral Nemo

Openrouter
131K
Context
$0.0200
Input /1M
$0.0400
Output /1M
N/A
Max Output

LangMart: Mistral: Mistral Nemo

Model Overview

Property Value
Model ID openrouter/mistralai/mistral-nemo
Name Mistral: Mistral Nemo
Provider mistralai
Released 2024-07-19

Description

A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA.

The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi.

It supports function calling and is released under the Apache 2.0 license.

Description

LangMart: Mistral: Mistral Nemo is a language model provided by mistralai. This model offers advanced capabilities for natural language processing tasks.

Provider

mistralai

Specifications

Spec Value
Context Window 131,072 tokens
Modalities text->text
Input Modalities text
Output Modalities text

Pricing

Type Price
Input $0.02 per 1M tokens
Output $0.04 per 1M tokens

Capabilities

  • Frequency penalty
  • Max tokens
  • Min p
  • Presence penalty
  • Repetition penalty
  • Response format
  • Seed
  • Stop
  • Structured outputs
  • Temperature
  • Tool choice
  • Tools
  • Top k
  • Top p

Detailed Analysis

Mistral NeMo (July 2024) is a 12B multilingual model developed in collaboration with NVIDIA, featuring strong reasoning, world knowledge, and coding capabilities with native function calling support. The NVIDIA partnership ensures exceptional optimization for NVIDIA hardware (GPUs, AI accelerators) and TensorRT inference, delivering superior performance-per-watt. NeMo targets the efficiency sweet spot between 7B (limited capability) and 24B (higher cost), offering excellent quality at 12B parameter efficiency. The model excels at multilingual applications spanning 50+ languages, reasoning tasks requiring logical chains, knowledge-intensive queries, and code generation/understanding. The 128K context window enables processing extensive documents, entire code repositories, and long conversations. NeMo is particularly well-suited for NVIDIA deployments, edge-to-cloud workflows, multilingual customer service, code-aware assistants, and applications requiring balanced intelligence and efficiency. The July 2024 release ensures current knowledge while maintaining open-source accessibility (Apache 2.0). Recommended for users with NVIDIA infrastructure or requiring strong multilingual 12B performance.