LLM·Dex
Open weightsMistralsmalltext

Mistral Nemo

12B model co-built with Nvidia, strong small-model multilingual performance.

Updated


Quick facts

Released
Jul 2024
Context
128K tokens
Output / 1M
Pricing not published
License
Apache-2.0

About Mistral Nemo

Mistral Nemo is a 12B model co-developed with Nvidia, designed for efficient inference on a single GPU. Apache-2.0 licensed, with strong multilingual coverage and FIM support for code use cases.

The Nvidia-optimized inference stack means Nemo runs at high throughput on H100 and L40S hardware, making it a popular default for European and US-EU enterprise self-hosting.

Benchmarks

Published scores from Mistral's model card or independent leaderboards. We do not publish numbers we cannot source, see methodology.

HumanEval
,
Python coding pass@1
MMLU
,
Broad academic knowledge
GPQA
,
Graduate-level reasoning
SWE-bench
,
Real software-engineering tasks
Benchmark scores not yet available. We only publish numbers we can source from official model cards or independent leaderboards, see methodology.

Capabilities

Strengths

  • Apache-2.0
  • Single-GPU fit
  • Multilingual

Tracked weaknesses

  • Quality limited by 12B size

Pricing

Per-million-token rates as published by Mistral.

Per-token pricing not yet published for Mistral Nemo. Check the official provider site for current tiers.

Call Mistral Nemo from your code

Drop-in snippet for the Mistral SDK. Set your API key in the environment and run.

typescript
import OpenAI from "openai";

const client = new OpenAI({
  // Use OPENAI_API_KEY for OpenAI, or your provider's key + baseURL.
  apiKey: process.env.OPENAI_API_KEY,
});

const completion = await client.chat.completions.create({
  model: "mistral-nemo",
  messages: [
    { role: "user", content: "What's the time complexity of quicksort?" },
  ],
});

console.log(completion.choices[0].message.content);

Best for

Tasks where Mistral Nemo ranks among LLMDex's top picks.

Compare Mistral Nemo with…

Frequently asked

  • How much does Mistral Nemo cost?
    Mistral has not published per-token API pricing for Mistral Nemo at the time of writing. Check the official site for current pricing tiers, or compare against alternative models on LLMDex.
  • What is Mistral Nemo's context window?
    Mistral Nemo supports a context window of 128K tokens.
  • Is Mistral Nemo open source?
    Mistral Nemo ships with open weights under the Apache-2.0 license. You can self-host it, fine-tune it, and (subject to the license terms) deploy it commercially.
  • When was Mistral Nemo released?
    Mistral Nemo was released on Jul 18, 2024 by Mistral.
Friday digest

Intelligence, distilled weekly.

One short email every Friday, new model launches, leaderboard moves, and pricing drops. Curated by hand. Free, no spam.