M

Mistral Nemo

Mistral AI

ChatEfficient

About

12B parameter model built in collaboration with NVIDIA. Delivers strong performance for its size with 128k context length. Good balance of speed and capability, making it ideal for applications that need to balance quality with low latency.

Provider

Mistral AI

Context Window

128k tokens

Max Output

4,096 tokens

Model ID

mistralai/mistral-nemo

Capabilities

128k contextEfficient inferenceStreaming

API Usage

Drop-in compatible with any OpenAI client library.

import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://api.deployai.dev/v1",
  apiKey: process.env.DEPLOYAI_API_KEY,
});

const completion = await client.chat.completions.create({
  model: "mistralai/mistral-nemo",
  messages: [
    { role: "user", content: "Hello, how are you?" }
  ],
});

console.log(completion.choices[0].message.content);

Related Models

View all