12B parameter model built in collaboration with NVIDIA. Delivers strong performance for its size with 128k context length. Good balance of speed and capability, making it ideal for applications that need to balance quality with low latency.
Mistral AI
128k tokens
4,096 tokens
mistralai/mistral-nemo
Drop-in compatible with any OpenAI client library.
import OpenAI from "openai";
const client = new OpenAI({
baseURL: "https://api.deployai.dev/v1",
apiKey: process.env.DEPLOYAI_API_KEY,
});
const completion = await client.chat.completions.create({
model: "mistralai/mistral-nemo",
messages: [
{ role: "user", content: "Hello, how are you?" }
],
});
console.log(completion.choices[0].message.content);Flagship reasoning model with top-tier multilingual support, function calling, and 128k context.
Most capable multimodal model. Excels at complex reasoning, coding, creative writing, and vision tasks.
Affordable and efficient small model for fast, focused tasks. Great cost-to-performance ratio.