G

Gemini 2.0 Flash

Google

ChatCodeVisionFastFast

About

Google's ultra-fast multimodal model with native tool use, code execution, and the largest context window available at 1 million tokens. Excels at processing very long documents, multimodal inputs, and tasks requiring code execution within the model's response.

Provider

Google

Context Window

1M tokens

Max Output

8,192 tokens

Model ID

google/gemini-2.0-flash

Capabilities

1M token contextMultimodalCode executionGroundingStreaming

API Usage

Drop-in compatible with any OpenAI client library.

import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://api.deployai.dev/v1",
  apiKey: process.env.DEPLOYAI_API_KEY,
});

const completion = await client.chat.completions.create({
  model: "google/gemini-2.0-flash",
  messages: [
    { role: "user", content: "Hello, how are you?" }
  ],
});

console.log(completion.choices[0].message.content);

Related Models

View all