Skip to main content

Minimal request

{
  "model": "gemini-2.5-flash",
  "systemInstruction": {
    "parts": [{ "text": "You are a concise assistant." }]
  },
  "contents": [
    {
      "role": "user",
      "parts": [{ "text": "Summarize Api.Go in three sentences." }]
    }
  ]
}

cURL example

curl "https://generativelanguage.googleapis.com/v1beta/models/gemini-2.5-flash:generateContent" \
  -H "x-goog-api-key: $GEMINI_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "systemInstruction": {
      "parts": [{ "text": "You are a concise assistant." }]
    },
    "contents": [
      {
        "role": "user",
        "parts": [{ "text": "Summarize Api.Go in three sentences." }]
      }
    ]
  }'

Python example

from google import genai
from google.genai import types

client = genai.Client(api_key="<GEMINI_API_KEY>")

response = client.models.generate_content(
    model="gemini-2.5-flash",
    contents="Summarize Api.Go in three sentences.",
    config=types.GenerateContentConfig(
        system_instruction="You are a concise assistant."
    ),
)

print(response.text)

Node.js example

import { GoogleGenAI } from "@google/genai";

const ai = new GoogleGenAI({ apiKey: process.env.GEMINI_API_KEY });

const response = await ai.models.generateContent({
  model: "gemini-2.5-flash",
  contents: "Summarize Api.Go in three sentences.",
  config: {
    systemInstruction: "You are a concise assistant."
  }
});

console.log(response.text);

Best practices

  • Start with a Flash model for lower-latency chat
  • Keep the parts model intact so media can be added later without redesign
  • If you do not need extra reasoning cost, set thinkingBudget to 0 on Gemini 2.5 Flash