跳转到主要内容

推荐 endpoint

最小请求

{
  "model": "gpt-4.1",
  "messages": [
    { "role": "user", "content": "边生成边解释 SSE 流式输出。" }
  ],
  "stream": true
}

cURL 示例

curl https://mass.apigo.ai/v1/chat/completions \
  -H "Authorization: Bearer $TIDEMIND_API_KEY" \
  -H "Content-Type: application/json" \
  -N \
  -d '{
    "model": "gpt-4.1",
    "messages": [
      { "role": "user", "content": "边生成边解释 SSE 流式输出。" }
    ],
    "stream": true
  }'

Python 示例

from openai import OpenAI

client = OpenAI(
    base_url="https://mass.apigo.ai/v1",
    api_key="<TIDEMIND_API_KEY>",
)

stream = client.chat.completions.create(
    model="gpt-4.1",
    messages=[{"role": "user", "content": "边生成边解释 SSE 流式输出。"}],
    stream=True,
)

for chunk in stream:
    delta = chunk.choices[0].delta.content or ""
    print(delta, end="")

Node.js 示例

import OpenAI from "openai";

const client = new OpenAI({
  baseURL: "https://mass.apigo.ai/v1",
  apiKey: process.env.TIDEMIND_API_KEY,
});

const stream = await client.chat.completions.create({
  model: "gpt-4.1",
  messages: [
    { role: "user", content: "边生成边解释 SSE 流式输出。" }
  ],
  stream: true,
});

for await (const chunk of stream) {
  process.stdout.write(chunk.choices[0]?.delta?.content || "");
}

最佳实践

  • 前端按 chunk 增量渲染,不要等完整 JSON 才显示
  • 如果后面要接工具和结构化输出,优先评估 responses 的流式事件模型
  • 服务端统一处理断流重试和最终结果拼接