推荐 endpoint
最小请求
{
"model": "claude-sonnet-4-20250514",
"max_tokens": 1600,
"thinking": {
"type": "enabled",
"budget_tokens": 2048
},
"messages": [
{
"role": "user",
"content": [{ "type": "text", "text": "比较三种缓存架构的取舍,并给出推荐。" }]
}
]
}
cURL 示例
curl https://api.anthropic.com/v1/messages \
-H "x-api-key: $ANTHROPIC_API_KEY" \
-H "anthropic-version: 2023-06-01" \
-H "content-type: application/json" \
-d '{
"model": "claude-sonnet-4-20250514",
"max_tokens": 1600,
"thinking": {
"type": "enabled",
"budget_tokens": 2048
},
"messages": [
{
"role": "user",
"content": [
{ "type": "text", "text": "比较三种缓存架构的取舍,并给出推荐。" }
]
}
]
}'
Python 示例
from anthropic import Anthropic
client = Anthropic(api_key="<ANTHROPIC_API_KEY>")
response = client.messages.create(
model="claude-sonnet-4-20250514",
max_tokens=1600,
thinking={"type": "enabled", "budget_tokens": 2048},
messages=[
{
"role": "user",
"content": [
{"type": "text", "text": "比较三种缓存架构的取舍,并给出推荐。"}
],
}
],
)
print(response.content[-1].text)
Node.js 示例
import Anthropic from "@anthropic-ai/sdk";
const client = new Anthropic({
apiKey: process.env.ANTHROPIC_API_KEY,
});
const response = await client.messages.create({
model: "claude-sonnet-4-20250514",
max_tokens: 1600,
thinking: {
type: "enabled",
budget_tokens: 2048,
},
messages: [
{
role: "user",
content: [
{ type: "text", text: "比较三种缓存架构的取舍,并给出推荐。" }
]
}
]
});
console.log(response.content.at(-1).text);
最佳实践
- 只在真正需要复杂推理时开启 thinking
budget_tokens从小值开始调,不要默认开很大- 如果后面还要接工具调用,先验证 thinking 与工具链路的事件处理
