Anthropic Extended Thinking
OpenAI Reasoning Models (o-series)
OpenAI’s o-series models reason natively — no configuration needed:o1, o3, o3-mini, o4-mini
Reasoning models don’t support
temperature, top_p, frequency_penalty, presence_penalty, or stop. These parameters are automatically stripped.Gemini Thinking
Gemini 2.5+ models support thinking with a token budget:xAI Reasoning Models
xAI Grok models with-reasoning in the name use chain-of-thought:
Thinking in Streaming
When usingprocessStream(), thinking content arrives as separate chunks:
Provider Support
| Provider | How | Configuration |
|---|---|---|
| Anthropic | thinking parameter | thinking: { type: 'enabled', budgetTokens: N } |
| OpenAI | Native (o-series models) | Just use o1/o3/o4-mini models |
| Gemini | thinkingConfig | thinking: { type: 'enabled', budgetTokens: N } |
| xAI | Reasoning model names | Use grok-*-reasoning models |
| Groq | Not supported | - |
| Bedrock | Not supported | - |
Next Steps
Structured Output
Get guaranteed JSON responses
Streaming
Real-time streaming responses