reasoning_content field, distinct from the final response.
Supported models
| Model | Slug | Reasoning |
|---|---|---|
| DeepSeek V3.1 | deepseek-ai/DeepSeek-V3.1 | Enabled by default |
| DeepSeek V3 0324 | deepseek-ai/DeepSeek-V3-0324 | Enabled by default |
| Minimax M2.5 | MiniMaxAI/MiniMax-M2.5 | Enabled by default |
| OpenAI GPT OSS 120B | openai/gpt-oss-120b | Enabled by default |
| Kimi K2.5 | moonshotai/Kimi-K2.5 | Opt-in via chat_template_args |
| GLM 4.7 | zai-org/GLM-4.7 | Opt-in via chat_template_args |
| GLM 4.6 | zai-org/GLM-4.6 | Opt-in via chat_template_args |
reasoning_effort.
Models not listed here don’t support reasoning.
Enable thinking
Enable thinking for Kimi K2.5 and GLM models by passingchat_template_args.
- Python
- JavaScript
- cURL
Pass
chat_template_args through extra_body since it extends the standard OpenAI API:Control reasoning depth
Thereasoning_effort parameter controls how thoroughly the model reasons through a problem.
Currently, only GPT OSS 120B supports this parameter.
| Value | Behavior |
|---|---|
low | Faster responses, less thorough reasoning |
medium | Balanced (default) |
high | Slower responses, more thorough reasoning |
- Python
- JavaScript
- cURL
Pass
reasoning_effort through extra_body since it extends the standard OpenAI API:Parse the response
The model’s thinking process appears inreasoning_content, separate from the final answer in content. Both fields are returned on the message object.
completion_tokens and count toward your total usage and billing.
Decide when to reason
Reasoning improves quality for tasks that benefit from step-by-step thinking: mathematical calculations, multi-step logic problems, code generation with complex requirements, and analysis requiring multiple considerations. For straightforward tasks like simple Q&A or text generation, reasoning adds latency and token cost without improving quality. In these cases, use a model without reasoning support or setreasoning_effort to low.