reasoning_content field, distinct from the final response.
Supported models
| Model | Slug | Reasoning |
|---|---|---|
| DeepSeek V3.2 | deepseek-ai/DeepSeek-V3.2 | Enabled by default |
| DeepSeek V3.1 | deepseek-ai/DeepSeek-V3.1 | Enabled by default |
| DeepSeek V3 0324 | deepseek-ai/DeepSeek-V3-0324 | Enabled by default |
| Kimi K2 Thinking | moonshotai/Kimi-K2-Thinking | Always enabled |
| GLM 4.7 | zai-org/GLM-4.7 | Enabled by default |
| GLM 4.6 | zai-org/GLM-4.6 | Enabled by default |
Control reasoning depth
Thereasoning_effort parameter controls how thoroughly the model reasons through a problem.
| Value | Behavior |
|---|---|
low | Faster responses, less thorough reasoning |
medium | Balanced (default) |
high | Slower responses, more thorough reasoning |
- Python
- JavaScript
- cURL
Pass
reasoning_effort through extra_body since it extends the standard OpenAI API:Parse the response
The modelβs thinking process appears inreasoning_content, separate from the final answer in content.
reasoning_tokens field in completion_tokens_details shows how many tokens the model used for reasoning. These tokens count toward your total usage and billing.
Decide when to reason
Reasoning improves quality for tasks that benefit from step-by-step thinking: mathematical calculations, multi-step logic problems, code generation with complex requirements, and analysis requiring multiple considerations. For straightforward tasks like simple Q&A or text generation, reasoning adds latency and token cost without improving quality. In these cases, use a model without reasoning support or setreasoning_effort to low.