Reasoning Tokens

Use reasoning intentionally for quality and cost control

Reasoning-enabled models can improve answer quality for complex tasks, but they consume extra output tokens. Treat reasoning as a controllable budget, not a default.

Reference payload

Use this baseline request shape and adapt model, provider sort strategy, and token limits to your workload.

request.json
{
  "model": "anthropic/claude-sonnet-4.5",
  "reasoning": {
    "max_tokens": 2000,
    "exclude": false
  }
}

Configuration examples

TheRouter.ai keeps request semantics consistent across providers, so you can tune behavior without rewriting your app layer.

TypeScript
const payload = {
  model: "openai/o3",
  messages: [{ role: "user", content: "Solve this scheduling problem" }],
  reasoning: { effort: "high", exclude: false },
};

Production note

Operate with guardrails
Higher reasoning budgets can reduce hallucinations on hard tasks, but can significantly increase output-token spend.

Use the activity feed and usage exports to validate that these settings improve reliability and cost in your real traffic mix.