Arize
Send TheRouter.ai traces to Arize AX for evaluation and monitoring
Arize is built for LLM evaluation and model observability. TheRouter.ai sends OTLP traces aligned with OpenInference semantics so you can inspect generations, costs, and latency.
broadcast-arize-config.json
{
"destination": "arize",
"enabled": true,
"api_key": "arize_...",
"space_key": "space_...",
"model_id": "support-assistant-prod",
"base_url": "https://otlp.arize.com",
"sampling_rate": 1
}cURL
curl https://api.therouter.ai/v1/chat/completions -H "Authorization: Bearer $THEROUTER_API_KEY" -H "Content-Type: application/json" -H "X-TheRouter.ai-Broadcast: true" -d '{
"model": "anthropic/claude-sonnet-4.5",
"messages": [{"role":"user","content":"Classify this ticket"}],
"trace": {"trace_name":"Arize Ingest Test","dataset":"support"}
}'Verification
Check Arize for a new trace under your configured model ID and confirm token/cost fields are populated in span attributes.