Skip to main content
GET
/
eval-runs
/
{eval_run_id}
{
  "id": "a03fa2f4-900d-482d-afe0-470d4cd8d1f4",
  "agent_id": "basic-agent",
  "model_id": "gpt-4o",
  "model_provider": "OpenAI",
  "name": "Test ",
  "eval_type": "reliability",
  "eval_data": {
    "eval_status": "PASSED",
    "failed_tool_calls": [],
    "passed_tool_calls": [
      "multiply"
    ]
  },
  "eval_input": {
    "expected_tool_calls": [
      "multiply"
    ]
  },
  "created_at": "2025-08-27T15:41:59Z",
  "updated_at": "2025-08-27T15:41:59Z"
}

Authorizations

Authorization
string
header
required

Bearer authentication header of the form Bearer <token>, where <token> is your auth token.

Path Parameters

eval_run_id
string
required

Query Parameters

db_id
string | null

The ID of the database to use

Response

Evaluation run details retrieved successfully

id
string
required
eval_type
enum<string>
required
Available options:
accuracy,
performance,
reliability
eval_data
object
required
agent_id
string | null
model_id
string | null
model_provider
string | null
team_id
string | null
workflow_id
string | null
name
string | null
evaluated_component_name
string | null
eval_input
object | null
created_at
string<date-time> | null
updated_at
string<date-time> | null
I