List Results
Retrieve Results of an Evaluation Run
agents.evaluation_runs.list_results(strevaluation_run_uuid, EvaluationRunListResultsParams**kwargs) -> evaluation_runAPIEvaluationRunlinksAPILinksmetaAPIMetapromptslistEvaluationRunListResultsResponse
get/v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results
To retrieve results of an evaluation run, send a GET request to /v2/gen-ai/evaluation_runs/{evaluation_run_uuid}/results
.
Parameters
evaluation_run_uuidstr
pageint
optional
Page number.
per_pageint
optional
Items per page.
Returns
EvaluationRunListResultsResponseclass
Gets the full results of an evaluation run with all prompts.
from do_gradientai import GradientAI
client = GradientAI()
response = client.agents.evaluation_runs.list_results(
evaluation_run_uuid="\"123e4567-e89b-12d3-a456-426614174000\"",
)
print(response.evaluation_run)
200 Example
{
"evaluation_run": {
"agent_deleted": true,
"agent_name": "\"example name\"",
"agent_uuid": "\"123e4567-e89b-12d3-a456-426614174000\"",
"agent_version_hash": "\"example string\"",
"agent_workspace_uuid": "\"123e4567-e89b-12d3-a456-426614174000\"",
"created_by_user_email": "[email protected]",
"created_by_user_id": "\"12345\"",
"error_description": "\"example string\"",
"evaluation_run_uuid": "\"123e4567-e89b-12d3-a456-426614174000\"",
"evaluation_test_case_workspace_uuid": "\"123e4567-e89b-12d3-a456-426614174000\"",
"finished_at": "2023-01-01T00:00:00Z",
"pass_status": true,
"queued_at": "2023-01-01T00:00:00Z",
"run_level_metric_results": [
{
"error_description": "\"example string\"",
"metric_name": "\"example name\"",
"metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED",
"number_value": 123,
"reasoning": "\"example string\"",
"string_value": "\"example string\""
}
],
"run_name": "\"example name\"",
"star_metric_result": {
"error_description": "\"example string\"",
"metric_name": "\"example name\"",
"metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED",
"number_value": 123,
"reasoning": "\"example string\"",
"string_value": "\"example string\""
},
"started_at": "2023-01-01T00:00:00Z",
"status": "EVALUATION_RUN_STATUS_UNSPECIFIED",
"test_case_description": "\"example string\"",
"test_case_name": "\"example name\"",
"test_case_uuid": "\"123e4567-e89b-12d3-a456-426614174000\"",
"test_case_version": 123
},
"links": {
"pages": {
"first": "\"example string\"",
"last": "\"example string\"",
"next": "\"example string\"",
"previous": "\"example string\""
}
},
"meta": {
"page": 123,
"pages": 123,
"total": 123
},
"prompts": [
{
"ground_truth": "\"example string\"",
"input": "\"example string\"",
"input_tokens": "\"12345\"",
"output": "\"example string\"",
"output_tokens": "\"12345\"",
"prompt_chunks": [
{
"chunk_usage_pct": 123,
"chunk_used": true,
"index_uuid": "\"123e4567-e89b-12d3-a456-426614174000\"",
"source_name": "\"example name\"",
"text": "\"example string\""
}
],
"prompt_id": 123,
"prompt_level_metric_results": [
{
"error_description": "\"example string\"",
"metric_name": "\"example name\"",
"metric_value_type": "METRIC_VALUE_TYPE_UNSPECIFIED",
"number_value": 123,
"reasoning": "\"example string\"",
"string_value": "\"example string\""
}
]
}
]
}