# Evaluation Test Cases ## Create **post** `/v2/gen-ai/evaluation_test_cases` To create an evaluation test-case send a POST request to `/v2/gen-ai/evaluation_test_cases`. ### Returns - **test\_case\_uuid:** `string` Test‑case UUID. ## Retrieve **get** `/v2/gen-ai/evaluation_test_cases/{test_case_uuid}` To retrive information about an existing evaluation test case, send a GET request to `/v2/gen-ai/evaluation_test_case/{test_case_uuid}`. ### Returns - **evaluation\_test\_case:** `APIEvaluationTestCase` ## Update **put** `/v2/gen-ai/evaluation_test_cases/{test_case_uuid}` To update an evaluation test-case send a PUT request to `/v2/gen-ai/evaluation_test_cases/{test_case_uuid}`. ### Returns - **test\_case\_uuid:** `string` - **version:** `number` The new verson of the test case. ## List **get** `/v2/gen-ai/evaluation_test_cases` To list all evaluation test cases, send a GET request to `/v2/gen-ai/evaluation_test_cases`. ### Returns - **evaluation\_test\_cases:** `array of APIEvaluationTestCase` Alternative way of authentication for internal usage only - should not be exposed to public api ## List Evaluation Runs **get** `/v2/gen-ai/evaluation_test_cases/{evaluation_test_case_uuid}/evaluation_runs` To list all evaluation runs by test case, send a GET request to `/v2/gen-ai/evaluation_test_cases/{evaluation_test_case_uuid}/evaluation_runs`. ### Returns - **evaluation\_runs:** `array of APIEvaluationRun` List of evaluation runs. ## Domain Types ### API Evaluation Test Case - **APIEvaluationTestCase:** `object { archived_at, created_at, created_by_user_email, 15 more }` - **archived\_at:** `string` - **created\_at:** `string` - **created\_by\_user\_email:** `string` - **created\_by\_user\_id:** `string` - **dataset:** `object { created_at, dataset_name, dataset_uuid, 3 more }` - **created\_at:** `string` Time created at. - **dataset\_name:** `string` Name of the dataset. - **dataset\_uuid:** `string` UUID of the dataset. - **file\_size:** `string` The size of the dataset uploaded file in bytes. - **has\_ground\_truth:** `boolean` Does the dataset have a ground truth column? - **row\_count:** `number` Number of rows in the dataset. - **dataset\_name:** `string` - **dataset\_uuid:** `string` - **description:** `string` - **latest\_version\_number\_of\_runs:** `number` - **metrics:** `array of APIEvaluationMetric` - **name:** `string` - **star\_metric:** `APIStarMetric` - **test\_case\_uuid:** `string` - **total\_runs:** `number` - **updated\_at:** `string` - **updated\_by\_user\_email:** `string` - **updated\_by\_user\_id:** `string` - **version:** `number` ### API Star Metric - **APIStarMetric:** `object { metric_uuid, name, success_threshold, success_threshold_pct }` - **metric\_uuid:** `string` - **name:** `string` - **success\_threshold:** `number` The success threshold for the star metric. This is a value that the metric must reach to be considered successful. - **success\_threshold\_pct:** `number` The success threshold for the star metric. This is a percentage value between 0 and 100.