Evaluation Runs
Evaluation Runs
Evaluation Runs
Run an Evaluation Test Case
Retrieve Results of an Evaluation Run
Retrieve Information About an Existing Evaluation Run
Retrieve Results of an Evaluation Run Prompt
API Evaluation Metric
ShowShow
If true, the metric is inverted, meaning that a lower value is better.
Hide ParametersShow Parameters
Hide ParametersShow Parameters
The maximum value for the metric.
The minimum value for the metric.
API Evaluation Metric Result
ShowShow
Error description if the metric could not be calculated.
Metric name
Hide ParametersShow Parameters
The value of the metric as a number.
Reasoning of the metric result.
The value of the metric as a string.
API Evaluation Prompt
ShowShow
The ground truth for the prompt.
The number of input tokens used in the prompt.
The number of output tokens used in the prompt.
The list of prompt chunks.
Hide ParametersShow Parameters
The usage percentage of the chunk.
Indicates if the chunk was used in the prompt.
The index uuid (Knowledge Base) of the chunk.
The source name for the chunk, e.g., the file name or document title.
Text content of the chunk.
Prompt ID
The metric results for the prompt.
API Evaluation Run
ShowShow
Whether agent is deleted
Agent name
Agent UUID.
Version hash
Agent workspace uuid
The error description
Evaluation run UUID.
Evaluation test case workspace uuid
Run end time.
The pass status of the evaluation run based on the star metric.
Run queued time.
Run name.
Run start time.
Evaluation Run Statuses
Hide ParametersShow Parameters
Test case description.
Test case name.
Test-case UUID.
Test-case-version.