Create Evaluation
Create a Evaluation
Send over a input that a AI generated a response for, and the model will generate an evaluation based on some desired metrics.
The evaluate API can be used for either single metric evaluations or multiple metric evaluations. It is also capable of evaluating in RAG settings with an optional context field and against a ground-truth answer if one is available.
Authorizations
Bearer authentication header of the form Bearer <token>
, where <token>
is your auth token.
Body
A request to an Atla evaluator via the /eval
endpoint.
The ID or name of the Atla evaluator model to use. This may point to a specific model version or a model family. If a model family is provided, the default model version for that family will be used.
The input given to a model which produced the model_output
to be evaluated.
The output of the model which is being evaluated. This is the model_output
from the model_input
.
The criteria used to evaluate the model_output
. Only one of evaluation_criteria
or metric_name
can be provided.
An optional reference ("ground-truth" / "gold standard") answer against which to evaluate the model_output
.
A list of few-shot examples for the evaluation.
The name of the metric to use for the evaluation. Only one of evaluation_criteria
or metric_name
can be provided.
Any additional context provided to the model which received the model_input
and produced the model_output
.