Ground Truth
Evaluate against a known reference response.
Evaluations against a “ground truth” reference are crucial for determining how well an AI’s responses align with established correct answers. By comparing the generated responses to a known reference, we can measure the accuracy and relevance of the AI’s output, ensuring that it meets the expected standards.
This method of evaluation helps identify discrepancies, such as hallucinations or omissions, and provides insights into areas where the AI’s performance can be improved. Using Atla to perform these evaluations involves passing a reference response in the messages block, enabling a detailed assessment of the AI’s alignment with the ground truth.
Atla has been trained on specific reference metrics ensuring the best performance.
Running evals against a ‘ground truth’ answer
If you have access to a ‘ground truth’ answer, Atla can use this reference response to evaluate answers against it. To achieve this, pass a reference via the reference parameter.