Definition
The answer correctness test compares and evaluates the factual accuracy of the generated response with respect to the reference ground truth. This metric is based on the Ragas factual correctness metric.Taxonomy
- Task types: LLM.
- Availability: and .
Why it matters
- Answer correctness ensures that your LLM generates factually accurate responses when compared to known ground truth answers.
- This metric is crucial for applications where factual accuracy is paramount, such as question-answering systems, educational tools, or information retrieval systems.
- It helps identify when your model is generating plausible-sounding but incorrect information.
Required columns
To compute this metric, your dataset must contain the following columns:- Outputs: The generated answer/response from your LLM
- Ground truths: The reference/correct answer to compare against
Test configuration examples
If you are writing atests.json
, here are a few valid configurations for the answer correctness test:
Related
- Ragas integration - Learn more about Ragas metrics.
- Answer relevancy test - Measure how relevant answers are to questions.
- Aggregate metrics - Overview of all available metrics.