Definition
The recall test measures the ability to find all positive instances, calculated as TP / (TP + FN). For binary classification, it considers class 1 as “positive.” For multiclass classification, it uses the macro-average of the recall score for each class, treating all classes equally.Taxonomy
- Task types: Tabular classification, text classification.
- Availability: and .
Why it matters
- Recall measures how many of the actual positive cases the model correctly identifies, making it crucial when missing positive cases is costly.
- It’s particularly important in applications like medical diagnosis, fraud detection, or safety systems where failing to detect positive cases can have serious consequences.
- Higher recall values indicate better model performance, with 1.0 representing no false negatives.
- Recall complements precision to provide a complete picture of model performance, especially in the context of the precision-recall trade-off.
Required columns
To compute this metric, your dataset must contain the following columns:- Predictions: The predicted class labels from your classification model
- Ground truths: The actual/true class labels
Test configuration examples
If you are writing atests.json
, here are a few valid configurations for the recall test:
Related
- Precision test - Measure positive prediction accuracy.
- F1 test - Harmonic mean of precision and recall.
- Geometric mean test - Alternative balanced metric.
- Accuracy test - Overall classification correctness.
- Aggregate metrics - Overview of all available metrics.