Skip to content
Glossary

Precision and Recall

Metrics evaluating a model's predictions: Precision measures exactness, while Recall assesses completeness.
Definition

Precision and Recall are fundamental metrics in the evaluation of machine learning models, particularly in classification tasks and information retrieval. Precision, also known as positive predictive value, measures the proportion of true positive predictions in all positive predictions made by the model. It reflects the accuracy of the model in identifying relevant instances.

Recall, also known as sensitivity, measures the proportion of true positive predictions out of all actual positive instances in the dataset. It assesses the model's ability to capture all relevant instances.

  • Precision = True Positives / (True Positives + False Positives)
  • Recall = True Positives / (True Positives + False Negatives)

These metrics are especially important in contexts where the balance between capturing relevant instances (Recall) and ensuring the relevance of the instances captured (Precision) is crucial. In many real-world applications, there is a trade-off between Precision and Recall, and improving one may lead to a decrease in the other. The F1 score, the harmonic mean of Precision and Recall, is often used to balance this trade-off.

Examples/Use Cases:

In email spam detection, a high Recall would mean that most spam emails are correctly identified, but low Precision might mean that many legitimate emails are incorrectly marked as spam, which could be inconvenient for users. Conversely, high Precision would ensure that almost all emails marked as spam are indeed spam, but low Recall could mean that many spam emails are not detected, posing a security risk.

In medical diagnostics, high Recall is crucial for conditions where missing a positive case (a diseased individual) could have serious implications, even if it means some healthy individuals are falsely flagged (lower Precision) and require further testing. These examples illustrate how Precision and Recall help in evaluating and tuning models according to the specific requirements and potential costs of errors in different applications.

Related Terms
← Back to Glossary

Need human evaluators for your AI research? Scale annotation with expert AI Trainers.