A beginner-friendly explanation of precision, recall and F1 score, including when each metric matters most.
What precision measures
Precision answers this question: when the model predicts positive, how often is it correct? A high precision score means the model is careful about making positive predictions.
This matters when false positives are expensive. For example, if you flag legitimate transactions as fraud too often, the business cost can be significant.
What recall measures
Recall answers a different question: among all actual positive cases, how many did the model find? A high recall score means the model catches more true positives.
Recall matters when missing a positive case is expensive. In medical screening or safety systems, false negatives can be more harmful than false positives.
Advertisement
AdSense slot placeholder
Why F1 score is useful
F1 score combines precision and recall into one number. It is the harmonic mean, which means it only stays high when both precision and recall are reasonably strong.
This makes F1 score useful when you want one balanced metric for model comparison, especially on imbalanced datasets.
- Use precision when false positives hurt most.
- Use recall when false negatives hurt most.
- Use F1 score when you need a balanced view of both.
Do not optimize metrics in isolation
No metric should be interpreted alone. A model can improve recall by predicting positive more often, but that may reduce precision. Likewise, improving precision may lower recall if the model becomes too conservative.
The right decision depends on business goals, user experience and the cost of each mistake. That is why confusion matrices and threshold analysis are still important.