Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Precision-Recall Curve

Precision-Recall Curve

A plot evaluating classifier performance by trading off precision against recall across thresholds.

Year: 2006Generality: 729
Back to Vocab

A precision-recall curve is a diagnostic tool for evaluating binary classification models by plotting precision on the y-axis against recall on the x-axis across all possible decision thresholds. Precision measures the fraction of positive predictions that are actually correct, while recall (also called sensitivity) measures the fraction of true positives that the model successfully identifies. As the decision threshold is lowered, a model typically captures more true positives (higher recall) at the cost of also accepting more false positives (lower precision), and the curve traces this trade-off across the full threshold range.

The curve is especially valuable when class distributions are heavily skewed. In such settings, the ROC curve can paint an overly optimistic picture because it accounts for true negatives, which are abundant in imbalanced datasets and inflate apparent performance. The precision-recall curve sidesteps this by focusing exclusively on the positive class, making it the preferred evaluation tool in domains like fraud detection, rare disease diagnosis, and information retrieval, where the minority class is the primary concern. A model with a curve that hugs the top-right corner of the plot — maintaining high precision even at high recall — is considered strong.

A single scalar summary of the curve is often computed as the area under the precision-recall curve (AUPRC), sometimes called average precision. This metric aggregates performance across all thresholds into one number, enabling straightforward comparison between models. Unlike accuracy, AUPRC is robust to class imbalance and rewards models that rank true positives highly. Practitioners also use the curve interactively to select an operating threshold that satisfies application-specific constraints — for instance, a medical screening tool might prioritize recall to minimize missed diagnoses, accepting lower precision as a consequence.

The concept originates in information retrieval research from the 1970s and 1980s, where precision and recall were standard metrics for evaluating document search systems. It migrated into machine learning evaluation practice during the mid-2000s as large, imbalanced datasets became common in spam filtering, bioinformatics, and computer vision, cementing its role as a standard benchmark tool in modern ML workflows.

Related

Related

AUCPR (Area Under the Precision-Recall Curve)
AUCPR (Area Under the Precision-Recall Curve)

A classification metric summarizing precision-recall trade-offs across decision thresholds.

Generality: 595
Average Precision
Average Precision

A single-score metric summarizing model performance across all precision-recall thresholds.

Generality: 700
Classification Threshold
Classification Threshold

A cutoff value that maps a model's probability output to a discrete class label.

Generality: 694
Confusion Matrix
Confusion Matrix

A table that breaks down a classifier's predictions against actual class labels.

Generality: 796
Accuracy
Accuracy

The fraction of correct predictions a classification model makes overall.

Generality: 875
Bias-Variance Curve
Bias-Variance Curve

A plot showing how model complexity affects the balance between bias and variance.

Generality: 694