Using learned patterns from data to estimate unknown or future outcomes.
Prediction is the process by which a trained machine learning model produces an output — a label, value, probability, or category — for inputs it has not explicitly seen before. Rather than retrieving a stored answer, the model generalizes from patterns learned during training to infer what the correct output should be. This applies equally to forecasting future events (will this customer churn next month?), classifying present observations (is this image a cat or a dog?), and estimating unknown quantities (what is the likely sale price of this house?). The unifying thread is inference under uncertainty: the model commits to an estimate even when the ground truth is not yet known or directly observable.
The mechanics of prediction vary by model architecture. In supervised learning, a model learns a mapping from input features to target outputs by minimizing prediction error on labeled training examples. At inference time, new inputs are passed through this learned function to generate predictions. Regression models output continuous values; classifiers output discrete categories or class probabilities; sequence models generate tokens or time-series forecasts. More complex architectures like neural networks compose many such transformations, enabling predictions over high-dimensional inputs like images, text, and audio.
Prediction quality depends on several interacting factors: the representativeness and size of training data, the appropriateness of the chosen model family, regularization to prevent overfitting, and careful evaluation on held-out data. Metrics like accuracy, mean squared error, AUC-ROC, and calibration scores quantify different aspects of predictive performance. Poorly calibrated models may be accurate on average yet systematically wrong for specific subgroups — a critical concern in high-stakes domains like medicine, credit scoring, and criminal justice.
Prediction is arguably the central task of applied machine learning. Nearly every deployed ML system — recommendation engines, fraud detectors, medical diagnostic tools, autonomous vehicles — ultimately reduces to making predictions that inform or automate decisions. As models grow more capable, the distinction between prediction and reasoning has blurred, but the core objective remains: produce the most accurate, reliable estimate possible given available information.