A statistical method that updates probability estimates as new evidence arrives.
Bayesian inference is a statistical framework for reasoning under uncertainty by combining prior knowledge with observed data to produce updated probability estimates. At its core, the approach applies Bayes' theorem: the posterior probability of a hypothesis is proportional to the product of the prior probability — encoding what was believed before seeing data — and the likelihood of the observed evidence given that hypothesis. This stands in contrast to frequentist inference, which treats parameters as fixed unknowns and draws conclusions solely from the data at hand. Bayesian inference is especially valuable when data is scarce, because prior distributions allow practitioners to incorporate domain expertise or historical information directly into the model.
In machine learning, Bayesian inference underpins a wide range of techniques, from Bayesian neural networks and Gaussian processes to probabilistic graphical models and hyperparameter optimization. Rather than producing a single point estimate for a model parameter, Bayesian methods yield a full posterior distribution, which naturally quantifies uncertainty. This uncertainty quantification is critical in high-stakes applications such as medical diagnosis, autonomous systems, and scientific discovery, where knowing how confident a model is matters as much as the prediction itself.
A longstanding practical barrier to Bayesian inference was computational: computing the posterior exactly requires integrating over all possible parameter values, which is analytically intractable for most real-world models. The development of Markov Chain Monte Carlo (MCMC) sampling methods and, later, variational inference techniques transformed Bayesian inference from a theoretical ideal into a practical tool. Variational inference in particular, which approximates the posterior with a simpler distribution by minimizing a divergence measure, scales to large datasets and deep models in ways that MCMC cannot easily match.
Bayesian inference has become increasingly central to modern machine learning as the field moves toward models that reason about their own uncertainty. Frameworks such as PyMC, Stan, and Pyro have made Bayesian modeling accessible to practitioners, while research into scalable approximate inference continues to expand the frontier of what Bayesian methods can handle. Its combination of principled uncertainty quantification, flexibility in incorporating prior knowledge, and strong theoretical foundations makes Bayesian inference one of the most enduring and versatile tools in the machine learning toolkit.