Inferring hidden causes or parameters from observed data by reversing forward models.
An inverse problem involves working backward from observed measurements to determine the underlying causes, parameters, or structures that produced them. This stands in contrast to a forward problem, where known inputs are used to predict outputs. In machine learning, inverse problems appear across a wide range of applications: reconstructing medical images from sensor readings, inferring physical properties of materials from spectroscopic data, recovering audio signals from corrupted recordings, or estimating climate model parameters from historical observations. The challenge is that the mapping from causes to effects is often many-to-one, making the reverse direction ambiguous.
Inverse problems are frequently described as ill-posed in the sense defined by Jacques Hadamard — they may lack a unique solution, or small perturbations in the observed data can lead to wildly different inferred causes. This instability makes naive inversion unreliable. Regularization is the primary tool for managing ill-posedness: by imposing constraints or penalties on the solution space, methods like Tikhonov regularization or sparsity-promoting L1 penalties bias the solution toward plausible answers. Bayesian inference provides a natural probabilistic framework for the same idea, encoding prior beliefs about likely causes and updating them given observed evidence to produce a posterior distribution over possible solutions.
Modern machine learning has dramatically expanded the toolkit for solving inverse problems. Deep neural networks can learn powerful priors directly from data, enabling learned regularization strategies that outperform hand-crafted alternatives in domains like image reconstruction and geophysical inversion. Diffusion models and normalizing flows have emerged as particularly effective generative approaches, capable of sampling from the posterior distribution over solutions rather than returning a single point estimate. Physics-informed neural networks further integrate known governing equations into the learning process, improving sample efficiency and physical consistency.
Inverse problems sit at the intersection of applied mathematics, statistics, and machine learning, and their importance spans science and engineering. Solving them well requires balancing fidelity to observed data against prior constraints — a tension that mirrors the broader machine learning challenge of fitting models without overfitting. As datasets grow larger and neural architectures more expressive, learned approaches to inverse problems are increasingly replacing classical analytical methods.