A framework where agents minimize prediction errors through both perception and action.
Active inference is a theoretical framework from computational neuroscience that describes how intelligent agents — biological or artificial — navigate the world by continuously minimizing the difference between their predictions and actual sensory experience. Rooted in Karl Friston's free energy principle, it proposes that the brain (or any adaptive system) maintains an internal generative model of the environment and works to keep prediction errors as small as possible. Crucially, this minimization happens through two complementary channels: updating internal beliefs to better match incoming data (perception), and taking actions that steer the world toward expected states (behavior).
This dual mechanism distinguishes active inference from purely passive learning approaches. Rather than simply reacting to stimuli or optimizing a fixed reward signal, an active inference agent selects actions that are predicted to reduce uncertainty about its environment. This gives the framework a natural account of curiosity and exploration: an agent is intrinsically motivated to gather information that resolves ambiguity in its world model. The underlying mathematics draws on variational Bayes and information theory, treating perception and action as two sides of the same inference problem.
In machine learning, active inference offers an alternative to reinforcement learning that sidesteps some of its known limitations, such as reward sparsity and the need for hand-crafted objectives. Instead of maximizing an external reward, an active inference agent minimizes expected free energy — a quantity that balances exploiting known rewarding states against exploring uncertain ones. This built-in exploration-exploitation trade-off emerges naturally from the probabilistic framework rather than requiring explicit engineering.
The appeal of active inference lies in its unifying scope: it provides a single principled account of perception, action, learning, and attention under one mathematical roof. Researchers have applied it to robotics, cognitive modeling, psychiatric disorder modeling, and adaptive control systems. While computationally demanding and still maturing as an engineering tool, active inference continues to attract interest as a biologically plausible and theoretically coherent foundation for building agents that learn by reducing uncertainty rather than chasing predefined rewards.