The assumption that an AI agent acts to maximize expected utility given available information.
The Principle of Rationality holds that an intelligent agent is rational when it selects actions that maximize its expected utility given its current beliefs about the world, the information available to it, and its computational constraints. This principle serves as a foundational design philosophy for AI systems, providing a normative standard against which agent behavior can be evaluated. Rather than prescribing a specific algorithm, it defines what optimal behavior looks like: an agent should never knowingly choose an action that yields lower expected value than an available alternative. This framing connects AI directly to classical decision theory and Bayesian reasoning, where beliefs are represented as probability distributions and actions are chosen to maximize expected outcomes under uncertainty.
In practice, the principle underpins a wide range of AI architectures. In reinforcement learning, agents are explicitly trained to maximize cumulative reward signals, embodying rationality through learned value functions or policies. In planning and search, rational agents evaluate action sequences against objective functions to select optimal plans. In game-theoretic settings, rational agents reason about the strategies of other agents to identify equilibrium behaviors. The principle also motivates the use of probabilistic graphical models and Bayesian networks, where rational inference means updating beliefs in accordance with Bayes' theorem as new evidence arrives.
The principle carries significant practical and theoretical weight because it provides a unified criterion for comparing AI designs: a system is better insofar as it more closely approximates rational behavior under its operating constraints. However, the principle also has well-known limitations. Bounded rationality, a concept developed to address real-world constraints, acknowledges that agents often cannot compute the truly optimal action due to limited time, memory, or knowledge. Modern AI research increasingly grapples with how to build systems that are approximately rational in tractable ways, and how to align the utility functions agents optimize with genuinely desirable human values — a challenge central to AI safety research.