Built-in assumptions that help a learning algorithm generalize beyond its training data.
Inductive bias refers to the set of assumptions a machine learning algorithm uses to generalize from observed training examples to unseen data. Because no finite dataset can fully specify the correct function a model should learn, every learning algorithm must make some prior assumptions about which hypotheses are more plausible than others. These assumptions — whether explicit or implicit — define the inductive bias and determine which patterns the model is predisposed to discover. Without some form of inductive bias, a learner would have no principled basis for preferring one generalization over another, making learning from limited data theoretically impossible.
Inductive bias manifests differently across algorithm families. Linear models assume that relationships between inputs and outputs are approximately linear, which works well in many settings but fails when the true function is highly nonlinear. Decision trees favor shorter, simpler rules consistent with Occam's razor. Convolutional neural networks embed a spatial locality bias — the assumption that nearby pixels are more related than distant ones — making them well-suited for image data. Recurrent networks assume sequential dependencies in time. In each case, the architectural or algorithmic choices encode prior beliefs about the structure of the problem, shaping what the model can and cannot learn efficiently.
The concept is tightly linked to the bias-variance tradeoff. A model with strong inductive bias may underfit if its assumptions are wrong, but it will generalize well with less data when those assumptions are correct. A model with weak inductive bias is more flexible but requires far more data to avoid overfitting. Choosing an appropriate inductive bias for a given problem is therefore one of the most consequential decisions in model design — it determines sample efficiency, generalization behavior, and the kinds of errors a model is likely to make.
Inductive bias has grown in importance as researchers have moved toward understanding why certain architectures succeed. The success of transformers in natural language processing, for instance, has prompted analysis of what inductive biases attention mechanisms encode compared to recurrent networks. Similarly, debates around foundation models and transfer learning often center on whether pretraining instills useful inductive biases for downstream tasks. Understanding and deliberately engineering inductive bias remains a central challenge in building reliable, data-efficient machine learning systems.