A learning approach that builds a complete global model before any predictions are made.
Eager learning is a machine learning paradigm in which a model is fully constructed from the training data before any prediction requests arrive. Rather than deferring computation until query time, eager learners process the entire training set upfront to produce an explicit, generalized hypothesis about the data. This stands in direct contrast to lazy learning methods—such as k-nearest neighbors—which store training examples and only compute a response when a specific query is presented. The distinction matters because it shapes how computational cost is distributed between training and inference.
The mechanics of eager learning involve algorithms that perform a global optimization or search over the training data to extract patterns, boundaries, or representations. Decision trees, neural networks, support vector machines, and logistic regression are canonical examples. Each of these methods ingests the full dataset, adjusts internal parameters or structure, and produces a compact model that can be applied rapidly to new inputs. Because the heavy lifting happens during training, prediction at runtime is typically fast and requires minimal additional computation.
The primary advantage of eager learning is inference efficiency. Once trained, a model can respond to queries almost instantaneously, making it well-suited for production systems where low-latency predictions are critical. However, this comes at a cost: eager models commit to a fixed hypothesis at training time, which means they may generalize poorly if the test distribution shifts significantly from the training distribution. Retraining is required to incorporate new data, unlike lazy methods that naturally adapt as the example set grows.
Eager learning became a central organizing concept in machine learning during the early 1990s, as researchers sought to formally characterize the tradeoffs between different learning strategies. The framework helped clarify why certain algorithms scale well to large datasets while others struggle, and it remains a useful conceptual lens for understanding the computational and statistical properties of modern methods, including deep learning architectures that epitomize the eager approach at massive scale.