How well a model learns from limited training data to achieve strong performance.
Sample efficiency refers to a learning algorithm's ability to achieve strong performance using relatively few training examples. It is a measure of how effectively a model extracts useful signal from each data point it encounters, rather than relying on sheer data volume to drive improvement. In practice, a highly sample-efficient algorithm reaches a given level of accuracy or capability with far less data than a less efficient counterpart, which has significant implications for cost, speed, and feasibility of deployment.
The concept is especially critical in domains where data collection is expensive, dangerous, or slow. In robotics, for instance, a physical robot may require thousands of real-world trials to learn a manipulation task—each trial taking time and risking hardware damage. In medical diagnostics, labeled training examples may require expert annotation and are inherently scarce. Sample efficiency therefore becomes a practical bottleneck that determines whether machine learning is viable at all in these settings. Reinforcement learning has been a particularly active area of concern, since agents must interact with environments to generate their own training signal, making every interaction costly.
Several techniques have been developed to improve sample efficiency. Transfer learning allows a model pretrained on a data-rich source domain to be fine-tuned on a target domain with limited examples, leveraging shared structure across tasks. Meta-learning, or "learning to learn," trains models across many tasks so they can rapidly adapt to new ones from just a handful of examples. Data augmentation synthetically expands training sets by applying label-preserving transformations. Model-based reinforcement learning improves efficiency by building an internal model of the environment, enabling the agent to simulate experience rather than always requiring real interactions.
Sample efficiency gained particular prominence in the machine learning community around 2016, as deep reinforcement learning systems like those playing Atari games demonstrated impressive capabilities but required tens of millions of frames of experience—far beyond what humans need to master the same tasks. This gap between human and machine data requirements energized research into more efficient learning paradigms and remains an open challenge central to making AI systems more practical and broadly applicable.