Predictable power-law relationships between model size, data, compute, and performance.
Scaling laws are empirical mathematical relationships that describe how the performance of machine learning models improves as key resources increase — specifically model parameter count, training dataset size, and computational budget. Rather than improving arbitrarily or unpredictably, model performance tends to follow smooth power-law curves with respect to each of these variables, meaning that each order-of-magnitude increase in scale yields a roughly consistent, predictable gain in capability. This regularity allows researchers to forecast how well a model will perform before training it, and to make principled decisions about how to allocate a fixed compute budget between model size and data volume.
The mechanics underlying scaling laws reflect deep statistical properties of how neural networks learn from data. Larger models have greater capacity to represent complex functions, while more data reduces overfitting and exposes the model to a richer distribution of patterns. Crucially, these factors interact: a very large model trained on too little data will underperform, and vice versa. The Chinchilla scaling laws, published by Hoffmann et al. in 2022, refined earlier estimates by showing that many prominent large language models had been significantly undertrained relative to their size — that optimal performance requires scaling data and parameters in roughly equal proportion.
Scaling laws matter because they transform AI development from an art into something closer to an engineering discipline. Instead of relying on intuition or trial-and-error, teams can use scaling predictions to plan multi-million-dollar training runs with reasonable confidence in the outcome. They also carry strategic implications: if performance scales smoothly and predictably with compute, then sustained investment in hardware and data becomes a reliable path to capability improvements, which has shaped the economics and competitive dynamics of frontier AI development.
However, scaling laws have important limitations. They describe average performance on benchmark metrics and do not guarantee that specific capabilities — reasoning, factual accuracy, or safety — emerge reliably at a given scale. Some abilities appear to emerge abruptly rather than smoothly, complicating extrapolation. Nonetheless, scaling laws remain one of the most practically influential empirical findings in modern deep learning research.