Optimal LLM training balances model size and data quantity for a fixed compute budget.
Chinchilla scaling refers to a set of empirically derived principles for training large language models (LLMs) efficiently by finding the optimal ratio between model size (number of parameters) and the volume of training data, given a fixed computational budget. The concept emerged from DeepMind's 2022 paper "Training Compute-Optimal Large Language Models," which introduced a 70-billion-parameter model called Chinchilla as a proof of concept. The key finding was that prior large models—including GPT-3 and Gopher—were significantly undertrained relative to their size, meaning they used far more parameters than could be justified by the amount of data they were trained on.
The core insight is that model size and training tokens should scale roughly in proportion to each other. Specifically, the research suggested that for every doubling of model parameters, the training dataset should also approximately double. This stands in contrast to earlier scaling intuitions, which prioritized growing model size while holding data relatively fixed. The Chinchilla model, despite having roughly four times fewer parameters than the 280-billion-parameter Gopher model, outperformed it across a wide range of benchmarks by being trained on approximately four times more data—demonstrating that data efficiency is as critical as raw model scale.
The practical implications of Chinchilla scaling are significant for both research and industry. Training a smaller, data-rich model can achieve superior performance while consuming less memory and compute during inference—a major cost consideration when deploying models at scale. This reframing shifted how many organizations approached LLM development, encouraging investment in high-quality, large-scale datasets rather than simply racing to build ever-larger parameter counts.
Chinchilla scaling has also sparked ongoing debate and refinement. Subsequent work has questioned whether the original compute-optimal ratios fully account for inference costs, suggesting that in practice it may be worthwhile to train smaller models on even more data than Chinchilla prescribes—a perspective reflected in models like Meta's LLaMA series. As a result, Chinchilla scaling is now understood less as a fixed law and more as a foundational framework for reasoning about the trade-offs inherent in large-scale model training.