Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Training Compute

Training Compute

The total computational resources consumed while training a machine learning model.

Year: 2018Generality: 650
Back to Vocab

Training compute refers to the aggregate computational work performed during the process of fitting a machine learning model to data — typically measured in floating-point operations (FLOPs). It encompasses every forward pass, backward pass, and parameter update executed across the full training run, and scales with model size, dataset size, and the number of training steps. As deep learning models have grown dramatically in scale, training compute has become one of the primary axes along which AI progress is measured and planned, often requiring clusters of specialized accelerators such as GPUs or TPUs running for weeks or months.

The significance of training compute is formalized in AI scaling laws, which describe empirical relationships between compute budget, model parameters, dataset tokens, and resulting model performance. Landmark work by researchers at OpenAI and DeepMind — including the Chinchilla scaling laws — demonstrated that compute is most efficiently spent when model size and data volume are scaled together in specific proportions. This insight transformed how practitioners allocate training budgets and has driven the field toward increasingly deliberate compute-optimal training strategies rather than simply maximizing model size.

Training compute matters beyond raw performance: it is a proxy for cost, energy consumption, and accessibility. The exponential growth in compute required by frontier models has concentrated cutting-edge AI development among a small number of well-resourced organizations, raising questions about research equity and environmental impact. Compute efficiency — achieving equivalent performance with fewer FLOPs — has consequently become a major research objective, motivating advances in architecture design, mixed-precision training, and data curation. Understanding training compute is now essential for anyone reasoning about the economics, capabilities, and limitations of modern AI systems.

Related

Related

Compute
Compute

The processing power and hardware resources required to train and run AI models.

Generality: 875
Training Cost
Training Cost

The total computational, energy, and financial resources required to train an AI model.

Generality: 620
Compute Efficiency
Compute Efficiency

How effectively a system converts computational resources into useful model performance.

Generality: 702
Evaluation-Time Compute
Evaluation-Time Compute

Computational resources consumed when an AI model runs inference on new data.

Generality: 627
Training
Training

The iterative process of optimizing a model's parameters using data.

Generality: 950
Machine Time
Machine Time

The computational time a system spends executing tasks, excluding human interaction.

Generality: 381