Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Accelerator

Accelerator

Specialized hardware that speeds up AI training and inference beyond CPU capabilities.

Year: 2012Generality: 792
Back to Vocab

An accelerator is a specialized hardware component designed to perform specific computational tasks far more efficiently than a general-purpose CPU. In machine learning, accelerators exploit the massively parallel nature of neural network operations — particularly matrix multiplications and tensor contractions — by executing thousands of operations simultaneously. The three dominant accelerator types in AI are graphics processing units (GPUs), tensor processing units (TPUs), and field-programmable gate arrays (FPGAs). GPUs, originally built for rendering graphics, proved remarkably well-suited to deep learning because their thousands of smaller cores can process large batches of data in parallel. TPUs are custom silicon designed specifically for tensor operations, offering higher throughput and energy efficiency for neural network workloads. FPGAs provide reconfigurable logic that can be tuned to specific inference pipelines, making them attractive for low-latency edge deployments.

The practical importance of accelerators became undeniable around 2012, when researchers demonstrated that training deep convolutional networks on GPUs reduced training time from weeks to days. This shift unlocked the modern deep learning era, making it feasible to train models with hundreds of millions of parameters on large datasets. Since then, hardware acceleration has become a central concern in AI research and industry, with chip design and model architecture co-evolving to maximize throughput and minimize energy consumption.

Beyond GPUs and TPUs, a new generation of AI-specific chips has emerged from companies including NVIDIA, Google, Intel, and numerous startups. These designs increasingly incorporate dedicated matrix engines, high-bandwidth memory, and on-chip interconnects optimized for the communication patterns of large-scale distributed training. The rise of large language models and foundation models has intensified demand, pushing accelerator clusters into the thousands of chips connected by high-speed fabrics.

Accelerators matter not just for speed but for economic and scientific feasibility. Without them, training state-of-the-art models would be prohibitively expensive in both time and energy. They also shape what kinds of models researchers explore, since architectural choices are often constrained by what runs efficiently on available hardware. Understanding accelerators is therefore essential for anyone working on the design, training, or deployment of modern AI systems.

Related

Related

Accelerator Chip
Accelerator Chip

Specialized hardware that dramatically speeds up AI training and inference workloads.

Generality: 781
Accelerated Computing
Accelerated Computing

Using specialized hardware to dramatically speed up AI and machine learning workloads.

Generality: 794
Inference Acceleration
Inference Acceleration

Techniques and hardware that speed up neural network prediction without sacrificing accuracy.

Generality: 694
GPU (Graphics Processing Unit)
GPU (Graphics Processing Unit)

Massively parallel processor that accelerates deep learning by handling thousands of simultaneous computations.

Generality: 871
ASIC (Application-Specific Integrated Circuit)
ASIC (Application-Specific Integrated Circuit)

Custom silicon chips designed to accelerate specific computational workloads with maximum efficiency.

Generality: 700
TPU (Tensor Processing Unit)
TPU (Tensor Processing Unit)

Google's custom chip designed to accelerate machine learning workloads at scale.

Generality: 550