Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. GPU (Graphics Processing Unit)

GPU (Graphics Processing Unit)

Massively parallel processor that accelerates deep learning by handling thousands of simultaneous computations.

Year: 2007Generality: 871
Back to Vocab

A Graphics Processing Unit (GPU) is a specialized processor originally designed to accelerate the rendering of images and video for display output. Unlike a CPU, which contains a small number of powerful cores optimized for sequential tasks, a GPU contains thousands of smaller, more efficient cores designed to handle many operations simultaneously. This massively parallel architecture makes GPUs exceptionally well-suited for the kinds of mathematical operations that dominate machine learning workloads — particularly matrix multiplications and convolutions — where the same computation must be applied to enormous arrays of numbers at once.

In deep learning, training a neural network requires iterating over millions or billions of numerical operations across large datasets, often repeatedly for many epochs. GPUs dramatically accelerate this process by distributing these operations across their thousands of cores, reducing training times from weeks on a CPU to hours or even minutes. The same parallel advantage applies during inference, where GPUs can process large batches of inputs simultaneously, enabling real-time applications in computer vision, natural language processing, and generative AI.

The pivotal moment for GPU adoption in AI came with Nvidia's release of CUDA (Compute Unified Device Architecture) in 2007, which gave researchers a practical programming interface for running general-purpose computations on GPU hardware. The landmark 2012 ImageNet competition — where AlexNet, trained on GPUs, dramatically outperformed CPU-based competitors — demonstrated the transformative potential of GPU-accelerated deep learning and triggered an industry-wide shift. Since then, GPU clusters have become the standard infrastructure for AI research and production systems alike.

Today, GPUs remain central to the AI hardware ecosystem, with Nvidia's data center GPUs such as the A100 and H100 serving as the workhorses of large-scale model training. Competitors including AMD and Intel have developed their own GPU and accelerator offerings, while cloud providers offer GPU instances on demand. The insatiable compute demands of modern foundation models and generative AI systems have made GPU availability a strategic resource, shaping research timelines, business models, and even geopolitical policy around semiconductor supply chains.

Related

Related

Accelerated Computing
Accelerated Computing

Using specialized hardware to dramatically speed up AI and machine learning workloads.

Generality: 794
Accelerator
Accelerator

Specialized hardware that speeds up AI training and inference beyond CPU capabilities.

Generality: 792
Accelerator Chip
Accelerator Chip

Specialized hardware that dramatically speeds up AI training and inference workloads.

Generality: 781
CUDA (Compute Unified Device Architecture)
CUDA (Compute Unified Device Architecture)

NVIDIA's parallel computing platform enabling GPUs to accelerate general-purpose and AI workloads.

Generality: 794
TPU (Tensor Processing Unit)
TPU (Tensor Processing Unit)

Google's custom chip designed to accelerate machine learning workloads at scale.

Generality: 550
GPU-Poor
GPU-Poor

Having insufficient GPU resources to train or run competitive AI models.

Generality: 94