Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. ASIC (Application-Specific Integrated Circuit)

ASIC (Application-Specific Integrated Circuit)

Custom silicon chips designed to accelerate specific computational workloads with maximum efficiency.

Year: 2015Generality: 700
Back to Vocab

An Application-Specific Integrated Circuit (ASIC) is a microchip engineered to perform a fixed, predetermined set of operations rather than serving as a general-purpose processor. Unlike CPUs or GPUs, which are designed for flexibility across diverse tasks, ASICs sacrifice programmability in exchange for dramatic gains in speed, power efficiency, and physical footprint for their target workload. In machine learning, this tradeoff is highly attractive: the core operations of neural networks—matrix multiplications, convolutions, activation functions—are well-defined and repetitive, making them ideal candidates for hardwired silicon logic.

ASICs for ML work by implementing the mathematical primitives of neural network computation directly in hardware. Rather than fetching and decoding general instructions at runtime, an ML-focused ASIC routes data through fixed arithmetic units optimized for tensor operations, often processing thousands of multiply-accumulate operations in parallel. Memory hierarchies are co-designed with the compute fabric to minimize data movement, which is frequently the dominant energy cost in large-scale inference and training. The result is throughput-per-watt figures that general-purpose hardware cannot match for the specific workloads the chip targets.

Google's Tensor Processing Unit (TPU), first deployed internally in 2015 and publicly disclosed in 2016, brought ML ASICs into mainstream awareness. The TPU demonstrated that purpose-built silicon could deliver order-of-magnitude improvements in inference throughput and energy efficiency over contemporary GPUs for certain workloads, validating the ASIC approach for hyperscale AI infrastructure. Since then, a wave of companies—including Cerebras, Graphcore, Groq, and numerous cloud providers—have developed proprietary ML ASICs targeting training, inference, or both.

ASICs matter to the AI field because hardware efficiency increasingly determines what models are economically feasible to train and deploy. As model sizes have grown into the hundreds of billions of parameters, the cost and energy demands of computation have become central concerns. ASICs allow organizations to extract far more useful computation per dollar and per watt than commodity hardware, enabling larger experiments, lower-latency products, and more sustainable deployment at scale. Their rigidity is a limitation—an ASIC optimized for transformer inference may be poorly suited to a future architecture—but for stable, high-volume workloads, they represent the frontier of practical AI acceleration.

Related

Related

Accelerator Chip
Accelerator Chip

Specialized hardware that dramatically speeds up AI training and inference workloads.

Generality: 781
Accelerated Computing
Accelerated Computing

Using specialized hardware to dramatically speed up AI and machine learning workloads.

Generality: 794
Accelerator
Accelerator

Specialized hardware that speeds up AI training and inference beyond CPU capabilities.

Generality: 792
FPGA (Field-Programmable Gate Array)
FPGA (Field-Programmable Gate Array)

Reconfigurable hardware chips that accelerate AI workloads with low latency and power.

Generality: 627
TPU (Tensor Processing Unit)
TPU (Tensor Processing Unit)

Google's custom chip designed to accelerate machine learning workloads at scale.

Generality: 550
GPU (Graphics Processing Unit)
GPU (Graphics Processing Unit)

Massively parallel processor that accelerates deep learning by handling thousands of simultaneous computations.

Generality: 871