Skip to main content

Envisioning is an emerging technology research institute and advisory.

LinkedInInstagramGitHub

2011 — 2026

research
  • Reports
  • Newsletter
  • Methodology
  • Origins
  • Vocab
services
  • Research Sessions
  • Signals Workspace
  • Bespoke Projects
  • Use Cases
  • Signal Scanfree
  • Readinessfree
impact
  • ANBIMAFuture of Brazilian Capital Markets
  • IEEECharting the Energy Transition
  • Horizon 2045Future of Human and Planetary Security
  • WKOTechnology Scanning for Austria
audiences
  • Innovation
  • Strategy
  • Consultants
  • Foresight
  • Associations
  • Governments
resources
  • Pricing
  • Partners
  • How We Work
  • Data Visualization
  • Multi-Model Method
  • FAQ
  • Security & Privacy
about
  • Manifesto
  • Community
  • Events
  • Support
  • Contact
  • Login
ResearchServicesPricingPartnersAbout
ResearchServicesPricingPartnersAbout
  1. Home
  2. Vocab
  3. Federated Training

Federated Training

Collaborative model training across distributed devices without centralizing raw data.

Year: 2016Generality: 694
Back to Vocab

Federated training is a distributed machine learning paradigm in which a shared model is trained across many decentralized devices or servers, each retaining its own local data. Rather than pooling raw data on a central server, each participant trains the model locally using its private dataset and transmits only model updates — typically gradients or weight deltas — back to a central coordinator. That coordinator aggregates the updates, often via a technique called Federated Averaging, to improve the global model, which is then redistributed to participants for the next round of training.

The approach was formalized by researchers at Google in 2016, motivated by the challenge of training predictive models on smartphones without compromising user privacy. Because sensitive data never leaves the originating device, federated training offers a compelling alternative to traditional centralized learning in domains where data collection is legally restricted or ethically fraught — healthcare records, financial transactions, and personal communications being prominent examples. It also reduces the bandwidth costs and latency associated with transmitting large raw datasets over a network.

Despite its advantages, federated training introduces significant technical challenges. Data across participants is typically non-IID (non-independent and identically distributed), meaning each device's local dataset reflects idiosyncratic usage patterns rather than the global distribution. This heterogeneity can slow convergence and degrade model quality. Communication efficiency is another concern: coordinating hundreds of thousands of devices requires careful scheduling, compression of updates, and tolerance for stragglers or dropouts. Researchers have developed techniques such as gradient compression, secure aggregation, and differential privacy to address these issues.

Federated training has moved well beyond its origins in mobile keyboard prediction and now underpins applications in medical imaging consortia, fraud detection networks, and autonomous vehicle fleets — any setting where data is naturally siloed and privacy is paramount. As regulatory frameworks like GDPR tighten restrictions on data centralization, federated approaches are increasingly viewed not merely as a technical convenience but as a foundational strategy for responsible AI development at scale.

Related

Related

Federated Learning
Federated Learning

A training approach that learns from decentralized data without ever centralizing it.

Generality: 711
Federated Analytics
Federated Analytics

Analyzing decentralized data in place, without centralizing or exposing raw data.

Generality: 507
FSDP (Fully Sharded Data Parallel)
FSDP (Fully Sharded Data Parallel)

Distributed training technique that shards model parameters and optimizer states across devices.

Generality: 485
Data Parallelism
Data Parallelism

Training technique that splits data across multiple processors running identical model copies simultaneously.

Generality: 794
Edge Model
Edge Model

An AI model that runs inference directly on local devices rather than the cloud.

Generality: 575
DAWN (Distributed Agents in a Worldwide Network)
DAWN (Distributed Agents in a Worldwide Network)

A decentralized architecture where autonomous AI agents collaborate across global infrastructure.

Generality: 292