Collaborative model training across distributed devices without centralizing raw data.
Federated training is a distributed machine learning paradigm in which a shared model is trained across many decentralized devices or servers, each retaining its own local data. Rather than pooling raw data on a central server, each participant trains the model locally using its private dataset and transmits only model updates — typically gradients or weight deltas — back to a central coordinator. That coordinator aggregates the updates, often via a technique called Federated Averaging, to improve the global model, which is then redistributed to participants for the next round of training.
The approach was formalized by researchers at Google in 2016, motivated by the challenge of training predictive models on smartphones without compromising user privacy. Because sensitive data never leaves the originating device, federated training offers a compelling alternative to traditional centralized learning in domains where data collection is legally restricted or ethically fraught — healthcare records, financial transactions, and personal communications being prominent examples. It also reduces the bandwidth costs and latency associated with transmitting large raw datasets over a network.
Despite its advantages, federated training introduces significant technical challenges. Data across participants is typically non-IID (non-independent and identically distributed), meaning each device's local dataset reflects idiosyncratic usage patterns rather than the global distribution. This heterogeneity can slow convergence and degrade model quality. Communication efficiency is another concern: coordinating hundreds of thousands of devices requires careful scheduling, compression of updates, and tolerance for stragglers or dropouts. Researchers have developed techniques such as gradient compression, secure aggregation, and differential privacy to address these issues.
Federated training has moved well beyond its origins in mobile keyboard prediction and now underpins applications in medical imaging consortia, fraud detection networks, and autonomous vehicle fleets — any setting where data is naturally siloed and privacy is paramount. As regulatory frameworks like GDPR tighten restrictions on data centralization, federated approaches are increasingly viewed not merely as a technical convenience but as a foundational strategy for responsible AI development at scale.