Federated learning
Federated learning is a distributed machine learning approach that enables collaborative training of models across multiple decentralized clients, such as mobile devices or servers, each holding local data samples that remain on-device, with only model updates aggregated centrally to improve a shared global model without raw data exchange.[1] This paradigm addresses key challenges in traditional centralized training by minimizing data transfer and enhancing privacy through data locality, though it requires careful handling of statistical heterogeneity and communication efficiency.[1] Originally proposed in 2016 by researchers at Google, it was motivated by scenarios like next-word prediction on smartphones, where billions of user interactions generate vast but siloed data.[2] The core algorithm involves iterative rounds where clients perform local stochastic gradient descent on their data and upload gradient or model difference updates to a central server, which averages them—often weighted by client data size—to refine the global model before redistribution.[1] This process reduces bandwidth needs compared to full data transmission and supports non-IID data distributions common in real-world edge environments, though convergence can be slower due to client drift from local optimizations. Early implementations demonstrated substantial reductions in communication costs, such as up to 100x fewer bits transferred for deep network training versus centralized baselines.[1] Federated learning has been applied in production systems for tasks like predictive text in Google's Gboard keyboard and speech recognition, leveraging vast edge data while complying with privacy regulations like GDPR by avoiding data centralization.[2] However, it does not inherently provide formal privacy guarantees, as aggregated updates can still leak sensitive information via model inversion or membership inference attacks, prompting integrations with differential privacy techniques to bound such risks probabilistically.[3] Ongoing research focuses on robustness to heterogeneous devices, secure aggregation against malicious clients, and scalability to thousands of participants, positioning it as a foundational method for privacy-preserving AI in domains including healthcare and finance.