Federated ADMM from Bayesian Duality
By: Thomas Möllenhoff , Siddharth Swaroop , Finale Doshi-Velez and more
Potential Business Impact:
Makes AI learn better from many computers.
ADMM is a popular method for federated deep learning which originated in the 1970s and, even though many new variants of it have been proposed since then, its core algorithmic structure has remained unchanged. Here, we take a major departure from the old structure and present a fundamentally new way to derive and extend federated ADMM. We propose to use a structure called Bayesian Duality which exploits a duality of the posterior distributions obtained by solving a variational-Bayesian reformulation of the original problem. We show that this naturally recovers the original ADMM when isotropic Gaussian posteriors are used, and yields non-trivial extensions for other posterior forms. For instance, full-covariance Gaussians lead to Newton-like variants of ADMM, while diagonal covariances result in a cheap Adam-like variant. This is especially useful to handle heterogeneity in federated deep learning, giving up to 7% accuracy improvements over recent baselines. Our work opens a new Bayesian path to improve primal-dual methods.
Similar Papers
FedAPM: Federated Learning via ADMM with Partial Model Personalization
Machine Learning (CS)
Helps AI learn better from different people's phones.
Federated Smoothing ADMM for Localization
Machine Learning (CS)
Helps computers find lost things better.
Distributionally Robust Federated Learning: An ADMM Algorithm
Machine Learning (CS)
Helps AI learn from different data sources.