Federated Learning on Riemannian Manifolds: A Gradient-Free Projection-Based Approach
By: Hongye Wang , Zhaoye Pan , Chang He and more
Potential Business Impact:
Trains AI models privately without seeing your data.
Federated learning (FL) has emerged as a powerful paradigm for collaborative model training across distributed clients while preserving data privacy. However, existing FL algorithms predominantly focus on unconstrained optimization problems with exact gradient information, limiting its applicability in scenarios where only noisy function evaluations are accessible or where model parameters are constrained. To address these challenges, we propose a novel zeroth-order projection-based algorithm on Riemannian manifolds for FL. By leveraging the projection operator, we introduce a computationally efficient zeroth-order Riemannian gradient estimator. Unlike existing estimators, ours requires only a simple Euclidean random perturbation, eliminating the need to sample random vectors in the tangent space, thus reducing computational cost. Theoretically, we first prove the approximation properties of the estimator and then establish the sublinear convergence of the proposed algorithm, matching the rate of its first-order counterpart. Numerically, we first assess the efficiency of our estimator using kernel principal component analysis. Furthermore, we apply the proposed algorithm to two real-world scenarios: zeroth-order attacks on deep neural networks and low-rank neural network training to validate the theoretical findings.
Similar Papers
Data-Free Black-Box Federated Learning via Zeroth-Order Gradient Estimation
Machine Learning (CS)
Lets computers learn together without sharing secrets.
Communication-Efficient Zero-Order and First-Order Federated Learning Methods over Wireless Networks
Machine Learning (CS)
Makes phones learn together without sharing secrets.
Gradient Projection onto Historical Descent Directions for Communication-Efficient Federated Learning
Machine Learning (CS)
Makes AI learn faster with less data sent.