Federated Flow Matching
By: Zifan Wang , Anqi Dong , Mahmoud Selim and more
Potential Business Impact:
Trains AI privately on scattered data.
Data today is decentralized, generated and stored across devices and institutions where privacy, ownership, and regulation prevent centralization. This motivates the need to train generative models directly from distributed data locally without central aggregation. In this paper, we introduce Federated Flow Matching (FFM), a framework for training flow matching models under privacy constraints. Specifically, we first examine FFM-vanilla, where each client trains locally with independent source and target couplings, preserving privacy but yielding curved flows that slow inference. We then develop FFM-LOT, which employs local optimal transport couplings to improve straightness within each client but lacks global consistency under heterogeneous data. Finally, we propose FFM-GOT, a federated strategy based on the semi-dual formulation of optimal transport, where a shared global potential function coordinates couplings across clients. Experiments on synthetic and image datasets show that FFM enables privacy-preserving training while enhancing both the flow straightness and sample quality in federated settings, with performance comparable to the centralized baseline.
Similar Papers
Flow Matching for Tabular Data Synthesis
Machine Learning (CS)
Creates private fake data that's better than old ways.
Flow Matching for Medical Image Synthesis: Bridging the Gap Between Speed and Quality
CV and Pattern Recognition
Makes AI create better medical pictures faster.
Federated Gaussian Mixture Models
Machine Learning (CS)
Lets phones learn together without sharing secrets.