Score: 2

Feature-Aware One-Shot Federated Learning via Hierarchical Token Sequences

Published: January 7, 2026 | arXiv ID: 2601.03882v1

By: Shudong Liu , Hanwen Zhang , Xiuling Wang and more

Potential Business Impact:

Helps AI learn from medical images better.

Business Areas:
Image Recognition Data and Analytics, Software

One-shot federated learning (OSFL) reduces the communication cost and privacy risks of iterative federated learning by constructing a global model with a single round of communication. However, most existing methods struggle to achieve robust performance on real-world domains such as medical imaging, or are inefficient when handling non-IID (Independent and Identically Distributed) data. To address these limitations, we introduce FALCON, a framework that enhances the effectiveness of OSFL over non-IID image data. The core idea of FALCON is to leverage the feature-aware hierarchical token sequences generation and knowledge distillation into OSFL. First, each client leverages a pretrained visual encoder with hierarchical scale encoding to compress images into hierarchical token sequences, which capture multi-scale semantics. Second, a multi-scale autoregressive transformer generator is used to model the distribution of these token sequences and generate the synthetic sequences. Third, clients upload the synthetic sequences along with the local classifier trained on the real token sequences to the server. Finally, the server incorporates knowledge distillation into global training to reduce reliance on precise distribution modeling. Experiments on medical and natural image datasets validate the effectiveness of FALCON in diverse non-IID scenarios, outperforming the best OSFL baselines by 9.58% in average accuracy.

Country of Origin
🇭🇰 🇨🇳 China, Hong Kong

Repos / Data Links

Page Count
9 pages

Category
Computer Science:
Machine Learning (CS)