Score: 2

Multi-Modal Style Transfer-based Prompt Tuning for Efficient Federated Domain Generalization

Published: January 9, 2026 | arXiv ID: 2601.05955v1

By: Yuliang Chen , Xi Lin , Jun Wu and more

Potential Business Impact:

Helps AI learn from many different data sources.

Business Areas:
Visual Search Internet Services

Federated Domain Generalization (FDG) aims to collaboratively train a global model across distributed clients that can generalize well on unseen domains. However, existing FDG methods typically struggle with cross-client data heterogeneity and incur significant communication and computation overhead. To address these challenges, this paper presents a new FDG framework, dubbed FaST-PT, which facilitates local feature augmentation and efficient unseen domain adaptation in a distributed manner. First, we propose a lightweight Multi-Modal Style Transfer (MST) method to transform image embedding under text supervision, which could expand the training data distribution and mitigate domain shift. We then design a dual-prompt module that decomposes the prompt into global and domain prompts. Specifically, global prompts capture general knowledge from augmented embedding across clients, while domain prompts capture domain-specific knowledge from local data. Besides, Domain-aware Prompt Generation (DPG) is introduced to adaptively generate suitable prompts for each sample, which facilitates unseen domain adaptation through knowledge fusion. Extensive experiments on four cross-domain benchmark datasets, e.g., PACS and DomainNet, demonstrate the superior performance of FaST-PT over SOTA FDG methods such as FedDG-GA and DiPrompt. Ablation studies further validate the effectiveness and efficiency of FaST-PT.

Country of Origin
🇮🇹 🇺🇸 🇨🇳 Italy, United States, China

Page Count
9 pages

Category
Computer Science:
Distributed, Parallel, and Cluster Computing