Score: 1

Balanced Learning for Domain Adaptive Semantic Segmentation

Published: December 7, 2025 | arXiv ID: 2512.06886v1

By: Wangkai Li , Rui Sun , Bohao Liao and more

Potential Business Impact:

Helps computers better understand pictures of things.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Unsupervised domain adaptation (UDA) for semantic segmentation aims to transfer knowledge from a labeled source domain to an unlabeled target domain. Despite the effectiveness of self-training techniques in UDA, they struggle to learn each class in a balanced manner due to inherent class imbalance and distribution shift in both data and label space between domains. To address this issue, we propose Balanced Learning for Domain Adaptation (BLDA), a novel approach to directly assess and alleviate class bias without requiring prior knowledge about the distribution shift. First, we identify over-predicted and under-predicted classes by analyzing the distribution of predicted logits. Subsequently, we introduce a post-hoc approach to align the logits distributions across different classes using shared anchor distributions. To further consider the network's need to generate unbiased pseudo-labels during self-training, we estimate logits distributions online and incorporate logits correction terms into the loss function. Moreover, we leverage the resulting cumulative density as domain-shared structural knowledge to connect the source and target domains. Extensive experiments on two standard UDA semantic segmentation benchmarks demonstrate that BLDA consistently improves performance, especially for under-predicted classes, when integrated into various existing methods. Code is available at https://github.com/Woof6/BLDA.

Country of Origin
🇨🇳 China

Repos / Data Links

Page Count
27 pages

Category
Computer Science:
CV and Pattern Recognition