Score: 1

Domain Generalization: A Tale of Two ERMs

Published: October 6, 2025 | arXiv ID: 2510.04441v1

By: Yilun Zhu , Naihao Deng , Naichen Shi and more

Potential Business Impact:

Helps computers learn from different examples better.

Business Areas:
A/B Testing Data and Analytics

Domain generalization (DG) is the problem of generalizing from several distributions (or domains), for which labeled training data are available, to a new test domain for which no labeled data is available. A common finding in the DG literature is that it is difficult to outperform empirical risk minimization (ERM) on the pooled training data. In this work, we argue that this finding has primarily been reported for datasets satisfying a \emph{covariate shift} assumption. When the dataset satisfies a \emph{posterior drift} assumption instead, we show that ``domain-informed ERM,'' wherein feature vectors are augmented with domain-specific information, outperforms pooling ERM. These claims are supported by a theoretical framework and experiments on language and vision tasks.

Page Count
20 pages

Category
Computer Science:
Machine Learning (CS)