Cross-Semantic Transfer Learning for High-Dimensional Linear Regression
By: Jiancheng Jiang, Xuejun Jiang, Hongxia Jin
Current transfer learning methods for high-dimensional linear regression assume feature alignment across domains, restricting their applicability to semantically matched features. In many real-world scenarios, however, distinct features in the target and source domains can play similar predictive roles, creating a form of cross-semantic similarity. To leverage this broader transferability, we propose the Cross-Semantic Transfer Learning (CSTL) framework. It captures potential relationships by comparing each target coefficient with all source coefficients through a weighted fusion penalty. The weights are derived from the derivative of the SCAD penalty, effectively approximating an ideal weighting scheme that preserves transferable signals while filtering out source-specific noise. For computational efficiency, we implement CSTL using the Alternating Direction Method of Multipliers (ADMM). Theoretically, we establish that under mild conditions, CSTL achieves the oracle estimator with overwhelming probability. Empirical results from simulations and a real-data application confirm that CSTL outperforms existing methods in both cross-semantic and partial signal similarity settings.
Similar Papers
Quantifying Dataset Similarity to Guide Transfer Learning
Machine Learning (Stat)
Tells computers if learning from old data helps.
Coefficient Shape Transfer Learning for Functional Linear Regression
Methodology
Helps computers learn from less data.
Source-free Semantic Regularization Learning for Semi-supervised Domain Adaptation
CV and Pattern Recognition
Teaches computers to learn from less labeled data.