OASIS: Open-world Adaptive Self-supervised and Imbalanced-aware System
By: Miru Kim, Mugon Joe, Minhae Kwon
Potential Business Impact:
Teaches computers to learn from messy, incomplete data.
The expansion of machine learning into dynamic environments presents challenges in handling open-world problems where label shift, covariate shift, and unknown classes emerge. Post-training methods have been explored to address these challenges, adapting models to newly emerging data. However, these methods struggle when the initial pre-training is performed on class-imbalanced datasets, limiting generalization to minority classes. To address this, we propose a method that effectively handles open-world problems even when pre-training is conducted on imbalanced data. Our contrastive-based pre-training approach enhances classification performance, particularly for underrepresented classes. Our post-training mechanism generates reliable pseudo-labels, improving model robustness against open-world problems. We also introduce selective activation criteria to optimize the post-training process, reducing unnecessary computation. Extensive experiments demonstrate that our method significantly outperforms state-of-the-art adaptation techniques in both accuracy and efficiency across diverse open-world scenarios.
Similar Papers
Open-Set Domain Adaptation Under Background Distribution Shift: Challenges and A Provably Efficient Solution
Machine Learning (CS)
Helps computers spot new things even when data changes.
Self-Supervised Pre-Training with Equilibrium Constraints
Machine Learning (CS)
Teaches computers to learn from mixed data better.
World Model Robustness via Surprise Recognition
Machine Learning (CS)
Keeps AI safe from bad or surprising information.