Score: 3

Handling Label Noise via Instance-Level Difficulty Modeling and Dynamic Optimization

Published: May 1, 2025 | arXiv ID: 2505.00812v3

By: Kuan Zhang , Chengliang Chai , Jingzhe Xu and more

BigTech Affiliations: Massachusetts Institute of Technology

Potential Business Impact:

Fixes computer mistakes from bad data.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Recent studies indicate that deep neural networks degrade in generalization performance under noisy supervision. Existing methods focus on isolating clean subsets or correcting noisy labels, facing limitations such as high computational costs, heavy hyperparameter tuning process, and coarse-grained optimization. To address these challenges, we propose a novel two-stage noisy learning framework that enables instance-level optimization through a dynamically weighted loss function, avoiding hyperparameter tuning. To obtain stable and accurate information about noise modeling, we introduce a simple yet effective metric, termed wrong event, which dynamically models the cleanliness and difficulty of individual samples while maintaining computational costs. Our framework first collects wrong event information and builds a strong base model. Then we perform noise-robust training on the base model, using a probabilistic model to handle the wrong event information of samples. Experiments on five synthetic and real-world LNL benchmarks demonstrate our method surpasses state-of-the-art methods in performance, achieves a nearly 75% reduction in computational time and improves model scalability.

Country of Origin
πŸ‡ΊπŸ‡Έ πŸ‡¨πŸ‡³ United States, China

Page Count
20 pages

Category
Computer Science:
Machine Learning (CS)