Detect and Correct: A Selective Noise Correction Method for Learning with Noisy Labels
By: Yuval Grinberg , Nimrod Harel , Jacob Goldberger and more
Potential Business Impact:
Fixes computer mistakes in learning data.
Falsely annotated samples, also known as noisy labels, can significantly harm the performance of deep learning models. Two main approaches for learning with noisy labels are global noise estimation and data filtering. Global noise estimation approximates the noise across the entire dataset using a noise transition matrix, but it can unnecessarily adjust correct labels, leaving room for local improvements. Data filtering, on the other hand, discards potentially noisy samples but risks losing valuable data. Our method identifies potentially noisy samples based on their loss distribution. We then apply a selection process to separate noisy and clean samples and learn a noise transition matrix to correct the loss for noisy samples while leaving the clean data unaffected, thereby improving the training process. Our approach ensures robust learning and enhanced model performance by preserving valuable information from noisy samples and refining the correction process. We applied our method to standard image datasets (MNIST, CIFAR-10, and CIFAR-100) and a biological scRNA-seq cell-type annotation dataset. We observed a significant improvement in model accuracy and robustness compared to traditional methods.
Similar Papers
Enhanced Sample Selection with Confidence Tracking: Identifying Correctly Labeled yet Hard-to-Learn Samples in Noisy Data
CV and Pattern Recognition
Teaches computers to learn from messy pictures.
SelectMix: Enhancing Label Noise Robustness through Targeted Sample Mixing
Machine Learning (CS)
Teaches computers to learn from messy, wrong information.
Best Transition Matrix Esitimation or Best Label Noise Robustness Classifier? Two Possible Methods to Enhance the Performance of T-revision
Machine Learning (CS)
Fixes computer learning when data has mistakes.