Abstract
Noisy labels are often present in large, accessible datasets. Learning with noisy labels can degrade the generalization performance of DNNs. While Semi-Supervised Learning (SSL) approaches have shown promise by predicting pseudo-labels to correct noisy labels, we find and demonstrate a fundamental limitation of SSL-based label correction methods: hard samples near decision boundaries significantly weaken the memorization effect that these methods rely on. This leads to erroneous pseudo-labels and creates a negative feedback loop where models gradually memorize these errors, further degrading performance. To overcome the issue, inspired by AdaBoost and building on these insights, we propose HEALON (Hard samplE Adaptive Labeling with Optimal reweighting for Noisy labels), a novel framework for learning with noisy labels that effectively addresses the hard sample challenge through an optimal weighting strategy that balances their influence during training. The framework primarily consists of two steps: Weighted Implicit Ensemble (
Keywords
Get full access to this article
View all access options for this article.
