Skip to main content

Learning from Noisy Labels with Coarse-to-Fine Sample Credibility Modeling

  • Conference paper
  • First Online:
Computer Vision – ECCV 2022 Workshops (ECCV 2022)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 13802))

Included in the following conference series:

Abstract

Training deep neural network (DNN) with noisy labels is practically challenging since inaccurate labels severely degrade the generalization ability of DNN. Previous efforts tend to handle part or full data in a unified denoising flow via identifying noisy data with a coarse small-loss criterion to mitigate the interference from noisy labels, ignoring the fact that the difficulties of noisy samples are different, thus a rigid and unified data selection pipeline cannot tackle this problem well . In this paper, we first propose a coarse-to-fine robust learning method called CREMA, to handle noisy data in a divide-and-conquer manner. In coarse-level, clean and noisy sets are firstly separated in terms of credibility in a statistical sense. Since it is practically impossible to categorize all noisy samples correctly, we further process them in a fine-grained manner via modeling the credibility of each sample. Specifically, for the clean set, we deliberately design a memory-based modulation scheme to dynamically adjust the contribution of each sample in terms of its historical credibility sequence during training, thus alleviating the effect from noisy samples incorrectly grouped into the clean set. Meanwhile, for samples categorized into the noisy set, a selective label update strategy is proposed to correct noisy labels while mitigating the problem of correction error. Extensive experiments are conducted on benchmarks of different modality, including image classification (CIFAR, Clothing1M etc.) and text recognition (IMDB), with either synthetic or natural semantic noises, demonstrating the superiority and generality of CREMA.

B. Zhang, Y.Li and Y.Tu—Contributed equally to this work.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 149.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 199.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Arazo, E., Ortego, D., Albert, P., O’Connor, N.E., McGuinness, K.: Unsupervised label noise modeling and loss correction. In: Proc. International Conference on Machine Learning (ICML) (2019)

    Google Scholar 

  2. Arpit, D., et al.: A closer look at memorization in deep networks. In: Proc. International Conference on Machine Learning (ICML), pp. 233–242 (2017)

    Google Scholar 

  3. Berthelot, D., Carlini, N., Goodfellow, I.J., Papernot, N., Oliver, A., Raffel, C.: Mixmatch: a holistic approach to semi-supervised learning. Proc. Advances in Neural Information Processing Systems (NeurIPS) (2019)

    Google Scholar 

  4. Chang, H.S., Learned-Miller, E., McCallum, A.: Active Bias: training more accurate neural networks by emphasizing high variance samples. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 1002–1012 (2017)

    Google Scholar 

  5. Dehghani, M., Severyn, A., Rothe, S., Kamps, J.: Avoiding your teacher’s mistakes: training neural networks with controlled weak supervision. (2017) arXiv preprint arXiv:1711.00313

  6. Dehghani, M., Severyn, A., Rothe, S., Kamps, J.: Learning to learn from weak supervision by full supervision. In: Proc. Advances in Neural Information Processing Systems Workshop (NeurIPSW) (2017)

    Google Scholar 

  7. Ghosh, A., Kumar, H., Sastry, P.: Robust loss functions under label noise for deep neural networks. In: Proc. Association for the Advancement of Artificial Intelligence (AAAI) (2017)

    Google Scholar 

  8. Goldberger, J., Ben-Reuven, E.: Training deep neural-networks using a noise adaptation layer. In: Proc. International Conference on Learning Representations (ICLR) (2017)

    Google Scholar 

  9. Han, B., et al.: Masking: A new perspective of noisy supervision. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 5836–5846 (2018)

    Google Scholar 

  10. Han, B., et al.: Co-teaching: robust training of deep neural networks with extremely noisy labels. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 8527–8537 (2018)

    Google Scholar 

  11. Han, J., Luo, P., Wang, X.: Deep self-learning from noisy labels. In: Proc. IEEE International Conference on Computer Vision (ICCV), pp. 5138–5147 (2019)

    Google Scholar 

  12. Hendrycks, D., Mazeika, M., Wilson, D., Gimpel, K.: Using trusted data to train deep networks on labels corrupted by severe noise. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 10456–10465 (2018)

    Google Scholar 

  13. Houle, M.E.: Local intrinsic dimensionality I: an extreme-value-theoretic foundation for similarity applications. In: Proc. International Conference on Similarity Search and Applications (SISAP), pp. 64–79 (2017)

    Google Scholar 

  14. Jiang, L., Zhou, Z., Leung, T., Li, L.J., Fei-Fei, L.: MentorNet: learning data-driven curriculum for very deep neural networks on corrupted labels. In: Proc. International Conference on Machine Learning (ICML) (2018)

    Google Scholar 

  15. Krizhevsky, A., Hinton, G., et al.: Learning multiple layers of features from tiny images (2009)

    Google Scholar 

  16. Li, J., Socher, R., Hoi, S.C.: Dividemix: learning with noisy labels as semi-supervised learning. (2020) arXiv preprint arXiv:2002.07394

  17. Li, J., Wong, Y., Zhao, Q., Kankanhalli, M.: learning to learn from noisy labeled data. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5046–5054 (2019)

    Google Scholar 

  18. Li, J., Xiong, C., Hoi, S.C.: Learning from noisy data with robust representation learning. In: Proc. IEEE International Conference on Computer Vision (ICCV), pp. 9485–9494 (2021)

    Google Scholar 

  19. Li, X., Liu, T., Han, B., Niu, G., Sugiyama, M.: Provably end-to-end label-noise learning without anchor points. Proc. International Conference on Machine Learning (ICML) (2021)

    Google Scholar 

  20. Li, Y., Yang, J., Song, Y., Cao, L., Luo, J., Li, L.J.: Learning from noisy labels with distillation. In: Proc. IEEE International Conference on Computer Vision (ICCV), pp. 1910–1918 (2017)

    Google Scholar 

  21. Litany, O., Freedman, D.: Soseleto: a unified approach to transfer learning and training with noisy labels. (2018) arXiv preprint arXiv:1805.09622

  22. Liu, T., Tao, D.: Classification with noisy labels by importance reweighting. IEEE Trans. Pattern Anal. Mach. Intell. (TPAMI) 38(3), 447–461 (2015)

    Article  Google Scholar 

  23. Lyu, Y., Tsang, I.W.: Curriculum loss: robust learning and generalization against label corruption. In: Proc. International Conference on Learning Representations (ICLR) (2020)

    Google Scholar 

  24. Ma, X., et al.: Dimensionality-driven learning with noisy labels. In: Proc. International Conference on Machine Learning (ICML) (2018)

    Google Scholar 

  25. Maas, A., Daly, R.E., Pham, P.T., Huang, D., Ng, A.Y., Potts, C.: Learning word vectors for sentiment analysis. In: Proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies, pp. 142–150 (2011)

    Google Scholar 

  26. Malach, E., Shalev-Shwartz, S.: Decoupling when to update from how to update. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 960–970 (2017)

    Google Scholar 

  27. Manning, C., Schutze, H.: Foundations of statistical natural language processing. MIT press (1999)

    Google Scholar 

  28. Nguyen, D.T., Mummadi, C.K., Ngo, T.P.N., Nguyen, T.H.P., Beggel, L., Brox, T.: SELF: learning to filter noisy labels with self-Ensembling. In: Proc. International Conference on Learning Representations (ICLR) (2020)

    Google Scholar 

  29. Patrini, G., Rozza, A., Krishna Menon, A., Nock, R., Qu, L.: Making deep neural networks robust to label noise: a loss correction approach. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1944–1952 (2017)

    Google Scholar 

  30. Permuter, H., Francos, J.M., Jermyn, I.: A study of gaussian mixture models of color and texture features for image classification and segmentation. Pattern Recogn. (PR). 39, 695–706 (2006)

    Article  MATH  Google Scholar 

  31. Pleiss, G., Zhang, T., Elenberg, E.R., Weinberger, K.Q.: Identifying mislabeled data using the area under the margin ranking. In: Proc. Advances in Neural Information Processing Systems (NeurIPS) (2020)

    Google Scholar 

  32. Reed, S., Lee, H., Anguelov, D., Szegedy, C., Erhan, D., Rabinovich, A.: Training deep neural networks on noisy labels with bootstrapping. In: Proc. International Conference on Learning Representations (ICLR) (2015)

    Google Scholar 

  33. Ren, M., Zeng, W., Yang, B., Urtasun, R.: Learning to reweight examples for robust deep learning. In: Proc. International Conference on Machine Learning (ICML) (2018)

    Google Scholar 

  34. Shu, J., et al.: Meta-Weight-Net: learning an explicit mapping for sample weighting. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 1917–1928 (2019)

    Google Scholar 

  35. Song, H., Kim, M., Lee, J.G.: SELFIE: refurbishing unclean samples for robust deep learning. In: Proc. International Conference on Machine Learning (ICML), pp. 5907–5915 (2019)

    Google Scholar 

  36. Tanaka, D., Ikami, D., Yamasaki, T., Aizawa, K.: Joint optimization framework for learning with noisy labels. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5552–5560 (2018)

    Google Scholar 

  37. Veit, A., Alldrin, N., Chechik, G., Krasin, I., Gupta, A., Belongie, S.: Learning from noisy large-scale datasets with minimal supervision. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017)

    Google Scholar 

  38. Wang, R., Liu, T., Tao, D.: Multiclass learning with partially corrupted labels. IEEE Trans. Neural Networks. Learn. Syst. 29(6), 2568–2580 (2017)

    Article  MathSciNet  Google Scholar 

  39. Wang, X., Hua, Y., Kodirov, E., Clifton, D.A., Robertson, N.M.: Proselflc: progressive self label correction for training robust deep neural networks. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 752–761 (2021)

    Google Scholar 

  40. Wang, X., Hua, Y., Kodirov, E., Robertson, N.M.: Imae for noise-robust learning: mean absolute error does not treat examples equally and gradient magnitude’s variance matters. (2019) arXiv preprint arXiv:1903.12141

  41. Wang, X., Kodirov, E., Hua, Y., Robertson, N.M.: Derivative manipulation for general example weighting. (2019) arXiv preprint arXiv:1905.11233

  42. Wang, Y., Ma, X., Chen, Z., Luo, Y., Yi, J., Bailey, J.: Symmetric cross entropy for robust learning with noisy labels. In: Proc. IEEE International Conference on Computer Vision (ICCV), pp. 322–330 (2019)

    Google Scholar 

  43. Wei, H., Feng, L., Chen, X., An, B.: Combating noisy labels by agreement: a joint training method with co-regularization. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13726–13735 (2020)

    Google Scholar 

  44. Xiao, T., Xia, T., Yang, Y., Huang, C., Wang, X.: Learning from massive noisy labeled data for image classification. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2691–2699 (2015)

    Google Scholar 

  45. Yao, Q., Yang, H., Han, B., Niu, G., Kwok, J.: Searching to exploit memorization effect in learning with noisy labels. In: Proc. International Conference on Machine Learning (ICML) (2020)

    Google Scholar 

  46. Yi, K., Wu, J.: Probabilistic end-to-end noise correction for learning with noisy labels. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7017–7025 (2019)

    Google Scholar 

  47. Yu, X., Han, B., Yao, J., Niu, G., Tsang, I.W., Sugiyama, M.: How does disagreement help generalization against label corruption? In: Proc. International Conference on Machine Learning (ICML) (2019)

    Google Scholar 

  48. Zhang, C., Bengio, S., Hardt, M., Recht, B., Vinyals, O.: Understanding deep learning requires rethinking generalization. In: Proc. International Conference on Learning Representations (ICLR) (2017)

    Google Scholar 

  49. Zhang, H., Cissé, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: Beyond empirical risk minimization. In: Proc. International Conference on Learning Representations (ICLR) (2018)

    Google Scholar 

  50. Zhang, M., Lee, J., Agarwal, S.: Learning from noisy labels with no change to the training process. In: Proc. International Conference on Machine Learning (ICML), pp. 12468–12478 (2021)

    Google Scholar 

  51. Zhang, W., Wang, Y., Qiao, Y.: Metacleaner: Learning to hallucinate clean representations for noisy-labeled visual recognition. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7365–7374 (2019)

    Google Scholar 

  52. Zhang, Y., Zheng, S., Wu, P., Goswami, M., Chen, C.: Learning with feature-dependent label noise: a progressive approach. In: Proc. International Conference on Learning Representations (ICLR) (2021)

    Google Scholar 

  53. Zhang, Z., Sabuncu, M.: Generalized cross entropy loss for training deep neural networks with noisy labels. In: Proc. Advances in Neural Information Processing Systems (NeurIPS), pp. 8778–8788 (2018)

    Google Scholar 

  54. Zhang, Z., Zhang, H., Arik, S.O., Lee, H., Pfister, T.: Distilling effective supervision from severe label noise. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 9294–9303 (2020)

    Google Scholar 

  55. Zheltonozhskii, E., Baskin, C., Mendelson, A., Bronstein, A.M., Litany, O.: Contrast to divide: self-supervised pre-training for learning with noisy labels. (2021) arXiv preprint arXiv:2103.13646

  56. Zhou, T., Wang, S., Bilmes, J.: Robust curriculum learning: From clean label detection to noisy label self-correction. In: Proc. International Conference on Learning Representations (ICLR) (2021)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yabiao Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Zhang, B. et al. (2023). Learning from Noisy Labels with Coarse-to-Fine Sample Credibility Modeling. In: Karlinsky, L., Michaeli, T., Nishino, K. (eds) Computer Vision – ECCV 2022 Workshops. ECCV 2022. Lecture Notes in Computer Science, vol 13802. Springer, Cham. https://doi.org/10.1007/978-3-031-25063-7_2

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-25063-7_2

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-25062-0

  • Online ISBN: 978-3-031-25063-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics