Skip to main content

Combating Quality Distortion in Federated Learning with Collaborative Data Selection

  • Conference paper
  • First Online:
Advances in Knowledge Discovery and Data Mining (PAKDD 2024)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 14647))

Included in the following conference series:

  • 204 Accesses

Abstract

Federated Learning (FL), a paradigm facilitating collaborative model training across distributed devices, has attracted substantial attention due to its potential to address privacy concerns and data localization requirements. However, the inherent inaccessibility of data poses a critical challenge in ensuring data quality within FL systems. Consequently, FL systems grapple with a range of data-related issues, encompassing erroneous samples, imbalanced data distributions, and data skew, all of which impose a significant impact on model performance. Therefore, the judicious selection of appropriate data for training is of paramount importance as it seeks to ameliorate these challenges.

This research paper tackles a crucial but often overlooked concern: the presence of low-quality data samples. In such circumstances, we introduce an innovative algorithm that strategically curates a subset of data for each training iteration, with the overarching objective of optimizing the model’s accuracy while simultaneously addressing privacy concerns and reducing communication costs. Our primary innovation lies in the global selection of data, in contrast to the conventional approach that relies on individualized, client-specific data selection.

Furthermore, we introduce a novel medical dataset tailored specifically for classification tasks. This dataset intentionally incorporates various attributes associated with low-quality data to effectively replicate real-world conditions. Through rigorous empirical evaluation, we show the effectiveness of our algorithm using this dataset. The results demonstrate a notable improvement of approximately 2–3% in model performance, particularly in scenarios characterized by imbalanced data distributions.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 119.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 139.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The configurations of this experiment is described in Sect. 4.1.

  2. 2.

    Datasets are available at https://github.com/duclong1009/S-Selection.

  3. 3.

    By using the transformation function ImageEnhance of the PIL library.

References

  1. da Costa, G.B.P., Contato, W.A., Nazare, T.S., Batista Neto, J.E.S., Ponti, M.: An empirical study on the effects of different types of noise in image classification tasks (2016)

    Google Scholar 

  2. Dodge, S., Karam, L.: Understanding how image quality affects deep neural networks (2016)

    Google Scholar 

  3. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition (2015)

    Google Scholar 

  4. He, Z., Rakin, A.S., Fan, D.: Parametric noise injection: trainable randomness to improve deep neural network robustness against adversarial attack. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2019

    Google Scholar 

  5. Holmstrom, L., Koistinen, P., et al.: Using additive noise in back-propagation training. IEEE Trans. Neural Netw. 3(1), 24–38 (1992)

    Article  Google Scholar 

  6. Jiang, A.H., et al.: Accelerating deep learning by focusing on the biggest losers (2019)

    Google Scholar 

  7. Katharopoulos, A., Fleuret, F.: Not all samples are created equal: deep learning with importance sampling. In: Proceedings of the 35th International Conference on Machine Learning, vol. 80, pp. 2525–2534. PMLR, July 2018

    Google Scholar 

  8. Killamsetty, K., Sivasubramanian, D., Ramakrishnan, G., De, A., Iyer, R.: Grad-match: gradient matching based data subset selection for efficient deep model training (2021)

    Google Scholar 

  9. Krizhevsky, A., et al.: Learning multiple layers of features from tiny images (2009)

    Google Scholar 

  10. Li, A., Zhang, L., Tan, J., Qin, Y., Wang, J., Li, X.-Y.: Sample-level data selection for federated learning. In: IEEE INFOCOM 2021 - IEEE Conference on Computer Communications, pp. 1–10 (2021)

    Google Scholar 

  11. Li, Q., Diao, Y., Chen, Q., He, B.: Federated learning on non-IID data silos: an experimental study. In: 2022 IEEE 38th International Conference on Data Engineering (ICDE), pp. 965–978. IEEE (2022)

    Google Scholar 

  12. Li, T., Sahu, A.K., Zaheer, M., Sanjabi, M., Talwalkar, A., Smith, V.: Federated optimization in heterogeneous networks (2020)

    Google Scholar 

  13. Li, X., Huang, K., Yang, W., Wang, S., Zhang, Z.: On the convergence of fedavg on non-IID data. arXiv preprint arXiv:1907.02189 (2019)

  14. McMahan, B., Moore, E., Ramage, D., Hampson, S., Aguera, B., Arcas: Communication-efficient learning of deep networks from decentralized data, 54, 1273–1282 (2017)

    Google Scholar 

  15. Paul, M., Ganguli, S., Dziugaite, G.K.: Deep learning on a data diet: finding important examples early in training (2023)

    Google Scholar 

  16. Pillutla, K., Laguel, Y., Malick, J., Harchaoui, Z.: Tackling distribution shifts in federated learning with superquantile aggregation. In: NeurIPS 2022 Workshop on Distribution Shifts: Connecting Methods and Applications (2022)

    Google Scholar 

  17. Qin, Z., et al.: Infobatch: lossless training speed up by unbiased dynamic data pruning (2023)

    Google Scholar 

  18. Sorscher, B., Geirhos, R., Shekhar, S., Ganguli, S., Morcos, A.S.: Beyond neural scaling laws: beating power law scaling via data pruning (2023)

    Google Scholar 

  19. Tolpegin, V., Truex, S., Gursoy, M.E., Liu, L.: Data poisoning attacks against federated learning systems. In: Chen, L., Li, N., Liang, K., Schneider, S. (eds.) ESORICS 2020. LNCS, vol. 12308, pp. 480–501. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58951-6_24

    Chapter  Google Scholar 

  20. Truong, T.N., Gerofi, B., Martinez-Noriega, E.J., Trahay, F., Wahib, M.: KAKURENBO: adaptively hiding samples in deep neural network training. In: Thirty-Seventh Conference on Neural Information Processing Systems (2023)

    Google Scholar 

  21. Wu, C., Yang, X., Zhu, S., Mitra, P.: Mitigating backdoor attacks in federated learning (2021)

    Google Scholar 

  22. Yang, S., Park, H., Byun, J., Kim, C.: Robust federated learning with noisy labels. IEEE Intell. Syst. 37(2), 35–43 (2022)

    Article  Google Scholar 

  23. Yang, S., Xie, Z., Peng, H., Xu, M., Sun, M., Li, P.: Dataset pruning: reducing training data by examining generalization influence. In: The Eleventh International Conference on Learning Representations (2023)

    Google Scholar 

  24. Yu, X., Han, B., Yao, J., Niu, G., Tsang, I.W., Sugiyama, M.: How does disagreement help generalization against label corruption? (2019)

    Google Scholar 

  25. Zhou, T., Konukoglu, E.: FedFA: federated feature augmentation (2023)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Phi Le Nguyen or Thao Nguyen Truong .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Nguyen, D.L., Nguyen, P.L., Truong, T.N. (2024). Combating Quality Distortion in Federated Learning with Collaborative Data Selection. In: Yang, DN., Xie, X., Tseng, V.S., Pei, J., Huang, JW., Lin, J.CW. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2024. Lecture Notes in Computer Science(), vol 14647. Springer, Singapore. https://doi.org/10.1007/978-981-97-2259-4_14

Download citation

  • DOI: https://doi.org/10.1007/978-981-97-2259-4_14

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-97-2261-7

  • Online ISBN: 978-981-97-2259-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics