Skip to main content

Differential Private (Random) Decision Tree Without Adding Noise

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2023)

Abstract

The decision tree is a typical algorithm in machine learning and has multiple expanded variations. However, regarding privacy, few in the variations reached practical level due to many challenges on balancing privacy preservation and performance. In this paper, we propose a method of applying privacy preservation to the (random) decision tree, which is a variation of the expanded decision tree proposed by Fan et al. in 2003, to achieve the following goals:

  • Model training with data belonging to multiple organizations and concealing these data among organizations.

  • No leakage of training data from trained models.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

      We use the Pedersen commitment and the expanded ElGamal encryption (plaintext m is encoded as \(g^m\)).

  2. 2.

      (f-out-of-M)-threshold decryption means any f out of M participants cooperate can decrypt the ciphertexts, but any participants less than f cannot.

References

  1. Mobile kukan toukei (guidelines). https://www.intage.co.jp/english/service/platform/mobile-kukan-toukei/ (https://www.docomo.ne.jp/english/binary/pdf/service/world/inroaming/inroaming_service/Mobile_Kukan_Toukei_Guidelines.pdf)

  2. Aminifar, A., Rabbi, F., Pun, K.I., Lamo, Y.: Privacy preserving distributed extremely randomized trees. In: Hung, C., Hong, J., Bechini, A., Song, E. (eds.) SAC ’21: The 36th ACM/SIGAPP Symposium on Applied Computing, Virtual Event, Republic of Korea, March 22–26, 2021, pp. 1102–1105. ACM (2021). https://doi.org/10.1145/3412841.3442110

  3. Becker, B., Kohavi, R.: Adult. UCI Machine Learning Repository (1996)

    Google Scholar 

  4. Dwork, C., Kenthapadi, K., McSherry, F., Mironov, I., Naor, M.: Our data, ourselves: privacy via distributed noise generation. In: Vaudenay, S. (ed.) EUROCRYPT 2006. LNCS, vol. 4004, pp. 486–503. Springer, Heidelberg (2006). https://doi.org/10.1007/11761679_29

    Chapter  Google Scholar 

  5. Dwork, C., Roth, A.: The algorithmic foundations of differential privacy. Found. Trends Theor. Comput. Sci. 9(3–4), 211–407 (2014). https://doi.org/10.1561/0400000042

    Article  MathSciNet  Google Scholar 

  6. Fan, W., Wang, H., Yu, P.S., Ma, S.: Is random model better? on its accuracy and efficiency. In: Proceedings of the 3rd IEEE International Conference on Data Mining (ICDM 2003), 19–22 December 2003, Melbourne, Florida, USA, pp. 51–58. IEEE Computer Society (2003). https://doi.org/10.1109/ICDM.2003.1250902

  7. Jagannathan, G., Pillaipakkamnatt, K., Wright, R.N.: A practical differentially private random decision tree classifier. Trans. Data Priv. 5(1), 273–295 (2012). www.tdp.cat/issues11/abs.a082a11.php

  8. Li, N., Qardaji, W.H., Su, D.: On sampling, anonymization, and differential privacy or, k-anonymization meets differential privacy. In: Youm, H.Y., Won, Y. (eds.) 7th ACM Symposium on Information, Compuer and Communications Security, ASIACCS ’12, Seoul, Korea, May 2–4, 2012, pp. 32–33. ACM (2012). https://doi.org/10.1145/2414456.2414474

  9. Rajkovic, V.: Nursery. UCI Mach. Learn. Reposit. (1997). https://doi.org/10.24432/C5P88W

  10. Vaidya, J., Shafiq, B., Fan, W., Mehmood, D., Lorenzi, D.: A random decision tree framework for privacy-preserving data mining. IEEE Trans. Dependable Secur. Comput. 11(5), 399–411 (2014). https://doi.org/10.1109/TDSC.2013.43

Download references

Acknowledgements

This work was supported in part by JST CREST Grant Number JPMJCR21M1, and JSPS KAKENHI Grant Number JP20K11826, Japan.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lihua Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Nojima, R., Wang, L. (2024). Differential Private (Random) Decision Tree Without Adding Noise. In: Luo, B., Cheng, L., Wu, ZG., Li, H., Li, C. (eds) Neural Information Processing. ICONIP 2023. Communications in Computer and Information Science, vol 1963. Springer, Singapore. https://doi.org/10.1007/978-981-99-8138-0_14

Download citation

  • DOI: https://doi.org/10.1007/978-981-99-8138-0_14

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-99-8137-3

  • Online ISBN: 978-981-99-8138-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics