Skip to main content

Asynchronous Compatible D-ML System for Edge Devices

  • Conference paper
  • First Online:
Advances in Artificial Intelligence and Security (ICAIS 2021)

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 1422))

Included in the following conference series:

  • 1149 Accesses

Abstract

The parameter server system (Parameter Server, PS) is the most widely used distributed machine learning system. Its core module is the communication module, which mainly includes three parts: communication content, communication frequency, and communication pace. The communication scheme of the existing PS system will have problems such as waste of computing power resources and a large proportion of communication overhead. This paper hopes to design and implement a distributed set that saves computing power and has a smaller communication overhead based on the communication module of the original PS system. Machine learning system, this article mainly has the following contributions: First, through the analysis and experiment of Bulk Synchronous Parallel (BSP) communication pace, based on the superiority of BSP algorithm level, it overcomes the waste of computing power of BSP at the system level, and designs and implements a new set of Communication pace, the new communication pace under the same training task can save more computing power and achieve higher model accuracy. Secondly, this article combines the quantization compression technology with the parameter server system, and uses the quantization technology to compress and decompress the communication content in the parameter server system, so that the volume of the communication content during the operation of the parameter server system is significantly reduced, thereby reducing the communication overhead and improving Speedup ratio of the system. Finally, based on the analysis and experiment of the change trend of the model accuracy rate, this paper designs and implements a dynamic communication frequency adjustment scheme based on the change trend of the model accuracy rate, which reduces the overall communication frequency during the operation of the parameter server system and reduces the parameter server. T Experiments show that the system designed in this paper is superior to the existing mainstream solutions in terms of communication pace, communication content, and communication frequency. When training the same model, this system compared with the existing parameter server system in model convergence time, model There is a noticeable improvement in the highest accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Liu, Y.: Disturbed ML: Algorithm, theory and Practice

    Google Scholar 

  2. Low, Y., Gonzalez, J., Kyrola, A., Bickson, D., Guestrin, C., Hellerstein, J.M.: Distributed graphlab: a framework for machine learning in the cloud. arXiv preprint arXiv:1204.6078 (2012)

  3. Ahmed, A., Shervashidze, N., Narayanamurthy, S., Josifovski, V., Smola, A.J.: Distributed large-scale natural graph factorization. In: Proceedings of the 22nd international conference on World Wide Web, pp. 37–48 (2013)

    Google Scholar 

  4. Dean, J., et al.: Large scale distributed deep networks. In Advances in neural information processing systems, pp. 1223–1231 (2012)

    Google Scholar 

  5. Chen, W.-Y., Song, Y., Bai, H., Lin, C.-J., Chang, E.Y.: Parallel spectral clustering in distributed systems. IEEE Trans. Patt. Anal. Mach. Intell. 33(3), 568–586 (2010)

    Google Scholar 

  6. Agarwal, A., Wainwright, M.J., Duchi, J.C.: Distributed dual averaging in networks. In: Advances in Neural Information Processing Systems, pp. 550–558 (2010)

    Google Scholar 

  7. Han, S., Mao, H., Dally, W.J.: Deep compression: compressing deep neural networks with pruning, trained quantization and Huffman coding. arXiv preprint arXiv:1510.00149 (2015)

  8. Gong, Y., Liu, L., Yang, M., Bourdev, L.: Compressing deep convolutional networks using vector quantization. arXiv preprint arXiv:1412.6115 (2014)

  9. Berkeley University of California. Ray. [EB/OL]. https://ray.io/

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ge, G., Zhu, F., Huang, Y. (2021). Asynchronous Compatible D-ML System for Edge Devices. In: Sun, X., Zhang, X., Xia, Z., Bertino, E. (eds) Advances in Artificial Intelligence and Security. ICAIS 2021. Communications in Computer and Information Science, vol 1422. Springer, Cham. https://doi.org/10.1007/978-3-030-78615-1_15

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-78615-1_15

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-78614-4

  • Online ISBN: 978-3-030-78615-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics