Skip to main content

MLDev: Data Science Experiment Automation and Reproducibility Software

  • Conference paper
  • First Online:
Data Analytics and Management in Data Intensive Domains (DAMDID/RCDL 2021)

Abstract

In this paper, we explore the challenges of automating experiments in data science. We propose an extensible experiment model as a foundation for integration of different open source tools for running research experiments. We implement our approach in a prototype open source MLDev software package and evaluate it in a series of experiments yielding promising results. Comparison with other state-of-the-art tools signifies novelty of our approach.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 69.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 89.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Data version control tool (dvc). https://dvc.org. Accessed 14 June 2021

  2. MLDev. An open source data science experimentation and reproducibility software. https://mlrep.gitlab.io/mldev. Accessed 14 June 2021

  3. Berkus, J.: The 5 types of open source projects. https://wackowiki.org/doc/Org/Articles/5TypesOpenSourceProjects. Accessed 14 June 2021

  4. Bisong, E.: Google colaboratory. In: Building Machine Learning and Deep Learning Models on Google Cloud Platform, pp. 59–64. Springer. Apress, Berkeley, CA (2019). https://doi.org/10.1007/978-1-4842-4470-8_7

  5. Bordes, A., Usunier, N., Garcia-Duran, A., Weston, J., Yakhnenko, O.: Translating embeddings for modeling multi-relational data. In: Neural Information Processing Systems (NIPS), pp. 1–9 (2013)

    Google Scholar 

  6. Bunel, R., Hausknecht, M., Devlin, J., Singh, R., Kohli, P.: Leveraging grammar and reinforcement learning for neural program synthesis. In: International Conference on Learning Representations (2018). https://openreview.net/forum?id=H1Xw62kRZ

  7. Chue Hong, N.: Reproducibility badging and definitions: a recommended practice of the national information standards organization. Nat. Inf. Stan. Organ. (NISO) (2021). https://doi.org/10.3789/niso-rp-31-2021

  8. Di Tommaso, P., Chatzou, M., Floden, E.W., Barja, P.P., Palumbo, E., Notredame, C.: Nextflow enables reproducible computational workflows. Nat. Biotechnol. 35(4), 316–319 (2017)

    Article  Google Scholar 

  9. Dmitriev, S.: Language oriented programming: the next programming paradigm. JetBrains onboard 1(2), 1–13 (2004)

    Google Scholar 

  10. Gundersen, O.E., Gil, Y., Aha, D.W.: On reproducible AI: towards reproducible research, open science, and digital scholarship in AI publications. AI Mag. 39(3), 56–68 (2018)

    Google Scholar 

  11. Hutson, M.: Artificial intelligence faces reproducibility crisis (2018)

    Google Scholar 

  12. Isdahl, R., Gundersen, O.E.: Out-of-the-box reproducibility: a survey of machine learning platforms. In: 2019 15th International Conference on eScience (eScience), pp. 86–95. IEEE (2019)

    Google Scholar 

  13. Khritankov, A.: Analysis of hidden feedback loops in continuous machine learning systems. arXiv preprint arXiv:2101.05673 (2021)

  14. Khritankov, A.: Hidden feedback loops in machine learning systems: a simulation model and preliminary results. In: Winkler, D., Biffl, S., Mendez, D., Wimmer, M., Bergsmann, J. (eds.) SWQD 2021. LNBIP, vol. 404, pp. 54–65. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-65854-0_5

    Chapter  Google Scholar 

  15. Kluyver, T., et al.: Jupyter Notebooks-a Publishing Format for Reproducible Computational Workflows, vol. 2016 (2016)

    Google Scholar 

  16. Nathani, D., Chauhan, J., Sharma, C., Kaul, M.: Learning attention-based embeddings for relation prediction in knowledge graphs. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 4710–4723. Association for Computational Linguistics, Florence, Italy, July 2019. https://doi.org/10.18653/v1/P19-1466, https://www.aclweb.org/anthology/P19-1466

  17. Pineau, J., Sinha, K., Fried, G., Ke, R.N., Larochelle, H.: ICLR reproducibility challenge 2019. ReScience C 5(2), 5 (2019)

    Google Scholar 

  18. Pineau, J., et al.: Improving reproducibility in machine learning research (a report from the Neurips 2019 reproducibility program). arXiv preprint arXiv:2003.12206 (2020)

  19. Storer, T.: Bridging the chasm: a survey of software engineering practice in scientific programming. ACM Comput. Surv. (CSUR) 50(4), 1–32 (2017)

    Article  Google Scholar 

  20. Trisovic, A., Lau, M.K., Pasquier, T., Crosas, M.: A large-scale study on research code quality and execution. arXiv preprint arXiv:2103.12793 (2021)

  21. Voelter, M.: Fusing modeling and programming into language-oriented programming. In: Margaria, T., Steffen, B. (eds.) ISoLA 2018. LNCS, vol. 11244, pp. 309–339. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-03418-4_19

    Chapter  Google Scholar 

  22. Vorontsov, K., Iglovikov, V., Strijov, V., Ustuzhanin, A., Khritankov, A.: Roundtable: challenges in repeatable experiments and reproducible research in data science. Proc. MIPT (Trudy MFTI) 13(2), 100–108 (2021). https://mipt.ru/science/trudy/

  23. Wang, J., Tzu-Yang, K., Li, L., Zeller, A.: Assessing and restoring reproducibility of Jupyter notebooks, pp. 138–149 (2020)

    Google Scholar 

  24. Zaharia, M., et al.: Accelerating the machine learning lifecycle with MLflow. IEEE Data Eng. Bull. 41(4), 39–45 (2018)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Anton Khritankov .

Editor information

Editors and Affiliations

A Quality Requirements for Experiment Automation Software

A Quality Requirements for Experiment Automation Software

This is a preliminary list of quality requirements for experiment automation and reproducibility software. The requirements are based on series of in-depth interviews of data science researchers, heads of data science laboratories, academics, students and software developers in MIPT, Innopolis university and HSE.

Quality categories are given in accordance with ISO/IEC 25010 quality model standard.

Functionality

  • Ability to describe pipelines and configuration of ML experiments.

  • Run and reproduce experiments on demand and as part of a larger pipeline.

  • Prepare reports on the experiments including figures and papers.

Usability

  • Low entry barrier for data scientists who are Linux users.

  • Ability to learn gradually, easy to run first experiment

  • Technical and programming skill needed to use experiment automation tools should be lower than running experiments without it.

  • Users should be able to quickly determine the source of the errors.

Portability and Compatibility

  • Support common ML platforms (incl. Cloud Google Colab), OSes (Ubuntu 16, 18, 20, MacOS) and ML libraries (sklearn, pandas, pytorch, tensorflow...)

  • Support experiments in Python, Matlab

  • Run third-party ML tools with command-line interface

Maintainability

  • Open project, that is everyone should be able to participate and contribute.

  • Contributing to the project should not require understanding all the internal workings.

  • Should provide backward compatibility for experiment definitions.

Security/Reliability

  • Confidentiality of experiment data unless requested by user otherwise (e.g. publish results).

  • Keep experiment data secure/safe for a long time

Efficiency

  • Overhead is negligible for small and large experiment compared with the user code.

Satisfaction and Ease of Use.

  • Must be at least as rewarding/satisfactory/easy-to-use as Jupyter Notebook.

  • Interface should be similar to other tools familiar to data scientists.

Freedom from Risk

  • Using experiment automation software should not risk having their projects completed and results published.

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Khritankov, A., Pershin, N., Ukhov, N., Ukhov, A. (2022). MLDev: Data Science Experiment Automation and Reproducibility Software. In: Pozanenko, A., Stupnikov, S., Thalheim, B., Mendez, E., Kiselyova, N. (eds) Data Analytics and Management in Data Intensive Domains. DAMDID/RCDL 2021. Communications in Computer and Information Science, vol 1620. Springer, Cham. https://doi.org/10.1007/978-3-031-12285-9_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-12285-9_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-12284-2

  • Online ISBN: 978-3-031-12285-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics