One of the greatest limitations of deep neural networks is the difficulty of interpreting what they learn from the data. Deep distilling addresses this problem by extracting human-comprehensible and executable code from observations.
References
Blazek, P. J., Venkatesh, K. & Lin, M. M. Nat. Comput. Sci. https://doi.org/10.1038/s43588-024-00593-9 (2024).
Blazek, P. J. & Lin, M. M. Nat. Comput. Sci. 1, 607–618 (2021).
Langley, P. Scientific Discovery: Computational Explorations of the Creative Processes (MIT Press, 1987).
Xu, F. et al. Explainable AI: A brief survey on history, research areas, approaches and challenges. In Natural Language Processing and Chinese Computing. NLPCC 2019 (eds Tang, J. et al.) Lecture Notes in Computer Science Vol 11839, 563–574 (Springer, 2019).
Karniadakis, G. E. et al. Nat. Rev. Phys. 3, 422–440 (2021).
Brunton, S. L., Proctor, J. L. & Kutz, J. N. Proc. Natl Acad. Sci. 113, 3932–3937 (2016).
McCulloch, W. S. & Pitts, W. Bull. Math. Biophys. 5, 115–133 (1943).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Competing interests
The author declares no competing interests.
Rights and permissions
About this article
Cite this article
Bakarji, J. Distilling data into code. Nat Comput Sci 4, 92–93 (2024). https://doi.org/10.1038/s43588-024-00598-4
Published:
Issue Date:
DOI: https://doi.org/10.1038/s43588-024-00598-4
- Springer Nature America, Inc.