Abstract
Verification techniques, as discussed in Chaps. 10 and 13, are to ascertain—with mathematical proof—whether a property holds on a mathematical model. The soundness and completeness required by the mathematical proof result in the scalability problem that verification algorithms can only work with either small models (e.g., the MILP-based method as in Sect. 11.2) or limited number of input dimensions (e.g., the reachability analysis as in Sect. 11.3). In practice, when working with real-world systems where the machine learning models are large in nature, other techniques have to be considered for the certification purpose. Similar to traditional software testing against software verification, neural network testing provides a certification methodology with a balance between completeness and efficiency. In established industries, e.g., avionics and automotive, the needs for software testing have been settled in various standards such as DO-178C and MISRA. However, due to the lack of logical structures and system specification, it is less straightforward on how to extend such standards to work with systems with neural network components. In the following, we discuss some existing neural network testing techniques. The readers are referred to the survey (Huang et al., Comput. Sci. Rev. 37:100270, 2020) for more discussion.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Niladri S Chatterji, Behnam Neyshabur, and Hanie Sedghi. The intriguing role of module criticality in the generalization of deep networks. ICLR2020, 2020.
Kelly Hayhurst, Dan Veerhusen, John Chilenski, and Leanna Rierson. A practical tutorial on modified condition/decision coverage. Technical report, NASA, 2001.
Wei Huang, Youcheng Sun, Xingyu Zhao, James Sharp, Wenjie Ruan, Jie Meng, and Xiaowei Huang. Coverage-guided testing for recurrent neural networks. IEEE Transactions on Reliability, pages 1–16, 2021.
Xiaowei Huang, Daniel Kroening, Wenjie Ruan, James Sharp, Youcheng Sun, Emese Thamo, Min Wu, and Xinping Yi. A survey of safety and trustworthiness of deep neural networks: Verification, testing, adversarial attack and defence, and interpretability. Computer Science Review, 37:100270, 2020.
Yue Jia and Mark Harman. An analysis and survey of the development of mutation testing. IEEE Transactions on Software Engineering, 37(5):649–678, 2011.
Gaojie Jin, Xinping Yi, Liang Zhang, Lijun Zhang, Sven Schewe, and Xiaowei Huang. How does weight correlation affect the generalisation ability of deep neural networks. In NeurIPS’20, 2020.
Augustus Odena and Ian Goodfellow. TensorFuzz: Debugging neural networks with coverage-guided fuzzing. arXiv preprint arXiv:1807.10875, 2018.
Kexin Pei, Yinzhi Cao, Junfeng Yang, and Suman Jana. DeepXplore: Automated whitebox testing of deep learning systems. In Proceedings of the 26th Symposium on Operating Systems Principles, pages 1–18. ACM, 2017.
RTCA. Do-178c, software considerations in airborne systems and equipment certification. 2011.
Ting Su, Ke Wu, Weikai Miao, Geguang Pu, Jifeng He, Yuting Chen, and Zhendong Su. A survey on data-flow testing. ACM Computing Surveys, 50(1):5:1–5:35, March 2017.
Youcheng Sun, Xiaowei Huang, and Daniel Kroening. Testing deep neural networks. CoRR, abs/1803.04792, 2018.
Youcheng Sun, Xiaowei Huang, Daniel Kroening, James Shap, Matthew Hill, and Rob Ashmore. Structural test coverage criteria for deep neural networks. 2018.
Youcheng Sun, Min Wu, Wenjie Ruan, Xiaowei Huang, Marta Kwiatkowska, and Daniel Kroening. Concolic testing for deep neural networks. In Automated Software Engineering (ASE), 33rd IEEE/ACM International Conference on, 2018.
Youcheng Sun, Min Wu, Wenjie Ruan, Xiaowei Huang, Marta Kwiatkowska, and Daniel Kroening. Deepconcolic: Testing and debugging deep neural networks. In 41st ACM/IEEE International Conference on Software Engineering (ICSE2019), 2018.
T.-W. Weng, H. Zhang, P.-Y. Chen, J. Yi, D. Su, Y. Gao, C.-J. Hsieh, and L. Daniel. Evaluating the Robustness of Neural Networks: An Extreme Value Theory Approach. In ICLR2018, 2018.
Matthew Wicker, Xiaowei Huang, and Marta Kwiatkowska. Feature-guided black-box safety testing of deep neural networks. In International Conference on Tools and Algorithms for the Construction and Analysis of Systems, pages 408–426. Springer, 2018.
Hong Zhu, Patrick AV Hall, and John HR May. Software unit test coverage and adequacy. ACM Computing Surveys, 29(4):366–427, 1997.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this chapter
Cite this chapter
Huang, X., Jin, G., Ruan, W. (2023). Testing Techniques. In: Machine Learning Safety. Artificial Intelligence: Foundations, Theory, and Algorithms. Springer, Singapore. https://doi.org/10.1007/978-981-19-6814-3_14
Download citation
DOI: https://doi.org/10.1007/978-981-19-6814-3_14
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-6813-6
Online ISBN: 978-981-19-6814-3
eBook Packages: Computer ScienceComputer Science (R0)