Abstract
A particular form of network used in brain imaging, the neural network, is the topic of this chapter. Properties of neural nets are presented together with a detailed description of the structure of neurons and artificial neurons that comprise these networks. A neural net is a conceptual model based upon the human brain. It has an observed input layer that connects to a middle, hidden layer consisting of an unobserved number of synapses and neurons and leads to an observed output layer. There are different functions for measuring the degree or presence of activity between the neurons. The model can be trained in various ways, but usually a design and test set approach is used where the model is given an input and allowed to develop unsupervised and eventually compared to the output (which is often a group) to assess fit using various measures such as mean absolute error and root mean square error. Tips are given on how to Interpret these fit indices. The learning rate, epoch, and batch, which control the amount of learning input data used, are described. A worked example in R is at the end of the chapter looking at stress and anxiety levels in a child and its mother being used to predict the presence, or absence, of hyperactivity.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Agatonovic-Kustrin, S., & Beresford, R. (2000). Basic concepts of artificial neural network (ANN) modeling and its application in pharmaceutical research. Journal of Pharmaceutical and Biomedical Analysis, 22(5), 717–727.
Anđelković, B. Research and development of new methods for the calculation of clamping circuits using neural networks and fuzzy logic. https://view.officeapps.live.com/op/view.aspx?src=http%3A%2F%2Fwww.hm.co.rs%2Fstatistika%2FPrincipi_fazi_i_neuro_modeliranja1.doc&wdOrigin=BROWSELINK. Last access 10.01.2022.
Bailey, J. D., Baker, J. C., Rzeszutek, M. J., & Lanovaz, M. J. (2021). Machine learning for supplementing behavioral assessment. Perspectives on Behavior Science, 1–15.
Bengio, Y. (2012). Practical recommendations for gradient-based training of deep architectures. Neural Networks: Tricks of the Trade: Second Edition, 437–478.
Bonaccorso, G. (2017). Machine learning algorithms. Packt Publishing Ltd.
Brownlee, J. (2016). Master Machine Learning Algorithms: Discover How They Work and Implement Them From Scratch. (2016). United States: Machine Learning Mastery.
Camuñas-Mesa, L. A., Linares-Barranco, B., & Serrano-Gotarredona, T. (2019). Neuromorphic spiking neural networks and their memristor-CMOS hardware implementations. Materials, 12(17), 2745.
Dangeti, P. (2017). Statistics for machine learning. Packt Publishing Ltd.
Goodfellow, I., Bengio, Y., & Courville, A. (2016). Deep learning. MIT Press.
Harper, G. Viability of Artificial Neural Networks in Mobile Healthcare. https://www.theseus.fi/bitstream/handle/10024/28868/Harper_Gavin.pdf?sequence=1. Last access 10.01.2022.
Heaton, J. (2015). Artificial intelligence for humans, volume 3: Deep learning and neural networks. Heaton Research, Inc.
Koutsoukas, A., Monaghan, K. J., Li, X., & Huan, J. (2017). Deep-learning: Investigating deep neural networks hyper-parameters and comparison of performance to shallow methods for modeling bioactivity data. Journal of Cheminformatics, 9(1), 1–13.
Lawrence, J. (1993). Introduction to neural networks. California Scientific Software.
Ma, S., & Ji, C. (1999). Performance and efficiency: Recent advances in supervised learning. Proceedings of the IEEE, 87(9), 1519–1535.
Marques, D. B., Barradas Filho, A. O., Romariz, A. R., Viegas, I. M., Luz, D. A., Barros Filho, A. K., et al. (2014). Recent developments on statistical and neural network tools focusing on biodiesel quality. International Journal of Computer Science and Application, 3(3), 97–110.
McCulloch, W. S., & Pitts, W. (1943). A logical calculus of the ideas immanent in nervous activity. The Bulletin of Mathematical Biophysics, 5, 115–133.
Minsky, M., & Papert, S. (1969). Perceptrons: An essay in computational geometry. MIT Press.
Russell, S. J. (2010). Artificial intelligence a modern approach. Pearson Education, Inc.
Shepherd, G. M., & Koch, C. (1990). Introduction to synaptic circuits. In G. M. Shepherd (Ed.), The synaptic organization of the brain. Oxfrod University Press.
Yang, L., & Shami, A. (2020). On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing, 415, 295–316.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Farahani, H., Blagojević, M., Azadfallah, P., Watson, P., Esrafilian, F., Saljoughi, S. (2023). Deep Neural Network. In: An Introduction to Artificial Psychology. Springer, Cham. https://doi.org/10.1007/978-3-031-31172-7_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-31172-7_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-31171-0
Online ISBN: 978-3-031-31172-7
eBook Packages: Behavioral Science and PsychologyBehavioral Science and Psychology (R0)