Skip to main content

Short-and-Long-Term Impact of Initialization Functions in NeuroEvolution

  • Conference paper
  • First Online:
Intelligent Systems (BRACIS 2022)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 13653))

Included in the following conference series:

  • 668 Accesses

Abstract

Neural evolutionary computation has risen as a promising approach to propose neural network architectures without human interference. However, the often high computational cost of these approaches is a serious challenge for their application and research. In this work, we empirically analyse standard practices with Coevolution of Deep NeuroEvolution of Augmenting Topologies (CoDeepNEAT) and the effect that different initialization functions have when experiments are tuned for quick evolving networks on a small number of generations and small populations. We compare networks initialized with the He, Glorot, and Random initializations on different settings of population size, number of generations, training epochs, etc. Our results suggest that properly setting hyperparameters for short training sessions in each generation may be sufficient to produce competitive neural networks. We also observed that the He initialization, when associated with neural evolution, has a tendency to create architectures with multiple residual connections, while the Glorot initializer has the opposite effect.

We thank Coordination for the Improvement of Higher Education Personnel - CAPES/PROAP and Amazonas State Research Support Foundation - FAPEAM/POSGRAD 2021. This research was partially supported by CAPES via student support grant #88887.498437/2020-00.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Ma, Y., Xie, Y.: Evolutionary neural networks for deep learning: a review. Int. J. Mach. Learn. Cybern. (2022). https://doi.org/10.1007/s13042-022-01578-8

  2. Kumar, S. K.: On weight initialization in deep neural networks. In: arXiv preprint arXiv:1704.08863 (2017)

  3. Initializing neural networks. https://www.deeplearning.ai/ai-notes/initialization/. Accessed 12 June 2022

  4. Goodfellow, I.J., Bengio, Y., Courville, A.: Deep Learning, 1st edn. Cambridge (2016)

    Google Scholar 

  5. Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 249–256. JMLR Workshop and Conference Proceedings, Sardinia (2010)

    Google Scholar 

  6. He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1026–1034. IEEE, Santiago (2010)

    Google Scholar 

  7. Koutnik, J., Gomez, F., Schmidhuber, J.: Evolving neural networks in compressed weight space. In: Proceedings of the 12th Annual Conference on Genetic and Evolutionary Computation, pp. 619–626. ACM, Portland (2010)

    Google Scholar 

  8. Togelius, J., Gomez, F., Schmidhuber, J.: Learning what to ignore: memetic climbing in topology and weight space. In: Proceedings of the 2008 IEEE Congress on Evolutionary Computation (IEEE World Congress on Computational Intelligence), pp. 3274–3281. IEEE, Hong Kong (2008)

    Google Scholar 

  9. Okada, H., Wada, T., Yamashita, A., Matsue, T.: Interval-valued evolution strategy for evolving neural networks with interval weights and biases. In: Proceedings of the International Conference on Soft Computing and Intelligent Systems, and the 13th International Symposium on Advanced Intelligence Systems, pp. 2056–2060. IEEE, Kobe (2012)

    Google Scholar 

  10. Desell, T.: Accelerating the evolution of convolutional neural networks with node-level mutations and epigenetic weight initialization. In: Proceedings of the Genetic and Evolutionary Computation Conference Companion, pp. 157–158. IEEE, Kyoto (2018)

    Google Scholar 

  11. Lyu, Z., ElSaid, A., Karns, J., Mkaouer, M., Desell, T.: An experimental study of weight initialization and weight inheritance effects on neuroevolution. In: Proceedings of Applications of Evolutionary Computation: 24th International Conference. ACM, Seville (2021)

    Google Scholar 

  12. Miikkulainen, R., Liang, J., Meyerson, E., Rawal, A.: Evolving deep neural networks. In: Artificial Intelligence in the Age of Neural Networks and Brain Computing, pp. 293–312 (2019)

    Google Scholar 

  13. Papavasileiou, E., Cornelis, J., Jansen, B.: A systematic literature review of the successors of ‘NeuroEvolution of augmenting topologies’. Evol. Comput. 29, 1–73 (2020)

    Article  Google Scholar 

  14. Bohrer, J.S., Grisci, B.I., Dorn, M.: Neuroevolution of neural network architectures using CoDeepNEAT and Keras. In: arXiv preprint arXiv:2002.04634 (2020)

  15. Zhou, X., Li, X., Hu, K., Zhang, Y., Chen, Z., Gao, X.: ERV-Net: an efficient 3D residual neural network for brain tumor segmentation. Expert Syst. Appl. 170, 114566 (2021)

    Google Scholar 

  16. Dogan, S, et al.: Automated accurate fire detection system using ensemble pretrained residual network. Expert Syst. Appl. 203, 117407 (2022)

    Google Scholar 

  17. Hoorali, F., Khosravi, H., Moradi, B.: IRUNet for medical image segmentation. Expert Syst. Appl. 191, 116399 (2022)

    Google Scholar 

  18. Li, H., Xu, Z., Tyalor, G., Studer, C., Goldstein, T.: Visualizing the loss landscape of neural nets. In: Advances in Neural Information Processing Systems (2018)

    Google Scholar 

  19. Intuitive Explanation of Skip Connections in Deep Learning. https://theaisummer.com/skip-connections/. Accessed 12 June 2022

  20. Keras Documentation - Glorot Uniform. https://www.tensorflow.org/api_docs/python/tf/keras/initializers/GlorotUniform. Accessed 10 July 2022

  21. Keras Documentation - Glorot Normal. https://www.tensorflow.org/api_docs/python/tf/keras/initializers/GlorotNormal. Accessed 10 July 2022

  22. Keras Documentation - He Uniform. https://www.tensorflow.org/api_docs/python/tf/keras/initializers/HeUniform. Accessed 10 July 2022

  23. Keras Documentation - He Normal. https://www.tensorflow.org/api_docs/python/tf/keras/initializers/HeNormal. Accessed 10 July 2022

  24. Searching for activation functions. https://arxiv.org/abs/1710.05941. Accessed 12 June 2022

  25. Mish: A self regularized non-monotonic neural activation function. https://arxiv.org/abs/1908.08681. Accessed 12 June 2022

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Rafael Giusti .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Evangelista, L.G.C., Giusti, R. (2022). Short-and-Long-Term Impact of Initialization Functions in NeuroEvolution. In: Xavier-Junior, J.C., Rios, R.A. (eds) Intelligent Systems. BRACIS 2022. Lecture Notes in Computer Science(), vol 13653. Springer, Cham. https://doi.org/10.1007/978-3-031-21686-2_21

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-21686-2_21

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-21685-5

  • Online ISBN: 978-3-031-21686-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics