Improving multi-layer spiking neural networks by incorporating brain-inspired rules

Research Paper

Abstract

This paper introduces seven brain-inspired rules that are deeply rooted in the understanding of the brain to improve multi-layer spiking neural networks (SNNs). The dynamics of neurons, synapses, and plasticity models are considered to be major characteristics of information processing in brain neural networks. Hence, incorporating these models and rules to traditional SNNs is expected to improve their efficiency. The proposed SNN model can mainly be divided into three parts: the spike generation layer, the hidden layers, and the output layer. In the spike generation layer, non-temporary signals such as static images are converted into spikes by both local and global feature-converting methods. In the hidden layers, the rules of dynamic neurons, synapses, the proportion of different kinds of neurons, and various spike timing dependent plasticity (STDP) models are incorporated. In the output layer, the function of classification for excitatory neurons and winner take all (WTA) for inhibitory neurons are realized. MNIST dataset is used to validate the classification accuracy of the proposed neural network model. Experimental results show that higher accuracy will be achieved when more brain-inspired rules (with careful selection) are integrated into the learning procedure.

Keywords

brain-inspired rules spiking neural network plasticity classification task 

受脑启发的学习规则对深层脉冲神经网络性能的提升

创新点

本文总结和归纳了七条受脑启发的学习准则,并应用于改善脉冲神经网络。这些学习准则都来源于对生物脑的实验研究,并各自从不同的侧面反映了生物网络的学习特性,如神经元的动态分配、突触的自适应生长和消亡机制、不同的突触可塑性学习机制(如不同类型的时序依赖突触可塑性)、网络背景噪声对学习的调控机制、兴奋性和抑制性神经元的比例对学习的调节机制等。本文通过组合上述不同的受脑启发的规则,通过实验研究验证了:随着越来越多的、经过仔细选择的、受脑启发的规则的引入,深层脉冲神经网络能够得到越来越好的分类性能。

Keywords

受脑启发的学习规则 脉冲神经网络 可塑性 分类 
052201 

References

  1. 1.
    Hinton G, Osindero S, Teh Y W. A fast learning algorithm for deep belief nets. Neural Comput, 2006, 18: 1527–1554MathSciNetCrossRefMATHGoogle Scholar
  2. 2.
    He K, Zhang X, Ren S, et al. Delving deep into rectifiers: surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE International Conference on Computer Vision, Santiago, 2015. 1026–1034Google Scholar
  3. 3.
    Maass W. Networks of spiking neurons: the third generation of neural network models. Neural Networks, 1997, 10: 1659–1671CrossRefGoogle Scholar
  4. 4.
    Eliasmith C, Stewart T, Choo X, et al. A large-scale model of the functioning brain. Science, 2012, 338: 1202–1205CrossRefGoogle Scholar
  5. 5.
    Zenke F, Agnes E, Gerstner W. Diverse synaptic plasticity mechanisms orchestrated to form and retrieve memories in spiking neural networks. Nat Commun, 2015, 6: 6922CrossRefGoogle Scholar
  6. 6.
    Song H F, Yang G R, Wang X J. Training excitatory-inhibitory recurrent neural networks for cognitive tasks: a simple and flexible framework. PLoS Comput Biol, 2016, 12: e1004792CrossRefGoogle Scholar
  7. 7.
    Beyeler M, Oros N, Dutt N D, et al. A GPU-accelerated cortical neural network model for visually guided robot navigation. Neural Networks, 2015, 72: 75–87CrossRefGoogle Scholar
  8. 8.
    Maffei G, Santos-Pata D, Marcos E, et al. An embodied biologically constrained model of foraging: from classical and operant conditioning to adaptive real-world behavior in DAC-X. Neural Networks, 2015, 72: 88–108CrossRefGoogle Scholar
  9. 9.
    Wade J J, McDaid L J, Santos J, et al. Swat: a spiking neural network training algorithm for classification problems. IEEE Trans Neur Net, 2010, 21: 1817–1830CrossRefGoogle Scholar
  10. 10.
    Beyeler M, Dutt N D, Krichmar J L. Categorization and decision-making in a neurobiologically plausible spiking network using a STDP-like learning rule. Neural Networks, 2013, 48: 109–124CrossRefGoogle Scholar
  11. 11.
    Izhikevich E M. Simple model of spiking neurons. IEEE Trans Neur Net, 2003, 14: 1569–1572CrossRefGoogle Scholar
  12. 12.
    Iakymchuk T, Rosado-Munoz A, Guerrero-Martinez J F, et al. Simplified spiking neural network architecture and stdp learning algorithm applied to image classification. EURASIP J Image Vide Process, 2015, 2015: 1–11CrossRefGoogle Scholar
  13. 13.
    Ionescu M, Paun G, Yokomori T. Spiking neural P systems. Fund Inform, 2006, 71: 279–308MathSciNetMATHGoogle Scholar
  14. 14.
    Zhao Y, Liu X, Wang W. Spiking neural P systems with neuron division and dissolution. PLoS ONE, 2016, 11: e0162882CrossRefGoogle Scholar
  15. 15.
    Jia Y, Huang C, Darrell T. Beyond spatial pyramids: receptive field learning for pooled image features. In: Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, 2012. 3370–3377Google Scholar
  16. 16.
    Kravitz D J, Saleem K S, Baker C I, et al. The ventral visual pathway: an expanded neural framework for the processing of object quality. Trend Cogn Sci, 2013, 17: 26–49CrossRefGoogle Scholar
  17. 17.
    Häusser M. The hodgkin-huxley theory of the action potential. Nat Neurosci, 2000, 3: 1165CrossRefGoogle Scholar
  18. 18.
    Brette R, Gerstner W. Adaptive exponential integrate-and-fire model as an effective description of neuronal activity. J Neurophysiology, 2005, 94: 3637–3642CrossRefGoogle Scholar
  19. 19.
    Sharkey N E, Jackson S A. An internal report for connectionists. Comput Arc Integrat Neural Symb Proc, 1995, 292: 223–244Google Scholar
  20. 20.
    Destexhe A. Conductance-based integrate-and-fire models. Neural Comput, 1997, 9: 503–514CrossRefGoogle Scholar
  21. 21.
    Heeger D J, Ress D. What does fMRI tell us about neuronal activity. Nat Rev Neurosci, 2002, 3: 142–151CrossRefGoogle Scholar
  22. 22.
    Sokal R R, Rohlf F J. Biometry: the Principles and Practice of Statistics in Biological Research. New York: WH Freeman and Company, 1969MATHGoogle Scholar
  23. 23.
    Chrol-Cannon J, Jin Y. Computational modelling of neural plasticity for self-organization of neural networks. Biosystems, 2014, 125: 43–54CrossRefGoogle Scholar
  24. 24.
    Ghosh-Dastidar S, Adeli H. Spiking neural networks. Int J Neural Syst, 2009, 19: 295–308CrossRefGoogle Scholar
  25. 25.
    Seress L, Ribak C E. Direct commissural connections to the basket cells of the hippocampal dentate gyrus: anatomical evidence for feed-forward inhibition. J Neurocytology, 1984, 13: 215–225CrossRefGoogle Scholar
  26. 26.
    Lytton W W, Sejnowski T J. Inhibitory Interneurons Can Rapidly Phase-lock Neural Populations, Chapter for Induced Rhythms in the Brain. New York: Springer, 1992. 357–366Google Scholar
  27. 27.
    Waddington A, Appleby P A, De Kamps M, et al. Triphasic spike-timing-dependent plasticity organizes networks to produce robust sequences of neural activity. Front Comput Neurosci, 2012, 6: 88CrossRefGoogle Scholar
  28. 28.
    Song S, Miller K D, Abbott L F. Competitive hebbian learning through spike-timing-dependent synaptic plasticity. Nat Neurosci, 2000, 3: 919–926CrossRefGoogle Scholar
  29. 29.
    Clopath C, Büsing L, Vasilaki E, et al. Connectivity reflects coding: a model of voltage-based stdp with homeostasis. Nat Neurosci, 2010, 13: 344–352CrossRefGoogle Scholar
  30. 30.
    Rolls E T, Deco G. Computational Neuroscience of Vision. New York: Oxford University Press, 2002MATHGoogle Scholar
  31. 31.
    Schaul T, Bayer J, Wierstra D, et al. Pybrain. J Mach Learn Res, 2010, 11: 743–746Google Scholar

Copyright information

© Science China Press and Springer-Verlag Berlin Heidelberg 2017

Authors and Affiliations

  1. 1.Institute of AutomationChinese Academy of SciencesBeijingChina
  2. 2.Center for Excellence in Brain Science and Intelligence TechnologyChinese Academy of SciencesShanghaiChina

Personalised recommendations