Parallel Brain Simulator: A Multi-scale and Parallel Brain-Inspired Neural Network Modeling and Simulation Platform
- 603 Downloads
The brain is naturally a parallel and distributed system. Reverse engineering a cognitive brain is considered to be a grand challenge. In this paper, we present the parallel brain simulator (PBS), a parallel and distributed platform for modeling the cognitive brain at multiple scales. Inspired by large-scale graph computation, PBS can be considered as a universal parallel execution engine, which is aimed at reducing the complexity of distributed programming and providing an easy to use programmable platform for computational neuroscientists and artificial intelligence researchers for modeling and simulation of large-scale neural networks. As illustrative examples and validations, three brain-inspired neural networks which are built on PBS are introduced, including the 1:1 human hippocampus network, the 1:1 mouse whole-brain network and the CASIA brain simulator built for cognitive robotics. We deploy PBS on both commodity clusters and supercomputers, and a scalable performance is achieved. In addition, we provide evaluations on the scalability and performance of both lumped synapse-based simulation and non-lumped synapse-based simulation with different data-graph distribution methods to show the effectiveness and usability of the PBS platform.
KeywordsNeural network simulator Graph computation Parallel simulation Spiking neural network PBS
This study was funded by the Strategic Priority Research Program of the Chinese Academy of Sciences (XDB02060007) and Beijing Municipal Science and Technology Commission (Z151100000915070).
Compliance with Ethical Standards
Conflict of Interest
Xin Liu, Yi Zeng, Tielin Zhang, and Bo Xu declare that they have no conflict of interest.
All procedures followed were in accordance with the ethical standards of the responsible committee on human experimentation (institutional and national) and with the Helsinki Declaration of 1975, as revised in 2008 (5). Additional informed consent was obtained from all patients for which identifying information is included in this article.
Human and Animal Rights
This article does not contain any studies with human or animal subjects performed by the any of the authors.
- 1.Aleksander I, Morton H. An introduction to neural computing. London: Chapman and Hall; 1990.Google Scholar
- 2.Ananthanarayanan R, Esser S, Simon H, Modha D. The cat is out of the bag: cortical simulations with 109 neurons and 1013 synapses. In: Proceedings of the SC conference on high performance networking and computing, 2009. 2009;1: p. 1–12 IEEE.Google Scholar
- 6.Brette R, Rudolph M, Carnevale T, Hines M, Beeman D, Bower JM, Diesmann M, Morrison A, Goodman PH, Zirpe M, Natschlger T, Pecevski D, Ermentrout B, Djurfeldt M, Lansner A, Rochel O, Vieville T, Muller E, Davison AP, Boustani SE, Destexh A. Simulation of networks of spiking neurons: a review of tools and strategies. J Comp Neurosci. 2007;23(3):349–98.CrossRefGoogle Scholar
- 9.Dean J, Ghemawat S. MapReduce: simplified data processing on large clusters. In: Proceedings of the USENIX symposium on operating systems design and implementation, 2004. OSDI, 2004. 2004; IEEE.Google Scholar
- 11.Eliasmith C, Anderson C. Neural engineering: computation, representation, and dynamics in neurobiological systems. Cambridge: MIT press; 2004.Google Scholar
- 14.Gewaltig MO, Morrison A, Plesser HE. NEST by example: an introduction to the Neural Simulation Tool NEST. In: Le Novère N, editor. Computational systems neurobiology. Springer; 2012. p. 533–558.Google Scholar
- 15.Gonzalez J, Low Y, Gu H, Bickson D. Powergraph: distributed graph-parallel computation on natural graphs. In: Proceedings of the USENIX symposium on operating systems design and implementation, 2012. OSDI, 2012. 2012; IEEE.Google Scholar
- 20.Jaeger H. The “echo state” approach to analysing and training recurrent neural networks-with an erratum note. In: Technical report, German National Research Center for Information Technology. 2001.Google Scholar
- 21.Kunkel S, Schmidt M, Eppler JM, Plesser HE, Masumoto G, Igarashi J, Ishii S, Fukai T, Morrison A, Diesmann M, Helias M. Spiking network simulation code for petascale computers. Front Neuroinform. 2014;10(8):78.Google Scholar
- 22.Lansner A, Diesmann M. Virtues, pitfalls, and methodology of neuronal network modeling and simulations on supercomputers. In: Proceedings of the Chapter 10 in Nicolas Le Novre computational systems biology, Springer 2012.Google Scholar
- 23.Le NN. Computational systems neurobiology. New York: Springer; 2012.Google Scholar
- 24.Low Y, Gonzalez J, Kyrola A, Bickson D, Guestrin C, Hellerstein JM. Distributed graphlab: a framework for machine learning and data mining in the cloud. In: Proceedings of the very large database endowment, 2012. VLDB, 2012. 2012;5(8), IEEE.Google Scholar
- 25.Malewicz G, Austern MH, Bik AJC, Dehnert JC, Horn I, Leiser N, Czajkowski G. Pregel: a system for large-scale graph processing. In: ACM special interest group on management of data, 2010. SIGMOD, 2010. 2010; IEEE.Google Scholar
- 28.Oh SW, Harris JA, Ng L, Winslow B, Cain N, Mihalas S, Wang Q, Lau C, Kuan L, Henry AM, Mortrud MT, Ouellette B, Nguyen TN, Sorensen SA, Slaughterbeck CR, Wakeman W, Li Y, Feng D, Ho A, Nicholas E, Hirokawa KE, Bohn P, Joines KM, Peng H, Hawrylycz MJ, Phillips JW, Hohmann JG, Wohnoutka P, Gerfen CR, Koch C, Bernard A, Dang C, Jones AR, Zeng H. A mesoscale connectome of the mouse brain. Nature. 2014;508:207–14.CrossRefPubMedGoogle Scholar
- 33.Yonezawa A, Watanabe T, Yokokawa M, Sato M, Hirao K. Advanced institute for computational science (aics): Japanese national high-performance computing research institute and its 10-petaflops supercomputer. In: Proceedings of the SC conference on high performance networking and computing, 2011 2011.13, pp. 1–8 IEEE.Google Scholar
- 34.Zhang T, Zeng Y, Xu B. A computational effort towards the microscale mouse brain connectome from the mesoscale. Manuscript, 2015.Google Scholar