Abstract
The spatial component of input signals often carries information crucial to a neuronβs function, but models mapping synaptic inputs to the transmembrane potential can be computationally expensive. Existing reduced models of the neuron either merge compartments, thereby sacrificing the spatial specificity of inputs, or apply model reduction techniques that sacrifice the underlying electrophysiology of the model. We use Krylov subspace projection methods to construct reduced models of passive and quasi-active neurons that preserve both the spatial specificity of inputs and the electrophysiological interpretation as an RC and RLC circuit, respectively. Each reduced model accurately computes the potential at the spike initiation zone (SIZ) given a much smaller dimension and simulation time, as we show numerically and theoretically. The structure is preserved through the similarity in the circuit representations, for which we provide circuit diagrams and mathematical expressions for the circuit elements. Furthermore, the transformation from the full to the reduced system is straightforward and depends on intrinsic properties of the dendrite. As each reduced model is accurate and has a clear electrophysiological interpretation, the reduced models can be used not only to simulate morphologically accurate neurons but also to examine computations performed in dendrites.
Similar content being viewed by others
References
Bai, Z., & Skoogh, D. (2006). A projection method for model reduction of bilinear dynamical systems. Linear Algebra and its Applications, 415, 406β425.
Braun, M. (1975). Differential equations and their applications. New York: Springer.
Bush, P., & Sejnowski, T. (1993). Reduced compartmental models of neocortical pyramidal cells. Journal of Neuroscience Methods, 46, 159β166.
Destexhe, A., Mainen, Z., & Sejnowski, T. (1998). Kinetic models of synaptic transmission. In C. Koch, & I. Segev (Eds.), Methods in neuronal modeling (Chapter 1, pp. 1β25). Cambridge: MIT Press.
Freund, R. (2000). Krylov-subspace methods for reduced-order modeling in circuit simulation. Journal of Computational and Applied Mathematics, 123, 395β421.
Freund, R. (2011). The SPRIM algorithm for structure-preserving order reduction of general RCL circuits. In P. Benner, M. Hinze, & E. ter Maten (Eds.), Model reduction for circuit simulation (pp. 25β52). New York: Springer.
Gabbiani, F., & Cox, S. (2010). Mathematics for neuroscientists. Boston: Elsevier/Academic Press.
Golding, N., Kath, W., & Spruston, N. (2001). Dichotomy of action-potential backpropagation in CA1 pyramidal neuron dendrites. Journal of Neurophysiology, 86, 2998β3010.
Grimme, E. (1997). Krylov projection methods for model reduction. PhD thesis, University of Illinois at Urbana-Champaign, Urbana, Illinois.
Gu, C. (2011). QLMOR: A projection-based nonlinear model order reduction approach using quadratic-linear representation of nonlinear systems. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 30, 1307β1320.
Gugercin, S., Antoulas, A., & Beattie, C. (2008). \({{\cal H}_2}\) model reduction for large-scale linear dynamical systems. SIAM Journal on Matrix Analysis and Applications, 30, 609β638.
Hodgkin, A., & Huxley, A. (1952). A quantitative description of membrane current and its application to conduction and excitation in nerve. Journal of Physiology, 117, 500β544.
Jarsky, T., Roxin, A., Kath, W., & Spruston, N. (2005). Conditional dendritic spike propagation following distal synaptic activation of hippocampal CA1 pyramidal neurons. Nature Neuroscience, 8, 1667β1676.
Johnston, D., & Amaral, D. (1998). Hippocampus. In G. Shepherd (Ed.), The synaptic organization of the brain (Chapter 10, pp.Β 417β458). New York: Oxford University Press.
Kellems, A., Chaturantabut, S., Sorensen, D., & Cox, S. (2010). Morphologically accurate reduced order modeling of spiking neurons. Journal of Computational Neuroscience, 28, 477β494.
Kellems, A., Roos, D., Xiao, N., & Cox, S. (2009). Low-dimensional, morphologically accurate models of subthreshold membrane potential. Journal of Computational Neuroscience, 27, 161β176.
Koch, C. (1999). Biophysics of computation: Information processing in single neurons. New York: Oxford University Press.
Krapp, H., & Gabbiani, F. (2005). Spatial distribution of inputs and local receptive field properties of a wide-field, looming sensitive neuron. Journal of Neurophyiology, 93, 2240β2253.
Li, P., & Pileggi, L. (2005). Compact reduced-order modeling of weakly nonlinear analog and RF circuits. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 23, 184β203.
Li, R., & Bai, Z. (2005). Structure-preserving model reduction using a Krylov subspace projection formulation. Communications in Mathematical Sciences, 3, 179β199.
Lin, Y., Bao, L., & Wei, Y. (2009). Order reduction of bilinear MIMO dynamical systems using new block Krylov subspaces. Computers and Mathematics with Applications, 58, 1093β1102.
Mohler, R. (1991). Nonlinear systems: Applications to bilinear control. Englewood Cliffs: Prentice Hall.
Odabasioglu, A., Celik, M., & Pileggi, L. (1998). PRIMA: Passive reduced-order interconnect macromodeling algorithm. IEEE Transactions on Computer-aided Design of Integrated Circuits and Systems, 17, 645β654.
OβShea, M., & Rowell, C. (1976). The neuronal basis of a sensory analyser, the acridid movement detector sysem. II. Response decrement, convergence, and the nature of excitatory afferents to the fan-like dendrites of the LGMD. Journal of Experimental Biology, 65, 289β308.
Phillips, J. (2000). Projection frameworks for model reduction of weakly nonlinear systems. In Proceedings of DAC 2000 (pp. 184β189).
Phillips, J. (2003). Projection-based approaches for model reduction of weakly nonlinear, time-varying systems. IEEE Transations on Computer-Aided Design Integrated Circuits Systems, 22, 171β187.
Pinsky, P., & Rinzel, J. (1994). Intrinsic and network rhythmogenesis in a reduced Traub model for CA3 neurons. Journal of Computational Neuroscience, 1, 39β60.
Poznanski, R. (1991). A generalized tapering equivalent cable model for dendritic neurons. Bulletin of Mathematical Biology, 53, 457β467.
Rall, W. (1959). Branching dendritic trees and motoneuron membrane resistivity. Experimental Neurology, 1, 491β527.
Roychowdhury, J. (1999). Reduced-order modeling of time-varying systems. IEEE Transactions on Circuits and Systems. II: Analog and Digital Signal Processing, 46, 1273β1288.
Rugh, W. (1981). Nonlinear system theory. Baltimore: Johns Hopkins University Press.
Schierwagen, A. (1989). A non-uniform equivalent cable model of membrane voltage changes in a passive dendritic tree. Frontiers in Neuroscience, 1, 19β42.
Spruston, N. (2008). Pyramidal neurons: Dendritic structure and synaptic integration. Nature Reviews Neuroscience, 9, 206β221.
Traub, R., Wong, K., Miles, R., & Michelson, H. (1991). A model of a CA3 hippocampal pyramidal neuron incorporating voltage-clamp data on intrinsic conductances. Journal of Neurophysiology, 66, 635β650.
Trefethen, L., & Bau, D. (1997). Numerical linear algebra. Philadelphia: Society for Industrial and Applied Mathematics.
Villemagne, C., & Skelton, R. (1987). Model reduction using a projection formulation. International Journal of Control, 46, 2141β2169.
Yan, B., & Li, P. (2011). Reduced order modeling of passive and quasi-active dendrites for nervous system simulation. Journal of Computational Neuroscience, 31, 247β271.
Acknowledgements
This work is supported by NSF grant DMS-0739420 and by a training fellowship from the Keck Center for Interdisciplinary Bioscience Training of the Gulf Coast Consortia (NIBIB Grant No. 1T32EB006350-01A1).
Author information
Authors and Affiliations
Corresponding author
Additional information
Action Editor: Brent Doiron
Appendices
Appendix A: Construction of the Volterra series
The Volterra series is a useful representation of nonlinear systems, but thorough accounts of its derivation and convergence are sparse in the literature. In this appendix we derive the Volterra series for the passive cable given synaptic input and establish its convergence by relating it to the well-studied Picard iterates. For an alternative approach to the Volterra series, see Section 5.3 of Mohler (1991).
Consider the full model for the passive cable driven by monosynaptic input,
where by Eq.Β (33), Aβ=βββC βββ1 G, \(b = C^{-1}Ee_p,\) and Bβ=βββC βββ1 N. Define f(t)ββ‘βeβββAt v(t). Then,
If g(t) were transient, the product f(t)g(t) would be relatively small, motivating the initial guess
Equation (51) is solved by the fixed point problem
where
Each iterate is used to generate the next guess for f via
This method is known as Picard iterations or the method of successive approximations, and as both f k and K are continuous, f k(t) converges to f(t), as shown in Section 1.10 of Braun (1975). We construct the Volterra series by defining v 1(t)ββ‘βeAt f 1(t) and v k(t)ββ‘βeAt (f kβββf kβββ1)(t) for kβ>β1. Its convergence follows via
Unwrapping each term reveals
implying that (v 1)β²(t)β=βAv 1(t)β+βbg(t), or
Similarly, for kββ₯β2,
implying that (v k)β²(t)β=βAv k(t)β+βBg(t)v kβββ1(t), or
The series easily generalizes for polysynaptic input by replacing the input of Eq.Β (53) with \(\sum_{j=1}^m E_{p_j} e_{p_j} g_{p_j}(t)\) and of Eq.Β (55) with \(-\sum_{j=1}^m g_{p_j}(t) {N^{(p_j)}} v^{k-1}(t)\).
Appendix B: Transfer functions for the Volterra series
When the cable is driven by current injections, we construct a reduced system to match the leading moments of the transfer functions for the full and reduced systems. The same procedure can lead to an appropriate reducer for the cable driven by synaptic conductances by considering the transfer functions for each term in the Volterra series, where each transfer function maps the synaptic conductance, g, to \(y^k = {e_{\rm SIZ}}^T v^k\). In this appendix we derive the transfer functions for each Volterra term given monosynaptic input. The transfer functions can be generalized for polysynaptic input.
Assume g(tβ<β0)β=β0 and define the kernel
where Aβ=βββC βββ1 G and \(b = C^{-1}Ee_p\). EquationΒ (52) can be written as
As expected, v 1 depends linearly on g, and by the convolution theorem,
and \({\cal L}\) denotes the Laplace transform. The transfer function \({\widetilde{H}}_1\) is then given by
Since the passive system is stable, \({\widetilde{H}}_1\) is well-defined if Re(s)ββ₯β0 and provides the mapping
We next strive to write the second Volterra term as a convolution, which would allow us to easily compute its transfer function. By Eq.Β (54), v 2 depends on the product of v 1 and g, both of which depend on g. We thus make the educated guess that v 2 can be written as
and solve for the kernel h 2. As Eq.Β (58) is not quite a convolution, we pause to show the advantage of its form. Define \({\overline g}(t_1,t_2) \equiv g(t_1)g(t_2)\), and define
Then, \(v^2(t) = {\overline v}(t,t)\), and by the convolution theorem,
where \({\widetilde{H}}_2 = {\cal L} h_2\). Therefore,
and the two-dimensional transfer function \({\widetilde{H}}_2\) does indeed map the input g to the output v 2 in the frequency domain. We now return to the computation of the kernel h 2 and corresponding transfer function. By Eq.Β (54),
where Bβ=βββC βββ1 N. By Eq.Β (56),
given the change of variables, Ο 2β=βΟ 1β+βs. EquationΒ (58) is then obtained by defining the kernel
The transfer function, \({\widetilde{H}}_2 = {\cal L} h_2\), is given by
where
Hence, if Re(s 1β+βs 2)ββ₯β0, then
Finally, \(H_2 \equiv e_{\rm SIZ}^T {\widetilde{H}}_2\) combined with Eq.Β (59) leads to
To simplify notation, the transfer function can be written in its regular form, defined in Section 2.3 of Rugh (1981) such that \(H_2(s_1,s_2) = H_2^{{\text{reg}}}(s_1+s_2,s_2),\) or
In a similar manner, one can iteratively compute the kernels and corresponding transfer functions for each Volterra term. As each Volterra term has an increasingly nonlinear dependence on g, it can be written as
where \({\overline g}(t_1,\cdots,t_k) \equiv \prod_{j=1}^k g(t_j)\). Given \({\overline v}(t_1,\cdots,t_k) \equiv\) \( (h_k \star {\overline g})(t_1,\cdots,t_k),\)
and \(v^k(t) = {\overline v}(t,\cdots,t)\). The transfer function for the output y k in its regular form is then given by
Rights and permissions
About this article
Cite this article
Hedrick, K.R., Cox, S.J. Structure-preserving model reduction of passive and quasi-active neurons. J Comput Neurosci 34, 1β26 (2013). https://doi.org/10.1007/s10827-012-0403-y
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10827-012-0403-y