## Abstract

Neuroscience incorporates knowledge from a range of scales, from single molecules to brain wide neural networks. Modeling is a valuable tool in understanding processes at a single scale or the interactions between two adjacent scales and researchers use a variety of different software tools in the model building and analysis process. Here we focus on the scale of biochemical pathways, which is one of the main objects of study in systems biology. While systems biology is among the more standardized fields, conversion between different model formats and interoperability between various tools is still somewhat problematic. To offer our take on tackling these shortcomings and by keeping in mind the FAIR (findability, accessibility, interoperability, reusability) data principles, we have developed a workflow for building and analyzing biochemical pathway models, using pre-existing tools that could be utilized for the storage and refinement of models in all phases of development. We have chosen the SBtab format which allows the storage of biochemical models and associated data in a single file and provides a human readable set of syntax rules. Next, we implemented custom-made MATLABĀ® scripts to perform parameter estimation and global sensitivity analysis used in model refinement. Additionally, we have developed a web-based application for biochemical models that allows simulations with either a network free solver or stochastic solvers and incorporating geometry. Finally, we illustrate convertibility and use of a biochemical model in a biophysically detailed single neuron model by running multiscale simulations in NEURON. Using this workflow, we can simulate the same model in three different simulators, with a smooth conversion between the different model formats, enhancing the characterization of different aspects of the model.

## Introduction

Computational systems biology is a data-driven field concerned with building models of biological systems. Methods from systems biology have proven valuable in neuroscience, particularly when studying the composition of synapses, molecular mechanisms of plasticity, learning and various other neuronal processes (Bhalla & Iyengar, 1999; Hellgren Kotaleski & Blackwell, 2010; Li et al., 2012). A wide variety of different software and toolboxes, each with their own strengths and weaknesses, are available within the field. This diversity, however, can obstruct model reuse as interoperability between the different software packages and the convertibility between various file types is only solved in part. Interoperability can either mean that the model built in one simulator can be run in another or that both simulators interoperate at run-time either at the same or different scales (Cannon et al., 2007). The former is addressed by standardizing model descriptions for which notable examples in systems biology include the standard machine-readable model formats, such as XML-based SBML (Systems Biology Markup Language; Hucka et al., 2003) and CellML (Hedley et al., 2001), and human readable format, such as SBtab (Lubitz et al., 2016). An analogous model description language for neurons and networks is the NeuroML (Neural Open Markup Language; Gleeson et al. 2010,Ā 2012).

We start by providing examples of available systems biology tools, for model building, parameter estimation and model analysis. We then proceed to describe our approach in developing a modular workflow to address some of the interoperability issues and present simulation results of an example use case in various simulators and frameworks, with further examples provided in Supplementary Materials. Our workflow starts with a human-readable representation of the model that is easily accessible to everyone and proceeds through various conversions into different simulation environments: MATLABĀ®, COPASI, NEURON, and STEPS (STochastic Engine for Pathway Simulation). Specifically, we will describe the conversion tools we created for this purpose.

## Examples of Software and Toolboxes used in Systems Biology

No software package is perfectly suited for every task, some have programmable interfaces with scripting languages, like the MATLABĀ® SimBiologyĀ® toolbox, some focus on providing a fixed array of functions that can be run via graphical user interfaces, like COPASI, although it now offers a Python toolbox for scripting (Welsh et al., 2018). Most toolboxes and software packages offer a mixture of the two approaches: fine-grained programmable interface as well as fixed high-level operations. At the extremes of this spectrum are powerful but inflexible high-level software on one side, and complex, hard to learn but very flexible libraries or toolboxes with an API (application programming interface) on the other. Some examples of general modeling toolboxes in MATLABĀ® are the SBPOP/SBToolbox2 (Schmidt & Jirstrand, 2006), and the PottersWheel Toolbox (Maiwald & Timmer, 2008). For Bayesian parameter estimation, there are the MCMCSTAT toolbox (Haario et al., 2006) in MATLABĀ®, as well as pyABC (Klinger et al., 2018) and pyPESTO (SchĆ¤lte et al., 2020) in Python, and the standalone *Markov Chain Monte Carlo* (MCMC) software GNU MCsim that allows to estimate the posterior distribution by sampling a high-dimensional probability distribution (Bois, 2009). For global sensitivity analysis, there is the Uncertainpy Python toolbox (TennĆøe et al.,Ā 2018). For simulations in neuroscience, some notable examples include NEURON (Hines & Carnevale, 1997) and STEPS (Hepburn et al., 2012). Both are used for simulations of neurons and can include reactionādiffusion systems and electrophysiology.

These software packages do not all use the same model definition formats. Most have some compatibility with SBML, others use their own formats (e.g. NEURON uses MOD files). In some cases, an SBML file exported from one of these packages cannot be imported into another package without errors; so manual intervention may be required.^{Footnote 1} Given an SBML file, a common task is to translate the contents into code that can be used in model simulations, for ordinary differential equations this is the right-hand side vector field function. There are several tools that facilitate the conversion between formats, e.g. the SBFC (The Systems Biology Format converter; Rodriguez et al., 2016) as well as the more general VFGEN (A Vector Field File Generator; Weckesser, 2008).

All toolboxes and software packages have great strengths and short-comings, and each programming language has different sets of (freely) available libraries which makes the development (or use) of numerical methods more or less feasible than in another language. One such example is the R package VineCopula for parameter dependency modeling, which has recently been used for modeling probability densities (in parameter spaces) between two MCMC runs (Eriksson et al., 2019). It infers the probability density function from a large enough sample, another method to do this is *kernel density estimation*.^{Footnote 2} The VineCopula package implements a much more advanced and robust method of density estimation based on vines and copulas and performs well in high dimensional cases. This package is not easily replaced in many other languages. The Julia language, on the other hand, has a far richer set of differential equation solvers than R or MATLABĀ® and comes with very efficient forward sensitivity analysis methods.

Each researcher must therefore make decisions that result in the best compromise for them. If a researcher is familiar with a given set of programming/scripting languages it is probably not reasonable to expect them to be able to collaborate with other groups in languages they do not know. For this reason, it is our firm opinion that the conversion of models between different formats is a very important task, and it is equally important to use formats that people can pick-up easily. To make format conversion flexible intermediate files are a great benefit which leads to a modular approach with possible validation between modules.

We also have to consider the FAIR data principlesāthe findability, accessibility, interoperability and reusability of data and associated infrastructure (Wilkinson et al., 2016). Here, we would like to address the interoperability principle by having developed a workflow for building biochemical pathway models using existing tools and custom-made, short, freely available scripts for the storage and refinement of models in all phases of development, ensuring interchangeability with other formats and toolkits at every step in the pipeline using standardized intermediate files (Fig.Ā 1 and Table 1).

FigureĀ 1 illustrates the relationship between the tools we used in this workflow. We created a use case based on a previously developed model for others to reuse and modify, and we use it to demonstrate how the depicted parts operate. In addition, we made two other use cases available for testing other parts of the workflow. More extensive testing was done with a model of the mitogen-activated protein kinase (MAPK) cascade provided by the FindSim workflow (Viswan et al., 2018) and the results can be found in the Supplementary Materials.

### The Workflow

While the standard model storage format in systems biology is SBML, it has some drawbacks: it does not lend itself to manual editing, the math in an SBML file is difficult to read and write manually, xml parsing is a difficult task that cannot be undertaken lightly by the novice programmer, species entries do not have a concentration unit attribute, time is handled very differently from any other model variable, etc. None of these issues are an error of course, but they are inconvenient for the inexperienced user. Therefore, this workflow is centered around building an easy-to-use infrastructure with models and data expressed in a spreadsheet-based storage format called SBtab (Lubitz et al., 2016). We chose SBtab as the primary modeling source file because it is human-readable and writable, it can contain both the model and the data, and because it is easy to write parsing scripts for it, such as a converter from SBtab to SBML using the libSBML^{Footnote 3} interface in R. This ease of convertibility is used in the second focus of the workflow, convertibility between SBtab and other common formats and simulation software, since in systems biology and in any other computational sciences, the lack of compatibility between different tools and formats can often pose problems. A partially working conversion tool between SBtab and SBML had already been developed by the SBtab team. However, it can currently only read one table at a time and does not produce any functional SBML files with our model example. To combat these shortcomings, we wrote scripts to convert the SBtab into SBML, either using the R language or MATLABĀ®, and validating it successfully in COPASI, STEPS and NEURON.

The bulk of our workflow is available as MATLABĀ® code, particularly the parameter estimation tools and functions for global sensitivity analysis. Sensitivity analysis can be used to determine the importance of different parameters in regulating different outputs. Local sensitivity analysis is based on partial derivatives and investigates the behavior of the output when parameters are perturbed in close vicinity to a specific point in parameter space. Global sensitivity analysis, on the other hand, is based on statistical approaches and has a much broader range. Global sensitivity analysis is more relevant for models that have a large uncertainty in their parameter estimates which is common for systems biology models where many of the parameters have not been precisely measured and the data are sparse.

A common approach within biochemical modeling is to use deterministic simulations and ordinary differential equations (ODE) that follow the law of mass action as it is computationally efficient and provides accurate (compared to averaged stochastic simulations) results for sufficiently large well-mixed biological systems. However, this approach has several restrictions in the case of neuronal biochemical cascades. First, such cascades are always subject to stochastic noise, which can be especially relevant in a compartment as small as a dendritic spine where the copy number of key molecules are small enough that the effect of randomness becomes significant (Bhalla, 2004). For precise simulation of stochasticity in reaction networks several stochastic solvers are available, e.g. Gillespieās Stochastic Simulation Algorithm (SSA) (Gillespie, 1976) and explicit and implicit tau-leaping algorithms (Gillespie, 2001). Second, the number of possible states of many biochemical cascades grow exponentially with the number of simulated molecule types, such that it becomes difficult to represent all these states in the model. In this case, for efficient simulation the reactions in the model could be represented and simulated in a network free form using rule-based modeling approaches (Chylek et al., 2015). Third, many biochemical networks are spatially distributed, this requires simulation of molecule diffusion (Hepburn et al., 2012). To tackle these problems, we developed the subcellular simulation setup application, a web-based software component for model development. It allows the extension and validation of deterministic chemical reaction network-based models by simulating them with stochastic solvers for reactionādiffusion systems (STEPS, Hepburn et al., 2012) and network free solvers (NFsim, Sneddon et al., 2011).

Regardless of the modeling approach (rule based or reaction network based ODE), the solvers yield a time-curve solution: x(t;Īø). This means that our workflow can only accept data that can be represented by one or several numerical solutions of this type. Time Series data and Dose Response curves can both be mapped to trajectory solutions. Dose Response curves are mapped point-by-point to solutions under varying input settings (doses). We have not used direct solvers to obtain steady states or other limit sets.

Although the workflow is in principle applicable to any biochemical pathway model our emphasis is on modeling biochemical signaling in neurons. Therefore, the last challenge we want to address is an important concept in the interoperability domain of computational neuroscience called multiscale modeling which concerns the integration of subcellular models into electrical models of single cells or in neuronal microcircuits. This can be achieved either by run-time interoperability between two simulators of different systems (Djurfeldt et al., 2010) or by expanding the capabilities of a single simulation platform as has been done with the NEURON software (McDougal et al., 2013). With this purpose in mind, we have written a conversion function from SBtab to the MOD format which is used by NEURON. As such, the inputs and the outputs of a biochemical cascade can be linked to any of the biophysiological measures of the electrical neuron model.

### Use Case

As a primary use case to illustrate the workflow we have chosen a previously developed pathway model of the emergence of eligibility trace observed in reinforcement learning in striatal direct pathway medium spiny neurons (MSN) that express the D1 receptor (Nair et al., 2016). Additional use cases are considered in the Supplementary Materials.

In this model, a synapse that receives excitatory input, which leads to an increase in calcium concentration, is potentiated only when the signal is followed by a reinforcing dopamine input. FigureĀ 2A represents a simplified model scheme illustrating these two signaling cascades, one starts with calcium as the input and the other one with dopamine. In simulation experiments the inputs are represented as a calcium train and a dopamine transient (Fig.Ā 3A). Calcium input refers to a burst of 10 spikes at 10Ā Hz reaching 5Ā Ī¼M. Dopamine input is represented by a single transient of 1.5Ā Ī¼M. The first cascade (species in blue) features the calcium-dependent activation of Ca^{2+}/calmodulin-dependent protein kinase II (CaMKII) and the subsequent phosphorylation of a generic CaMKII substrate which serves as a proxy for long term potentiation (LTP) and is the main output of the model. The second cascade (species in red) represents a G-protein dependent cascade following the dopamine input and resulting in the phosphorylation of the striatal dopamine- and cAMP-regulated phosphoprotein, 32Ā kDa (DARPP-32) that turns into an inhibitor of protein phosphatase 1 (PP1) which can dephosphorylate both CaMKII and its substrate. The phosphorylation of the substrate is maximal when two constraints are met. First, the time window between the calcium and dopamine inputs has to be short, corresponding to the input-interval constraint which is mediated by DARPP-32 via PP1 inhibition. Second, intracellular calcium elevation has to be followed by the dopamine input, corresponding to the input-order constraint that is mediated by another phosphoprotein, the cyclic AMP-regulated phosphoprotein, 21Ā kDa (ARPP-21), thanks to its ability to sequester calcium/calmodulin if dopamine arrives first (Fig.Ā 3D).

In the originally published model, CaMKII is autophosphorylated in two compartments, both the cytosol and the post synaptic density (PSD), with a custom-written MATLABĀ® rate function that was calculated based on the probability of two neighboring subunits being fully activated as described in Li et al. (2012). To make it possible to run the model in different software we replace the rate equation of autophosphorylation with a similar set of reactions in both compartments so that the model would only contain bimolecular reactions. The reactions represent a simplified version of the autophosphorylation reactions in Pepke et al. (2010), where in our case only the fully activated CaMKII can be phosphorylated. The same set of reactions is used in both compartments and the schematics is available in Fig.Ā 2B along with the required six new parameters. We used our parameter estimation script to find parameter values and bounds that preserved the qualitative behavior of the model. In this primary use-case, we used simulated data (real data is used in the supplementary materials model) from the original model with different timings of the dopamine input relative to the calcium input (Fig.Ā 2C) to obtain a comprehensive picture of its behavior which we want the updated model to reproduce.

### SBtab

As described above we have chosen SBtab as the format at the root of our workflow for the storage of the model and associated data. In this section we illustrate how we use this format. More information, documentation and examples are available from the authors of SBtab.^{Footnote 4} SBtab allows the storage of biochemical models and associated data in a single file and provides a set of syntax rules and conventions to structure data in a tabulated form making it easy to write, modify and share. To ensure interoperability, SBtab provides an online tool to convert the models into the SBML format.^{Footnote 5} SBtab is suitable for storing data that comes in spreadsheet or table formats, e.g. concentration time series or dose response curves, but it is likely that *any* data format that can be reasonably stored as a table will work well in SBtab. The SBtab file is intended to be updated manually during the process of model building. Additional instructions on how to make SBtab files work well within our toolchain can be found in the Subcellular Workflow documentation.^{Footnote 6} Some of the columns and sheets that we use should be considered as extensions to the format and are discussed in the documentation. SBtab is easy to parse so adjustments to parsers can be made quickly.

The SBtab file should include separate sheets for compartments, compounds, reactions, assignment expressions, parameters, inputs, outputs, and experiments (as well as data tables). We illustrate the functionalities of SBtab here with the use case. The use case model has 99 compounds, 138 reactions and 227 parameters. An example of the SBtab reaction table can be found in Table 2.

One of our goals with this study was to reproduce the original model behavior after replacing a single module inside the model to convert it to bimolecular reactions only. The data we used therefore represents the simulated time series (20Ā s) of the concentrations of four selected species in response to different input combinations using the original model. Each individual data sheet (named E0-E9, Fig.Ā 2C) in SBtab represents the outputs of one experiment. A separate sheet called Experiments allows to define the input parameters differently for each experimental setup. By setting the initial concentrations of the unused species to zero the data could be mapped to a specific sub-module of the model (the remaining species). In this case the initial conditions are the same for all experiments. The Experiments table can also support annotations relevant to each dataset. We used nine different timings (corresponding to E0-E8) between the calcium and the dopamine signal starting with a dopamine signal preceding calcium by four seconds and finishing with dopamine following calcium after four seconds as this corresponds to the time frame originally used in model development (Ītā=ā{-4,-3,-2,-1,0,1,2,3,4}). Additionally, we used simulations with calcium as the only input (E9) (Fig.Ā 2). The time series of the input species are in a separate sheet following each experiment sheet. An example of how experimental data is stored can be found in Table 3.

### Model Pre-Processing Tools

Model building entails frequent changes to the model structure by adding new species, reactions, and parameters. This can result in the emergence or disappearance of Wegscheider cyclicity conditions that refer to the relationships between reaction rate coefficients arising from conditions of thermodynamic equilibrium (Wegscheider,Ā 1901; Vlad & Ross,Ā 1994). Identified thermodynamic constraints show parameter dependencies that follow from physical laws and can reduce the number of independent parameters. These conditions are frequently difficult to determine by human inspection, especially for large systems. Similarly, identifying conserved moieties, like conserved total concentration of a protein, allows the reduction of the ODE model size, which leads to increased performance. In order to address these model pre-processing needs our toolkit includes scripts in MATLAB^{Ā®}/GNU Octave that use the stoichiometric matrix of the reaction network as an input to determine the thermodynamic constraints as described in Vlad and Ross (1994), and conservation laws (Tables 1 and 4). These diagnostic tools output any identified constraints that, if needed, are to be implemented manually before the parameter estimation step. It should be noted that such constraints need to be re-examined after each addition of new reactions as the structure of the model might change and make previously true constraints invalid. This is true for all major changes to the model.

### MATLABĀ® Tools

The bulk of our workflow is developed in MATLABĀ® as it provides an easy-to-use biochemical modeling application with a graphical user interface called SimBiologyĀ® along with a wide range of toolboxes for mathematical analysis. The workflow is divided into import and analysis scripts. We have written software for three types of analysis: diagnostics tools which are used to run the model and visualize how the model fits the data, parameter estimation, and global sensitivity analysis. To ensure an easy and user-friendly usage, all operations are controlled by a single settings file where all specification options needing user input are represented as modifiable variables. An example settings file of the use case model along with instructive comments can be found in the model GitHub repository.^{Footnote 7} Only this settings file and the model in the SBtab format are needed as input from the user to run all our MATLABĀ® scripts. After running the analysis, the inputs and the data points along with the simulated model fits are plotted and the results are stored inside the model folder. This process is entirely automatic, but the user can always explore and retrieve more data from the created files. For example, after running the parameter estimation analysis, a plot is generated with the original parameter set, the prior bounds, and optimized parameters, but the procedure for retrieving the optimized parameters and using them to create a new SBtab or a new settings file for subsequent runs is not yet automated. For additional explanations of these functionalities please see our GitHub repository documentation.^{Footnote 8}

### Import from SBtab to MATLABĀ®

The import tools we have developed generate all the model and data files that are needed to run any of the analysis options in MATLABĀ® that can be found in Table 1. These files are saved in subfolders of the main model that are created at run time. The files include a version of the model without any inputs in the MATLABĀ® (.mat) and SimBiologyĀ® (.sbproj) format, as well as several versions of the model corresponding to each experiment specified in the SBtab. The latter includes three versions of the model that are used for different purposes: equilibration, default and detailed. Equilibration does not have any inputs and is used to equilibrate the species, whereas the default and detailed versions are used for simulation of experiments. They have all the relevant inputs and outputs that are going to be measured when simulating an experiment, and only differ in the step size of the simulations (both of which can be chosen in the settings file with detailed usually being a smaller step size). Additionally, while not needed for the rest of our MATLABĀ® workflow, these import scripts also generateĀ *.tsv* files corresponding to the individual SBtab sheets (useful for tracking changes in GitHub), and an SBML file using MATLABĀ® built-in functions (level 2 version 4 encoding as of MATLABĀ® 2021a). The latter can be used by any simulator that can import SBML files but requires further processing for which an R script can be found in a separate repository.^{Footnote 9} Note that we have another converter from SBtab into SBML that runs in R instead of MATLABĀ®. We also generate other helper files that assist in the correct simulation of the MATLABĀ® model. A description of these and a more detailed explanation of the organization and of the created files and folders can be found in our documentation.^{Footnote 10}

### Parameter Estimation

MATLABĀ® offers a wide range of tools for function optimization. We have developed scripts that transform our parameter estimation problem into an objective function that can be optimized by various MATLABĀ® built-in optimizer functions. At the time of writing, these include āfminconā, simulated annealing (āsimulannealbndā), pattern search (āpatternsearchā), genetic algorithm (āgaā), particle swarm (āparticleswarmā), and surrogate optimization algorithms (āsurrogateoptā), for which MATLABĀ® provides thorough documentation. The code is built with flexibility in mind, so introduction of other MATLABĀ® built-in or custom optimization algorithms should be straightforward. The optimizers used are the ones chosen in the settings file. Our code supports the use and comparison of multiple optimizers at the same time, and multiple uses of the same optimization algorithm are also supported. This is particularly useful for using optimizers that are inherently single-threaded, e.g. the simulated annealing algorithm, since multiple simulated annealing optimizations can be performed in multiple computing cores. After performing the optimization, a file containing the results is produced from which the optimized outputs can be retrieved and used to manually update the SBtab or SimBiologyĀ® model. One of the equations used to calculate the score for how well the model outputs fit the experimental data can be found below (Eq.Ā 1). We have incorporated a few other ways of calculating the score, and custom scoring methods could also be added depending on the need (see documentation).

Here, \(Y\) represents the data that is going to be used to constrain the model, sourced either from experiments or previous models, and *y* represents the outputs of the model mapped to the data resulting of the simulation of the model under parameterization \(\theta\). The allowed mismatch \(\tau\) between the two simulation results is analogous to the standard deviation of a Gaussian noise model in data fitting. The resulting \(F\) is the objective function for optimization. The error is summed over \(n\), the number of points in each experimental output, \(m\), the number of experimental outputs in an experiment (which is four in our use case, see Fig.Ā 3B), and \(l\), the number of experiments (E0-E9 in our use case) (see Fig.Ā 2C).

Parameter estimation is generally based on experimental data. In this use case, however, we used simulated data of the concentrations of several species using the original version of the model in SimBiologyĀ® (the additional use cases provided in the Supplementary Materials use actual experimental data). After modifying the model, we minimized the difference between the old behavior and the updated modelās response through optimization. The simulation results from the old model can be considered as analogous to experimental data in a normal parameter estimation setting. Here, we merely aim to make an updated model agree with its earlier iteration, which itself was adjusted based on experimental data. When changing a module in a model it is crucial to protect the unchanged parts, which is why we performed parameter estimation using the key species that intersect the calcium and dopamine cascades, namely PP1, calmodulin and DARPP-32 (Fig.Ā 2 and Fig.Ā 3B). In this use case, the Particle Swarm Algorithm was chosen to perform the optimization, but all algorithms were capable of reasonable optimizations. The parameters obtained were then used to generate all the figures where optimized parameters are referred to. The choice of total amount of reactions, used to replace the original function that represented the CaMKII phosphorylation, were constrained by the optimization. We considered the outputs that we were measuring (Fig.Ā 3B) and added reactions until the addition of more did not meaningfully improve the fits.

### Validation in MATLABĀ®

We developed diagnostics scripts that can be used to reproduce the various experiments defined in SBtab. These scripts generate plots of the experimental inputs to the model (adapted for Fig.Ā 3A), the provided data and the outputs measured from model simulation (adapted for Fig.Ā 3B) given some choice of parameters, and plots of the scores calculated for the differences between the various experimental outputs and simulated model outputs. We used these tools to confirm that our parameter estimation resulted in a good fit for most of the species and the updated model was able to closely reproduce the results seen with the original model (Fig.Ā 3C). In our repository we provide the updated model in SBtab (.xlxs and.tsv), SBML (.xml) and MATLABĀ® SimBiologyĀ® (.sbproj and.mat). In addition to the general-purpose tools, we also wrote a use case-specific script, which uses data from the original model and reproduces the time-dependency of the substrate phosphorylation given different delays of the start of calcium and dopamine stimuli, using the optimized model (Fig.Ā 3C, D).

### Global Sensitivity Analysis

In many cases parameter estimation of biochemical pathway models does not result in one unique value for a parameter. Structural and practical unidentifiability (Raue et al., 2009) results in a large set of parameter values that all correspond to solutions with a good fit to the data, i.e., there is a large uncertainty in the parameter estimates (Eriksson et al., 2019). When this is the case, *local* sensitivity analysis is not so informative, since this can be different depending on which point in parameter space it is performed at. A global sensitivity analysis (GSA), on the other hand, covers a larger range of the parameter space. Several methods for GSA exist (Zi, 2011) but we have focused on a method by Sobol and Saltelli (Saltelli, 2002, 2004; Sobol, 2001) as implemented by Halnes et al. (2009) which is based on the decomposition of variances (Saltelli, 2004). Single parameters or subsets of parameters that have a large effect on the variance of the output get a high sensitivity score in this method. Intuitively, this method can be understood as varying all parameters but one (or a small subset) at the same time within a multivariate distribution to determine what effect this has on the output variance. If there is a large reduction in the variance, the parameter that was kept fixed is important for this output (Saltelli, 2004).

Let the vector *Ī* denote the parameters of the model, and *y*ā=ā*f(Ī)* be a scalar output from the model. In the sensitivity analysis *Ī* are stochastic variables, sampled from a multivariate distribution, whose variation gives a corresponding uncertainty of the output, quantified by the variance *V(Y)*. Note that in the setting and interpretations described here, the different *Ī*_{i} are assumed to be independent from each other (for cases with dependent *Ī*_{i} see e.g. (Saltelli, 2004) and (Eriksson et al., 2019)). We consider two types of sensitivity indices: the *first order effects S*_{i} and the *total order effects S*_{Ti}. The first order effects describe how the uncertainty in the output depends on the parameter *Ī*_{i} alone, i.e., how much of the variance of the output can be explained by the parameter *Ī*_{i} by itself. As an example, *S*_{i}ā=ā0.1 means that 10% of the output variance can be explained by *Ī*_{i} alone. The total order effects give an indication on the interactive effect the parameter *Ī*_{i} has with the rest of the parameters on the output. Parameters are said to interact when their effect on the output cannot be expressed as a sum of their single effects on the output.

The first order sensitivity index of the parameter *Ī*_{i} is defined as^{Footnote 11}

where *Ī*_{āi} corresponds to all elements of *Ī* except *Ī*_{i} The total order sensitivity index of the parameter *Ī*_{i} is defined as

If there is a large difference between S_{i} and S_{Ti}, this is an indication that this parameter takes part in interactions. For a detailed description see chapter 5 of Saltelli (2004).

The optimization described earlier takes place on log transformed space (log_{10}(*Ī*)). For the sensitivity analysis we perform the sampling on a lognormal distribution, meaning that log_{10}(*Ī*)ā~ā*N(Ī¼, Ļ).* Below we use *Ī¼*ā=ālog_{10}(*Ī**) and *Ļ*ā=ā0.1, where *Ī** correspond to the optimal values received from the optimization.^{Footnote 12} We illustrate this method using only the six parameters corresponding to the model module that has been replaced (Fig.Ā 2) and the results can be seen in Fig.Ā 4 Only four of the parameters (k216-k219/k222-k225; Fig.Ā 4) seem to be important for the output within the investigated parameter region, with the parameter k219/k225 dominating in experiments E3 to E9. There also seem to be some interactive effects between the parameters since S_{Ti} is larger than S_{i,} especially for the first four experiments (Fig.Ā 4A,Ā B).

## Compatibility and Validation with Other Simulation Environments

### Conversion to SBML and Simulations in COPASI

COPASI is one of the more commonly used modeling environments in systems biology and it can read SBML files (Hoops et al., 2006). Our first validation step is to use the SBML model retrieved from the SBtab model in COPASI. The online conversion tool from SBtab to SBML did not work for our specific use case; we wrote a new conversion function that can be found in a separate GitHub repository.^{Footnote 13} It interprets the biological model and converts it into plain ODEs in VFGENās custom format (.vf), the VFGEN file can then be used to create output in various languages^{Footnote 14} (Weckesser, 2008). The conversion script (written in R) converts the SBtab saved as a series ofĀ *.tsv* files or oneĀ .*ods* file into a VFGEN vector field file and as by-products also the SBML and a MOD file (see chapter *Conversion to a MOD file and simulations in NEURON*). To create an SBML model, libsbml must be installed with R bindings. The SBML file can be imported directly into COPASI.

Another way to convert the model into SBML is through a single MATLABĀ® SimBiologyĀ® function. The models, however, are created with long IDās that carry no biological information (the IDs are similar to hexadecimal hashes) for all model components, the units are not properly recognized, and some units may just be incorrectly defined in the output. We, therefore, created a script in R that asks for default units for the model and replaces the ones in SimBiology's SBML file. It fixes most issues (units, IDs, and the time variable in assignments) allowing the model to be properly imported in COPASI.^{Footnote 15} To illustrate that the SBML-converted model imported into COPASI produces the same results as in SimBiologyĀ®, we used a simplified calcium input corresponding to one double exponential spike analogous to the dopamine transient and simulated the model with deterministic solvers in both COPASI (LSODA solver) and MATLABĀ® SimBiologyĀ® under similar conditions. Both simulation environments produced almost overlapping results (Fig.Ā 5B, C) validating the converted model in the SBML format.

### Simulations in STEPS

The web-based subcellular simulation setup application^{Footnote 16} allows importing, combining and simulating models expressed in the BioNetGen language (BNGL; Harris et al., 2016). It supports the import of SBML (level 2 version 4) models and their transformation to rule-based BNGL form using Atomizer (Tapia & Faeder, 2013). The BioNetGen file format was extended to provide diffusion parameters, links to tetrahedral meshes describing the geometry of model compartments, as well as the additional parameters for solvers and stimulation protocols required for spatially distributed models. The subcellular simulation setup application is integrated with the network free solver NFsim (Sneddon et al., 2011) and it supports simulations of spatially distributed systems using STEPS (Hepburn et al., 2012). STEPS provides spatial stochastic and deterministic solvers for simulations of reactions and diffusion on tetrahedral meshes. Furthermore, the subcellular simulation setup application provides a number of facilities for the visualization of modelsā geometries and the results of simulations.

To demonstrate the compatibility of the subcellular simulation setup application with the workflow for model development described above, we imported the SBML version of the use case model to the setup application and simulated it with the STEPS TetOpSplit solver. We have used a simple two-compartmental spine model with a tetrahedral compartment corresponding to the spine compartment of the use case model. There is also a PSD compartment on one of the faces. The results of the model simulations with a STEPS solver were qualitatively similar to the results obtained with the deterministic model simulated in MATLABĀ®. Examples of simulated time courses for molecule concentrations as shown in Fig.Ā 3 in comparison with corresponding MATLABĀ® curves are shown in Fig.Ā 6.

### Conversion to a MOD File and Simulations in NEURON

As we suggested before, conversion between different modeling frameworks and formats facilitates collaboration. But conversion is made harder by the differences in the capabilities of different modeling packages. A model, such as the one above, could be useful for multiscale simulations investigating how neuronal network activity shapes synaptic plasticity. Many cellular level models are built and simulated in the NEURON environment which also supports simplified reactionādiffusion systems. It is useful to be able to integrate a subcellular level model into a cellular level model specified using NEURON. Models in NEURON are built by adding features with MOD files that are written in the NMODL programming language. A schematic illustration of our use case set up in NEURON is depicted in Fig.Ā 7. Conversion from SBML to MOD is already possible via NeuroML (instructions can be found in Lindroos et al., 2018); however, it is not an automated or user-friendly approach. We wrote an R script, which reads an SBtab model^{Footnote 17} and writes a MOD file, as well as VFGEN and an SBML^{Footnote 18} file. The script can optionally perform analysis of conservation laws and output a model where some of the differential equations for the state variables are substituted by algebraic equations arising from the conservation laws, thereby reducing the number of differential equations to be solved. Instructions on how to use the SBtab to VFGEN/MOD/SBML converter in R can be found in a separate GitHub repository^{Footnote 19}. The subcellular model in SBtab form is not aware of its coupling to a larger model of the cell and the user must edit the resulting MOD file manually to use it within a larger scope and assign a role to this model component. This typically means assigning input to the model and using the output in some way. An example is given below with our use case. Another point to stress is that the script converts the time unit of the parameters to milliseconds, NEURONās default unit for time, but does not change the concentration units. Thus, when coupling a biochemical cascade to other quantities in the neuronal model, care must be taken to rescale the coupled variables so that they match the units in the rest of the neuronal model. We also illustrate that with one of the inputs in the use case.

We validate the biochemical cascade model and our conversion tools in NEURON by qualitatively reproducing the results obtained in MATLABĀ® SimBiologyĀ®. Our goal is to show that the cascade model can be integrated into a single neuron model and bridge spatial and temporal scales of system behavior by linking the output of the cascade to changes in the synaptic properties and ultimately to the electrical behavior of the neuron model. Therefore, the biochemical model in the MOD format was incorporated into a single biophysically detailed and compartmentalized D1 MSN model from Lindroos et al. (2018). To integrate the MOD file into single cell models a few user-specific modifications have to be made to interface this model with the larger electrochemical system. First, the input is modified so that the calcium burst is represented by calcium influx from the cellās calcium channels and adjusted so that the overall calcium level would be similar to the input used in MATLABĀ® simulations. The calcium in the neuron model is expressed in millimolar units and when coupling it to the biochemical cascade we rescale it to nanomolar units (the units used in the biochemical cascade model). The dopamine transient is represented by an assignment expression (available in the Expression table of the SBtab) that creates its double exponential form (this can be used in other languages as well). For an adequate comparison of the substrate phosphorylation curve (shown in Fig.Ā 3D) obtained from the NEURON and MATLAB simulations it is necessary to provide the same input in the simulations. Hence, in the comparison in Fig.Ā 7C we used the calcium input recorded from the NEURON simulation in the MATLAB simulation.

In the D1 MSNs this biochemical signaling cascade causes synaptic strengthening via several mechanisms, one of which is the phosphorylation of AMPA receptors. As mentioned above, the model instead includes a generic substrate whose level of phosphorylation is the output of the cascade. For the purpose of illustrating the workflow with a proof-of-concept example, we have here linked the fraction of phosphorylated substrate to the AMPA receptor conductance (the conductance is scaled by 1ā+ā*f* (where *f*: fraction of the phosphorylated substrate), i.e. when there is very little substrate phosphorylation, very little change in the AMPA conductance is elicited, and vice versa. Modifying the model to include the reactions for AMPA receptor phosphorylation will be made in a future study.

## Discussion

In order to address the growing need for interoperability in biochemical pathway modeling within the neuroscience field, we have developed a workflow that can be used to refine models in all phases of development, keeping in mind the fact that many of the users (including us) are scientists and not professional programmers. For the model and data storage we have chosen the SBtab format which can be easily read and modified by both modelers and experimentalists, and can be converted into other formats, e.g. SBML, MATLABĀ® SimBiologyĀ® or MOD. Our workflow is modularized into different steps allowing the use of each step depending on the need and ensuring interoperability with other tools, such as those described in a similar endeavor named FindSim (Viswan et al., 2018). There are distinct advantages to the workflow, by enforcing a common standard for information exchange, it inherently makes the models more generalizable and reduces the likelihood that simulation results are artifacts of a particular simulator, and nonetheless, it gives users the flexibility to leverage the strengths of each different simulation environment and provides distinct stages of processing that would not be possible in any single simulator. The presented workflow aims to use software components that are free (apart from MATLABĀ®) and solve incremental sub-tasks within the workflow (with open standard intermediate files) to make the workflow easy to branch into scenarios we have not previously considered. It is also possible to circumvent MATLABĀ® entirely, if desired (e.g. conversion from SBtab to a MOD file that is then used by NEURON).

When deciding which software packages to use we find that an important aspect that must be considered is the cost and licensing. For some researchers, price may be a relevant concern, in other cases a researcher may have to undergo considerable overhead to make their institution/lab purchase a license and possibly operate a license server. Other than MATLABĀ®, we made the choice to disregard commercial products, keeping in line with the fieldās trend towards *open source* platforms. We will also expand our tools to support the use of models with more complex geometries, with several compartments (which is also an SBML feature), and tetrahedral meshes that can be used with STEPS in the subcellular simulation setup application. Such advanced geometries can in principle be defined within SBtab tables. When it comes to multiscale simulations, there is the possibility of using the ReactionāDiffusion module (RXD) in NEURON. Currently, however, it does not support the import of SBML as it lacks the concept of spatially extended models, but SBML support might be added to the future versions of RXD (McDougal et al., 2013).

Another interesting consideration is whether the user wants to define any model directly using rules (as in rule-based modeling). This would make the use of Atomizer unnecessary. The transformation from rules to classical reactions seems easier than the reverse, so even if a rule-based simulation is not necessary or too slow, a rule-based description may be shorter and more fundamental in terms of model translation.

When it comes to model analysis, we have here implemented a functionality for global sensitivity analysis. This is a thorough, but computationally demanding approach and the analysis usually needs to be run in parallel on a high performance computing environment. There are also faster but more approximate screening methods that could have been used (Saltelli, 2004). In the future we also intend to incorporate uncertainty quantification into the workflow (Eriksson et al., 2019).

While the current workflow is standalone, there are potential alignments to other systems that could assist adoption by the community. For example, other systems such as Galaxy are more general in scope focusing on bioinformatics and are aimed at naive users (Afgan et al., 2016). The current workflow focuses on neuroscience and assumes experienced users. While the current version requires user installation of dependencies, in the future Dockerised versions could potentially help ease installation requirements. Similarly, the Common Workflow Language (Amstutz et al., 2016) is another recent development that could facilitate the exchange of workflow information.

In summary, we have here presented a workflow for biochemical pathway modeling and provided a concrete use case of reward dependent synaptic plasticity, with two additional examples in the supplementary material and the workflow repository. Multiscale models are crucial when trying to understand the brain using modeling and simulations, e.g. how network activity shapes synaptic plasticity or how neuromodulation might affect cellular excitability on sub second timescales. Structured approaches for bridging from detailed cellular level neuron models, to more simplified or abstract cellular-, network-, and even brain region models are developing (Amsalem et al., 2020; Carlu et al., 2020; Schmutz et al., 2020). In the currently illustrated workflow, we add to these efforts by bridging from the subcellular scale to the cellular level scale. We updated parts of the use case model to accomplish a model with only bimolecular reactions that are easier to represent in standards such as SBML. The idea is that users can look at our model as a concrete test case, rerun the workflow (or parts thereof) and then replace the current example modelĀ (Fujita et al., 2010, Hass et al.,Ā 2019) with their own models. In this particular use case, we specifically focused on creating scripts to achieve interoperability between human readable model specification standards and machine- readable standards, and we also wanted to facilitate how a subcellular signaling model could be implemented in different solvers with different strengths, in this case both SimBiologyĀ® in MATLABĀ®, as well as STEPS and NEURON. NEURON is currently the most used simulation software for detailed cellular level neuron models, and STEPS can, as said, simulate signaling cascades in arbitrary dendritic morphologies both in a deterministic and stochastic manner. MATLABĀ® on the other hand has many functions, for example for parameter estimation and we included an implementation for global sensitivity analysis. Several other software is, however, used within the computational neuroscience community for cellular or subcellular model simulations (Akar et al., 2019; Oliveira et al., 2010; Ray & Bhalla, 2008; Resasco et al., 2012). To successively make as many of those tools interoperable with standards for both model and data specification, various parameter estimation and model analysis methods, visualization software, etc., will further facilitate the creation of FAIR multiscale modeling pipelines in the future.

## Information Sharing Statement

Source code of the Subcellular Workflow is available at https://github.com/jpgsantos/Subcellular_workflow/tree/1.0 and licensed under GNU General Public License v3.0., documentation of the Subcellular Workflow is available at https://subcellular-workflow.readthedocs.io/en/1.0/. We stored the files relevant to each of the models in independent repositories, https://github.com/jpgsantos/Model_Nair_2016/releases/tag/1.0 for the main model discussed in this paper, https://github.com/jpgsantos/Model_Viswan_2018/releases/tag/1.0 for the model discussed in the supplementary material, and https://github.com/jpgsantos/Model_Fujita_2010/releases/tag/1.0 for the model mentioned in the supplementary material but mostly defined in the GitHub. The models are stored in the SBtab format (Lubitz et al., 2016). Model reduction, parameter estimation and global sensitivity analysis tools are written in MATLABĀ® (RRID:SCR_001622) and require the SimBiologyĀ® toolbox. Conversion script to VFGEN (Weckesser, 2008), MOD and SBML (RRID:SCR_007422) is written in R (RRID:SCR_001905). Conversion to SBML requires the use of libSBML (RRID:SCR_014134). Validations are run in COPASI (RRID:SCR_014260; Hoops et al., 2006), NEURON (RRID:SCR_005393; Hines & Carnevale, 1997) and with the subcellular simulation setup application (RRID:SCR_018790; available at https://subcellular.humanbrainproject.eu/model/simulations) that uses a spatial solver provided by STEPS (RRID:SCR_008742; Hepburn et al., 2012) and network-free solver NFsim (available at http://michaelsneddon.net/nfsim/). The medium spiny neuron model (Lindroos et al., 2018) used in NEURON simulations is available in ModelDB database (RRID:SCR_007271) with access code 237,653. The FindSim use case model is available in https://github.com/BhallaLab/FindSim (Viswan et al., 2018).

## Availability of Data and Material

see Information Sharing Statement.

## Code Availability

see Information Sharing Statement.

## Change history

### 14 March 2022

The original version of this paper was updated to add the missing compact agreement Open Access funding note.

## Notes

for example: at the

*time of writing*the SimBiologyĀ® toolbox (in MATLABĀ®) exports the*time*variable as a normal variable, without the required definition URL, it must be added manually.Kernel density estimation does not work very well in high dimensional problems.

A library providing an application programming interface for SBML.

Where $$V$$ is the variance operator and $$E$$ (conditional) expected value.

Other distributions can be used as well.

Conversion function available in https://github.com/a-kramer/SBtabVFGEN and instructions in https://github.com/a-kramer/SBtabVFGEN/blob/master/README.md

Including, but not limited to: Python (NumPy), C (GNU Scientific Library (GSL), CVODE), GNU Octave (LSODE), R (deSolve), and core MATLABĀ® (e.g. for ode15s).

It should be noted that the units may not show up correctly in COPASI (depending on the version) even if they are correct in the SBML file itself.

The online application for subcellular simulations can be found in https://subcellular.humanbrainproject.eu/model/simulations with the documentation in https://humanbrainproject.github.io/hbp-sp6-guidebook/online_usecases/subcellular_level/subcellular_app/subcellular_app.html

A file given either as a series of tab separated text files or one open document spreadsheet file. Some content of the SBtab model is mandatory, some optional.

libSBML must installed with R bindings: i.e. SBML output is optional.

Conversion function available in https://github.com/a-kramer/SBtabVFGEN and instructions in https://github.com/a-kramer/SBtabVFGEN/blob/master/README.md

## References

Afgan, E., Baker, D., van den Beek, M., Blankenberg, D., Bouvier, D., Äech, M., Chilton, J., Clements, D., Coraor, N., Eberhard, C., GrĆ¼ning, B., Guerler, A., Hillman-Jackson, J., Von Kuster, G., Rasche, E., Soranzo, N., Turaga, N., Taylor, J., Nekrutenko, A., & Goecks, J. (2016). The Galaxy platform for accessible, reproducible and collaborative biomedical analyses: 2016 update.

*Nucleic Acids Research.,**44*(W1), W3āW10.Akar, N. A., Cumming, B., Karakasis, V., KĆ¼sters, A., Klijn, W., Peyser, A., & Yates, S. (2019). Arbor ā A morphologically-detailed neural network simulation library for contemporary high-performance computing architectures.

*27th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP)*, Pavia, Italy, pp. 274ā282.Amsalem, O., Eyal, G., Rogozinski, N., Gevaert, M., Kumbhar, P., SchĆ¼rmann, F., & Segev, I. (2020). An efficient analytical reduction of detailed nonlinear neuron models.

*Nature Communications,**11*(1), 288.Amstutz, P., Crusoe, M., TijaniÄ, N., Chapman, B., Chilton, J., Heuer, M., Kartashov, A., Leehr, D., MĆ©nager, H., Nedeljkovich, M., Scales, M., Soiland-Reyes, S., & Stojanovic, L. (2016): Common Workflow Language, v1.0. Specification, Common Workflow Language working group. https://w3id.org/cwl/v1.0/

Bhalla, U. S., & Iyengar, R. (1999). Emergent properties of networks of biological signaling pathways.

*Science,**283*(5400), 381ā387.Bhalla, U. S. (2004). Models of cell signaling pathways.

*Current Opinion in Genetics & Development,**14*, 375ā381.Bois, F. Y. (2009). GNU MCSim: Bayesian statistical inference for SBML-coded systems biology models.

*Bioinformatics,**25*, 1453ā1454.Cannon, R. C., Gewaltig, M. -O., Gleeson, P., Bhalla, U. S., Cornelis, H., Hines, M. L., et al. (2007). Interoperability of neuroscience modeling software: Current status and future directions.

*Neuroinformatics,**5*(2), 127ā138.Carlu, M., Chehab, O., Dalla Porta, L., Depannemaecker, D., HĆ©ricĆ©, C., Jedynak, M., KĆ¶ksal ErsĆ¶z, E., Muratore, P., Souihel, S., Capone, C., Zerlaut, Y., Destexhe, A., & di Volo, M. (2020). A mean-field approach to the dynamics of networks of complex neurons, from nonlinear Integrate- and Fire to Hodgkin-Huxley models.

*Journal of Neurophysiology,**123*(3), 1042ā1051.Chylek, L. A., Harris, L. A., Faeder, J. R., & Hlavacek, W. S. (2015). Modeling for (physical) biologists: an introduction to rule-based approach.

*Physical Biology,**12*(4), 045007.Djurfeldt, M., Hjorth, J., & Eppler, J. M., et al. (2010). Run-Time Interoperability Between Neuronal Network Simulators Based on the MUSIC Framework.

*Neuroinform,**8,*43ā60. https://doi.org/10.1007/s12021-010-9064-z.Eriksson, O., Jauhiainen, A., Maad Sasane, S., Kramer, A., Nair, A. G., Sartorius, C., & Hellgren Kotaleski, J. (2019). Uncertainty quantification, propagation and characterization by Bayesian analysis combined with global sensitivity analysis applied to dynamical intracellular pathway models.

*Bioinformatics,**35*(2), 284ā292.Fujita, K. A., Toyoshima, Y., Uda, S., Ozaki, Y., Kubota, H., & Kuroda, S. (2010). Decoupling of receptor and downstream signals in the Akt pathway by its low-pass filter characteristics.

*Science signaling,**3*(132), ra56.Gillespie, D. T. (1976). A General Method for Numerically Simulating the Stochastic Time Evolution of Coupled Chemical Reactions.

*Journal of Computational Physics,**22*(4), 403ā434.Gillespie, D. T. (2001). Approximate accelerated stochastic simulation of chemically reacting systems.

*Journal of Chemical Physics,**115*(4), 1716ā1733.Gleeson P, Crook S, Cannon RC, Hines ML, Billings GO, et al. (2010) NeuroML: A Language for Describing Data Driven Models of Neurons and Networks with a High Degree of Biological Detail. PLOS Computational Biology 6(6): e1000815. https://doi.org/10.1371/journal.pcbi.1000815

Gleeson, P., Steuber, V., Silver, R. A., & Crook, S. (2012). NeuroML. In: Le NovĆØre N. (eds)

*Computational Systems Neurobiology*. Springer, Dordrecht.Haario, H., Laine, M., Mira, A., & Saksman, E. (2006). DRAM: Efficient adaptive MCMC.

*Statistics and Computing,**16*, 339ā354.Halnes, G., Ulfhielm, E., EklĆ¶f Ljunggren, E., Hellgren Kotaleski, J., & Rospars, J. P. (2009). Modelling and sensitivity analysis of the reactions involving receptor, G-protein and effector in vertebrate olfactory receptor neurons.

*Journal of Computational Neuroscience,**27*(3), 471ā491.Harris, L. A., Hogg, J. S., Tapia, J. J., Sekar, J. A., Gupta, S., Korsunsky, I., Arora, A., Baruda, D., Sheehan, R. P., & Faeder, J. R. (2016). BioNetGen 2.2: advances in rule-based modeling.

*Bioinformatics,*Ā*32*(21), 3366ā3368.Hass, H., Loos, C., Raimndez-Alvarez, E., Timmer, J., Hasenauer, J., & Kreutz, C. (2019). Benchmark problems for dynamic modeling of intracellular processes.

*Bioinformatics,**35*, 3073ā3082.Hedley, W. J., Nelson, M. R., Bullivant, D. P., & Nielsen, P. F. (2001). A short introduction to CellML.

*Philosophical Transactions of the Royal Society, Series A,**359*, 1073ā1089.Hellgren Kotaleski, J., & Blackwell, K. T. (2010). Modelling the molecular mechanisms of synaptic plasticity using systems biology approaches.

*Nature Reviews Neuroscience,**11*, 239ā251.Hepburn, I., Chen, W., Wils, S., & De Schutter, E. (2012). STEPS: Efficient simulation of stochastic reaction-diffusion models in realistic morphologies.

*BMC Systems Biology,**6*, 36.Hines, M. L., & Carnevale, N. T. (1997). The NEURON simulation environment.

*Neural Computation,**9*, 1179ā1209.Hoops, S., Sahle, S., Gauges, R., Lee, C., Pahle, J., Simus, N., Singhal, M., Xu, L., Mendes, P., & Kummer, U. (2006). COPASI: A COmplex PAthway SImulator.

*Bioinformatics,**22*, 3067ā3074.Hucka, M., Finney, A., Sauro, H. M., Bolouri, H., Doyle, J. C., Kitano, H., et al. (2003). The systems biology markup language (SBML): A medium for representation and exchange of biochemical network models.

*Bioinformatics,**19*, 524ā531.Klinger, E., Rickert, D., & Hasenauer, J. (2018). pyABC: Distributed, likelihood-free inference.

*Bioinformatics,**34*(20), 3591ā3593.Li, L., Stefan, M. I., & Le Novere, N. (2012). Calcium input frequency, duration and amplitude differentially modulate the relative activation of calcineurin and CaMKII.

*PLoS One*,*7*, e43810.Lindroos, R., Dorst, M. C., Du, K., FilipoviÄ, M., Keller, D., Ketzef, M., Kozlov, A. K., Kumar, A., Lindahl, M., Nair, A. G., PĆ©rĆ©z-Fernandez, J., Grillner, S., Silberberg, G., & Hellgren Kotaleski, J. (2018). Basal ganglia neuromodulation over multiple temporal and structural scales-simulations of direct pathway MSNs investigate the fast onset of dopaminergic effects and predict the role of Kv4.2.

*Frontiers in Neural Circuits,**12*, 3.Lubitz, T., Hahn, J., Bergmann, F. T., Noor, E., Klipp, E., & Liebermeister, W. (2016). SBtab: A flexible table format for data exchange in systems biology.

*Bioinformatics,**32*(16), 2559ā2561.Maiwald, T., & Timmer, J. (2008). Dynamical modeling and multi-experiment fitting with PottersWheel.

*Bioinformatics,**24*(18), 2037ā2043.McDougal, R. A., Hines, M. L., & Lytton, W. W. (2013). Reaction-diffusion in the NEURON simulator.

*Frontiers in Neuroinformatics,**7*, 28.Nair, A. G., Bhalla, U. S., & Kotaleski J. H. (2016). Role of DARPP-32 and ARPP-21 in the emergence of temporal constraints on striatal Calcium and Dopamine integration.

*PLoS Computational Biology,*Ā*12*(9), e1005080.Oliveira, R. F., Terrin, A., Di Benedetto, G., Cannon, R. C., Koh, W., Kim M., Zaccolo, M., & Blackwell K. T. (2010). The role of type 4 phosphodiesterases in generating microdomains of cAMP: large scale stochastic simulations.

*PLoS One,**5*(7), e11725.Pepke, S., Kinzer-Ursem, T., Mihalas, S., & Kennedy, M. B. (2010). A dynamic model of interactions of Ca2+, calmodulin, and catalytic subunits of Ca2+/calmodulin-dependent protein kinase II.

*PLoS Computational Biology*,*6*(2), e1000675.Raue, A., Kreutz, C., Maiwald, T., Bachman, J., Schilling, M., KlingmĆ¼ller, U., & Timer, J. (2009). Structural and practical identifiability analysis of partially observed dynamical models by exploiting the profile likelihood.

*Bioinformatics,**25*(15), 1923ā1929.Ray, S., & Bhalla, U. S. (2008). PyMOOSE: Interoperable scripting in Python and MOOSE.

*Frontiers in Neuroinformatics,**2*, 6.Resasco, D. C., Gao, F., Morgan, F., Novak, I. L., Schaff, J. C., & Slepchenko, B. M. (2012). Virtual Cell: Computational tools for modeling in cell biology.

*Wiley Interdisciplinary Reviews: Systems Biology and Medicine,**4*(2), 129ā140.Rodriguez, N., Pettit, J. -B., Dalle Pezze, P., Li, L., Henry, A., van Iersel, M. P., et al. (2016). The systems biology format converter.

*BMC Bioinformatics,**17,*154.Saltelli, A. (2002). Making best use of model evaluations to compute sensitivity indices.

*Computer Physics Communications,**145*, 280ā297.Saltelli, A. (2004).

*Sensitivity analysis in practice: A guide to assessing scientific models*. Wiley.Schmidt, H., & Jirstrand, M. (2006). Systems Biology Toolbox for MATLAB: A computational platform for research in systems biology.

*Bioinformatics,**22*(4), 514ā515.Schmutz, V., Gerstner, W., & Schwalger, T. (2020). Mesoscopic population equations for spiking neural networks with synaptic short-term plasticity.

*The Journal of Mathematical Neuroscience,**10*(1), 5.SchĆ¤lte, Y., FrĆ¶hlich, F., Stapor, P., Wang, D., Vanhoefer, J., Weindl, D., et al. (2020). ICB-DCM/pyPESTO: pyPESTO 0.2.0 (Version v0.2.0). Zenodo.

Sneddon, M. W., Faeder, J. R., & Emonet, T. (2011). Efficient modeling, simulation and coarse-graining of biological complexity with NFsim.

*Nature Methods,**8*(2), 177ā183.Sobol, I. M. (2001). Global sensitivity indices for nonlinear mathematical models and their Monte Carlo estimates.

*Mathematics and Computers in Simulation,**55*(1ā3), 271ā280.Tapia, J. J., & Faeder, J. R. (2013). The Atomizer: extracting implicit molecular structure from reaction network models. In:

*Proceedings of the International Conference on Bioinformatics, Computational Biology and Biomedical Informatics (BCBā13)*, ACM, New York, NY, pp. 726ā727.TennĆøe, S., Halnes, G., & Einevoll, G. T. (2018). Uncertainpy: A python toolbox for uncertainty quantification and sensitivity analysis in computational neuroscience.

*Frontiers in Neuroinformatics,**14*(12), 49.Viswan, N. A., HarshaRani, G. V., Stefan, M. I., & Bhalla, U. S. (2018). FindSim: A framework for integrating neuronal data and signaling models.

*Frontiers in Neuroinformatics,**12*, 38.Vlad, M. O., & Ross, J. (1994). Thermodynamic approach to nonequilibrium chemical fluctuations.

*Journal of Chemical Physics,**100*(10), 7295ā7309.Weckesser, W. (2008). VFGEN: A code generation tool.

*Journal of Numerical Analysis, Industrial and Applied Mathematics,**3*(1ā2), 151ā165.Wegscheider, R. (1901). Ćber simultane Gleichgewichte und die Beziehungen zwischen Thermodynamik und Reactionskinetik homogener Systeme.

*Monatshefte FĆ¼r Chemie,**22*, 849ā906.Welsh, C. M., Fullard, N., Proctor, C. J., Martinez-Guimera, A., Isfort, R. J., Bascom, C. C., Tasseff, R., Przyborski, S. A., & Shanley, D. P. (2018). PyCoTools: A Python toolbox for COPASI.

*Bioinformatics,**34*(21), 3702ā3710.Wilkinson, M. D., Dumontier, M., Aalbersberg, I. J., Appleton, G., Axton, M., Baak, A., ... & Mons, B. (2016). The FAIR Guiding Principles for scientific data management and stewardship.

*Scientific data*,*3*(1), 1-9.Yagishita, S., Hayashi-Takagi, A., Ellis-Davies, G. C. R., Urakubo, H., Ishii, S., & Kasai, H. (2014). A critical time window for dopamine actions on the structural plasticity of dendritic spines.

*Science,**345*(6204), 1616ā1620. https://doi.org/10.1126/science.1255514Zi, Z. (2011). Sensitivity analysis approaches applied to systems biology models.

*IET Systems Biology,**5*(6), 336ā346.

## Acknowledgements

We thank Pavlo Getta for engineering support, Geir Halnes for sharing scripts for global sensitivity analysis with us, and Sahil Moza for helping with the FindSim use case. The simulations were partly performed on resources provided by the Swedish National Infrastructure for Computing (SNIC) at Lunarc.

## Funding

Open access funding provided by Royal Institute of Technology. The study was supported by the Swedish research council (VR-M-2017ā02806, VR-M-2020ā01652), Swedish e-Science Research Centre (SeRC), EU/Horizon 2020 no. 785907 (HBP SGA2) and no. 945539 (HBP SGA3); EPFL Blue Brain Project Fund and the ETH Board Funding to the Blue Brain Project; scholarship PD/BD/114180/2016 from FCT FundaĆ§Ć£o para a CiĆŖncia e Tecnologia.

## Author information

### Authors and Affiliations

### Corresponding authors

## Ethics declarations

### Ethics Approval

not applicable.

### Consent to Participate

not applicable.

### Consent for Publication

not applicable.

### Conflicts of Interest/Competing Interests

The authors declare that they have no conflict of interest.

## Additional information

### Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Supplementary Information

Below is the link to the electronic supplementary material.

## Rights and permissions

**Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

## About this article

### Cite this article

Santos, J.P.G., Pajo, K., Trpevski, D. *et al.* A Modular Workflow for Model Building, Analysis, and Parameter Estimation in Systems Biology and Neuroscience.
*Neuroinform* **20**, 241ā259 (2022). https://doi.org/10.1007/s12021-021-09546-3

Accepted:

Published:

Issue Date:

DOI: https://doi.org/10.1007/s12021-021-09546-3

### Keywords

- Interoperability
- Multiscale modeling
- SBtab
- Global sensitivity analysis
- Parameter estimation
- Systems biology