Post-genomic era and P4 medicine

Mendel's studies of inheritance patterns laid the foundation for our current understanding of monogenic or single-gene diseases in human. The laws of Mendelian inheritance, however, could not explain the polygenic or multifactorial inheritance of complex diseases (sporadic vs. familial inheritance). Complex diseases such as cancer or Alzheimer's disease (AD) represent highly heterogeneous clinical states, which reflect combined effects of various genes and their interaction with environmental factors. Analysis and annotation of genomic information has been the main focus of research in the recent past. Since the publication of the Human Genome Project on 26 June 2000, tremendous progress has been made in whole genome sequencing, analysis and interpretation. This progress is reflected in the increasing number of diseases for which genetic testing is available, according to the GeneTests database [1] (Figure 1).

Figure 1
figure 1

The rise of genome-based diagnostic applications. The number of genetic tests for diagnosis of different disease indications - offered by medical laboratories worldwide - has been growing since 1993.

Despite the increasing availability of genetic tests offered by medical laboratories, a fundamental question remains unanswered: how effectively can such molecular data sets be translated into clinical applications, ideally at the point of care, while compatible with patient-centric view of personalized medicine? Although recent efforts such as sequencing human exomes [2], 1000 Genomes Project [3] and Personal Genome Project [4] have been dedicated to characterizing the vast majority of common single-nucleotide polymorphisms (SNPs) and structural variants across the genome, they only provide a one-dimensional view of genome function. Besides, with the advent of new high-throughput technologies such as DNA sequencers, gene expression microarrays and mass spectrometry, other forms of high-dimensional molecular data have emerged that need to be analyzed and used for more accurate diagnosis and treatment prescription.

The realization of P4 medicine promises relies on the ability to manage and integrate different heterogeneous data types across multiple scales from molecular data sets to clinical information and medical history of patients. Handling patients' data and its exchange amongst healthcare information systems has been the focus of informatics endeavors in the realm of personalized medicine [5]. For example, CDISC (Clinical Data Interchange Standards Consortium) brings together leading global biotechnology and pharmaceutical development companies as well as government institutions, academic research centers and other non-profit organizations to develop and support global, platform-independent data standards that enable interoperability of clinical data amongst information systems [6]. Although standardization of data storage and representation is necessary for the future of P4 medicine but it is certainly not enough. The main challenge to be addressed here by translational bioinformatics is the problem of linking omics molecular data and affected biological pathways to clinical readouts and epidemiological background of patients with complex diseases. Integration, analysis and interpretation of such heterogeneous data types in the context of complex molecular systems requires an integrative approach to studying systems of biological components, referred to as systems biology. In the post-genomic era, therefore, the greatest challenge is to determine how heterogeneous ‘-omics’ data can be integrated into a coherent holistic model that can explain the disease phenotype in a personalized way. The goal of systems approaches is to decipher the complexity of diseases by integrating all types of heterogeneous biological information into predictive and actionable models (Figure 2). Such models will be central to the premises underlying P4 medicine because they help investigators bridge the ‘translational gap’ between basic and clinical research more efficiently.

Figure 2
figure 2

Disease-specific models at the center of data integration. Different data types across multiple biological scales can be aggregated and integrated into an integrative mechanistic model.

Following the recommendations of the EPMA white paper on developing integrative medical approaches [7], we review the current trend in systems modeling and its application to the P4 paradigm. Next, we address the four Ps individually, i.e. predictive, preventive, personalized and participatory, in the context of translational disease modeling and provide scenarios for each case using the latest developments in this area. Finally, we conclude the review with a summary on the emerging implications of integrative disease modeling in P4 medicine.

Shifting from genome-based to network-based views of human disease

With the paradigm shift in the post-genomic era from single-dimension biological data to multiple dimensions of heterogeneous data, the concept of personalized medicine, which originally defined on the basis of SNPs, is now undergoing a revolutionary change. It is increasingly recognized that analysis of SNPs in isolation does not lead to a complete understanding of complex disease processes and - as a consequence - heterogeneous data must be analyzed and interpreted in an integrated fashion. Systems biology analyses have led to the emergence of new approaches, which take advantage of a network-based integrative view on different biological components involved in pathogenesis of complex diseases rather than a genome-centric single view with no sufficient context. For example, genome-wide association studies (GWAS) have been extensively used to unravel complex associations between genes and diseases at the level of SNPs but they do not provide sufficient context for understanding the complexity of pathological dysregulation involved in the initiation and progression of complex diseases. The knowledge provided by GWAS has a number of limitations that restricts its application, when used slone: the affected genes are difficult to definitively identify and localize, the alteration of gene function by identified SNPs in the context of disease cannot be immediately elucidated and it often remains unknown which particular pathways might be modulated by the SNPs that were found [810]. Instead, molecular networks that represent molecular states of the perturbed biological system underlying disease (also known as disease maps) provide a suitable framework for transitioning from ‘descriptive’ to ‘mechanistic’ mode by linking genetic information to disease processes and clinical phenotypes. It is within such a framework that associations amongst biological entities become apparent and a bigger picture of the disease mechanism emerges [11]. To this end, construction of disease-specific molecular maps constitutes the first step towards integrative disease modeling: for example, a computationally tractable interaction map of Parkinson's disease that integrates pathways implicated in Parkinson's pathogenesis has been recently constructed and made openly accessible [12]. However, disease molecular maps only serve as a backbone for the addition of other complementary data from multiple biological scales so that such maps, enriched with genetic variation information, can support researchers in modeling and generating novel hypotheses from those models, which can be further validated in the wet lab (Figure 3).

Figure 3
figure 3

The cycle of systems modeling.

Model-driven approach to integrating biomedical knowledge and data

In parallel to the unprecedented growth of experimental data produced by high-throughput technologies, the rate of published knowledge through the biomedical literature has been constantly increasing. Biomedical literature including scientific publications and health records contain valuable information in free-form densely written text. When extracted and converted from unstructured to structured form, literature-derived information can complement data-driven approaches by refining hypotheses generated from high-throughput data sets. Such information helps to distinguish signal from noise when analyzing and interpreting high-throughput data sets [13]. For example, network models of prior biological knowledge have been used to measure the amplitude of perturbation that a given stimulus such as drug treatment or environmental agent induces in a biological system through alterations in gene expression; in this way, the biological impact caused by environmental factors, toxic substances, or drug treatments can be scored and compared [14].

Integration of both experiment-derived and literature-derived information into a single framework has been successfully used for analysis and interpretation of biological mechanisms underlying disease using computational models. Usually modeling biological mechanisms can be driven either by high-throughput experimental data or by prior knowledge of molecular biology published in the literature. The former approach, known as ‘data-driven modeling”, is solely based on the data and no assumption is made about the underlying mechanisms [15]. The limitation of the data-driven approach is, however, the small amount of available quantitative data and the heterogeneity of such data sets. These limitations can be largely compensated with the qualitative statements from the literature [16]. Therefore, the so-called ‘knowledge-driven modeling’ approach makes use of systematically captured expert knowledge, text-mining technologies and semantic resources such as ontologies to build or validate biological networks [17].

Based on the complementary nature of these two modeling approaches, we propose a hybrid approach - the so-called ‘model-driven approach’ - that combines both data- and knowledge-driven methods. The models generated by this approach could represent correlation or cause and effect, depending on the type of associations between pairs of variables in the network model (Figure 4). Correlation network models are routinely built using high-dimensional data such as protein-protein interactions or gene expression data by establishing pairwise relations (i.e. edges) for all variables (i.e. nodes) but they confound direct and indirect associations (‘is_somehow_related_with’) and do not distinguish between cause and effect. In contrast, causal network models aim at representing response variables and covariates and, thus, the directionality of associations between cause and effect. When relationships between variables represent conditional dependencies (e.g. Given disease symptoms, compute the probabilities of the presence of various diseases), the model is a Bayesian network, which requires information on prior distribution. In the absence of such information, Biological Expression Language (BEL) offers an alternative. BEL is the state-of-the-art causal network modeling language that integrates literature-derived ‘cause and effect’ relationships into a data-driven platform and produces casual network models [18].

Figure 4
figure 4

Model-driven approach to integrating biological data. Proposed model-driven approach combines biomedical knowledge and data into a single disease model, which could represent correlations (left) or cause and effect (right).

We have recently developed BEL-based, computer-processable models representing the physiology of amyloid beta precursor protein (APP); these models represent both the normal and the diseased condition. We have used these models for differential network analysis.

Central role of integrative modeling in the future of P4 medicine: moving beyond genomics

With the increasingly complex relationship between basic research and clinical application, there is a pressing need to bridge the translational gap from bench to clinic using integrative methods. The mission of ‘translational bioinformatics’ is, therefore, to provide infrastructure and techniques that enable integrative modeling of the whole biological system across multiple scales. Since the objective of P4 medicine is to enable a multitude of purposes in the frame of individualized healthcare including predictive modeling, preventive measures and personalized treatments, a significant amount of personalized data and information should be managed for proper individualized diagnosis and prognosis. As mentioned above, integrative modeling approaches provide a suitable medium for fusion of such data and interpretation of the information derived from such models. We believe that integrative modeling approaches become the major building block of future P4 medicine efforts as their potential in bridging the translational gap is going to be increasingly appreciated and we will soon witness the first examples for model-driven, personalized treatment optimization.

Application of integrative methods enforces a paradigm shift from the conventional concept of personalized medicine - merely based on genetic makeup of patients (SNPs) - to a modern concept that includes SNPs as one piece of information within a bigger picture amongst other data types. An outcome of this activity is generation of various models that can provide decision support for both researchers and clinicians (Figure 5). For instance, the integration of brain imaging information from each patient into a molecular network model that explains the mode of action of CNS drugs could support clinicians in making decisions about diagnostic or prognostic measures. Similarly, enriched disease models can support the identification of the optimal drugs for targeted treatment of individual patients through the so-called drug-biomarker co-development. Taken together, we envision that general disease models will be personalized and tailored to the need of individual patients through enriching the model with specific patient data.

Figure 5
figure 5

Decision support options provided by the integrative disease modeling approach. The integrative modeling approach can provide support to both basic researchers and clinicians in making decisions about diagnostic or prognostic measures.

Contribution of integrative modeling to P4 medicine

The anticipation that integrative modeling lies at the core of the future P4 medicine raises an important question: how integrative systems models could contribute to solving the real-world problems in predictive, preventive, personalized and participatory areas? In the following, we review the potential contribution of integrative modeling to each P and provide a scenario per P to showcase the applicability of the integrative models to all aspects of P4 medicine.

Predictive: early diagnosis for the prevention of chronic degenerative diseases

Given the increasingly crucial role of molecular data in the clinical management of patients with complex diseases such as cancer, predictions based on diagnostic biomarkers are emerging as major players in individualized medicine. Biomarkers are defined as indicators of normal biological processes, pathogenic processes, or pharmacological responses to a therapeutic. Diagnostic biomarkers not only support detection of prodromal signs but also determine the progression rout of the disease through indication of stage of the disease or subclinical manifestation of the disease [19].

Predictive diagnostics are needed for early treatment of complex diseases such as Alzheimer's disease or cancer, which present clinical heterogeneity. Conventional diagnostic methods based on measuring just a single parameter encounter the danger of low sensitivity to accurately differentiate patients with highly heterogeneous clinical manifestations. Genome-based molecular fingerprints, including gene expression profiling, have proven successful in linking genome-level events to clinical metrics but they remain far from the clinical application, as they often require invasive sampling and their measurements suffer from lack of consistency. An alternative to such a ‘signature-based approach’ would be a ‘multi-panel approach’ by which gene expression information is linked to proteomics and pathophysiology of disease so that many different types of biomolecules (i.e. a panel of biomarkers) are being associated to pathological processes [20]. Integrative approaches that make use of network models as integration platform have shown a great promise for supporting discovery of highly discerning subsets of molecular biomarkers from the vast combinatorial space of molecular candidates [21]. For instance, the integration of gene expression profiles obtained from primary breast tumors in a protein interaction network model led to identification of sub-network biomarkers that represent metastatic tumor progression [22].

Although predictive diagnostics are conventionally represented by genome-based methods and tests, imaging markers are emerging as an integral part of medical diagnosis, particularly for diagnosis of neurological disorders. Hence, translating imaging readouts to diagnostic molecular biomarkers is expected to be extremely valuable. To this end, we have developed an algorithm that incorporates the diagnostic knowledge of imaging and potential protein biomarkers specific to Alzheimer's disease into a brain-specific protein interaction network [23]. As a result, three network models representing disease stages were generated and subjected to pathway analysis (Figure 6). The results indicated that, in contrast to the widely accepted amyloid pathway hypothesis, particular pathways are causally related to the disease mechanism. This was not appreciated before.

Figure 6
figure 6

Modeling staging mechanism in Alzheimer's disease. Schematic representation of the proposed workflow for translation of imaging information into stage-specific disease maps.

Preventive: reducing the likelihood of disease and disability

The primary goal of preventive measures against complex diseases is to identify at-risk individuals long before the development of disease symptoms (for instance, individuals at risk of developing Alzheimer's disease) so that preventive treatments can be planned. Accordingly, preventive biomarkers aim at screening a population and stratifying individuals at a high risk of developing disease by measuring the association between their molecular profile (e.g. gene expression or genetic variation) and disease phenotype. However, finding efficient preventive biomarkers is a non-trivial task that requires characterization of composition, distribution and function of key molecules involved in the disease mechanism.

Computational modeling approaches have been used already to assist the process of stratification biomarker discovery for cardiovascular diseases [24]. Recently, an interesting area for application of computational models to disease prevention is emerging that uses mechanistic models to reveal interdependencies between the disease and risk mechanisms (Figure 7). Such interdependencies are reflected in comorbidities that co-occur with the disease and increase the disease risk.

Figure 7
figure 7

Deterministic factors involved in shaping the disease phenotype. The disease phenotype is the result of interactions between causative molecules, environmental factors and disease-modifying molecules.

In the following, we present a scenario that shows therapeutic prevention of diabetes can reduce the risk of developing Alzheimer's disease in at-risk populations.

Epidemiological findings indicate that type II diabetes mellitus is linked to developing and exacerbating AD pathology [25, 26] so that Alzheimer's has been even proposed by some authors to be ‘type III diabetes’ [27, 28]. Very recently, a 9-year prospective study on 3,069 elderly adults without dementia demonstrated that patients who suffered from diabetes had significantly worse cognitive decline in comparison with those who did not have the disease, suggesting the contribution of diabetes mellitus severity to accelerated cognitive impairment [29]. On the other hand, pharmacological studies provide evidence that application of anti-diabetes drugs confer protective effects to demented brains. In mouse models of Alzheimer's disease, application of an anti-diabetic drug called extendin-4 reversed insulin pathology and improved cognition significantly [30]. In another study, it was shown that the diabetes drug Liraglutide prevents key neurodegenerative developments in a mouse model of Alzheimer's disease [31]. Rosaglitazone, another anti-diabetic agent, also showed beneficial effects for dementia treatment when administered at low dose [32]. In spite of such growing evidence, the molecular mechanism underlying this protective effect is still unclear and thus, aggregation and analysis of disparate knowledge on this topic in the frame of a unified computational model may facilitate the mission of finding a preventive treatment for Alzheimer's disease. In this direction, we have recently proposed a mechanistic model for genomic hormone interactions underlying dementia, which not only reveals the possible molecular connections between the insulin signaling pathway and learning/memory functions but also explains the preventive mode of action by serendipitous off-target effects of several approved drugs [33].

Personalized: targeted therapy as an emergent trend in drug discovery and development

The concept of personalized medicine was first coined in the context of genetics but its definition has been extended beyond genomics to the customization of healthcare measures to individual patients [34]. The ultimate goal of personalized medicine is to use integrative data to better target the delivery of healthcare and treatments to individuals [35]. Therefore, strategies and methods are required to enable classification of individuals into subpopulations based on their susceptibility to disease or response to treatment (Figure 8). Accordingly, successful stratification of individuals is the key to the following preventive and therapeutic interventions [36].

Figure 8
figure 8

Stratification of patient population based on disease mechanism.

Prognostic and predictive biomarkers are considered to be the key contributors to the success of personalized medicine. Prognostic biomarkers stratify patients undergoing a specific therapy based on the risk of disease progression whereas predictive (stratification or companion) biomarkers identify patients who are more likely to successfully respond to specific therapeutic intervention. A major drawback in delivering right pharmacological treatment to right patients is that the current therapeutic options are based on the ‘one-size-fits-all’ paradigm and thus not appropriate for stratification of responders from non-responders to treatment. Due to this fact and also because of a couple of other reasons, including the pressing need for better safety and efficacy indices and the need to control increasing healthcare expenditures, a new trend is emerging in pharmaceutical businesses that shifts from mass therapies towards targeted therapies based on so-called companion biomarkers [37]. Despite previous success stories of companion biomarker applications, a significant gap exists between the R&D expenditure and available clinically validated biomarkers. Therefore, better understanding of the mode of action of drugs as well as the alternative pathways that can reduce therapeutic effects will be crucial to discovery of novel companion biomarkers.

We have recently reviewed current biomarker-discovery technologies highlighting challenges and opportunities for oncology biomarker discovery with the emphasis on computational integrative modeling approaches as an emerging trend in biomarker prediction [38]. The overall trend indicates that there is a move away from correlative biomarkers towards causative biomarkers. Thus, the aim of next-generation integrative models is to capture causal relationships between candidate biomarker and clinical outcome. A prime example is development of predictive biomarker content for identification of patients with ulcerative colitis who could be potential responders to targeted anti-TNF therapy with Infliximab [39]. Based on the prior knowledge in the literature, a causal network model was constructed that describes mechanistic knowledge underlying ulcerative colitis in the form of the ‘cause-relationship-effect’ pattern. Next, gene expression profiles of responders and non-responders were incorporated into the causal model and a mechanistic strength value was calculated on the gene expression network activity signature of TNF signaling for each patient in the population. The model demonstrated that non-responders have different TNF signature compared to responders, which was due to sustained TNF-like downstream signaling in non-responders after treatment with Infliximab, controlled by alternative upstream controllers.

Participatory: enriching models with patient-centric information and experiences

In parallel to the concepts of personalized medicine and targeted therapy, patient-centric medicine is developing based on patient-oriented research. The aim of patient-oriented research is to identify the best treatment strategy guided by disease heterogeneity diagnostics [40]. The internet has facilitated participation of individual patients in the healthcare through sharing their experiences in blogs and other social media. Indeed, online patient communities represent a true model for participatory medicine. For instance, PatientsLikeMe [41] is an online patient community that provides robust methods for patients to share their experiences on effectiveness of treatments. Since such virtual environments act as repositories of self-reported data on the nature of disease, treatment response and side effects, they have the potential to be used for refining the existing drug-response models or optimizing clinical outcomes. For instance, Wicks et al. (2011) used the self-reported data of 348 ALS patients - posted on the PatientsLikeMe website - and based on these reports demonstrated that treatment with lithium carbonate has not been beneficial for the patients [42].

Together with blogs and online resources, electronic health records (EHRs) present valuable sources for integrating patient data into disease models. EHRs have been used to stratify neuropsychiatric patient cohorts based on complete phenotypic profiles than the conventional primary diagnosis [43]. Very recently, we have applied our Alzheimer's disease ontology (ADO) to 650 AD patients' health records and demonstrated that four major comorbidities reported in these records also link mechanistically to the pathology of AD [44]. There are emerging initiatives like the ‘Quantified Self’ initiative [45], which will lay the foundation for the concept of virtual patient. The virtual patient is a model that can be compared (and populated/parameterized) by individuals. Quantified Self initiative collects self-tracking tools and applications in one place and facilitates finding and access to self-recording/self-reporting application softwares.

In summary, the integration of medical data from participating patients into disease models is imperative if these models are to be used for patient stratification.

Emerging applications of integrative disease modeling in P4 medicine

With the increasing recognition that integrative systems modeling will play a major part in all future activities in P4 medicine, we envision that translational bioinformatics in general and systems modeling in particular offer more solutions to current challenges in P4 medicine. In the following, we provide an outlook on the emerging trend and future prospects in this area.

Molecular-based classification of diseases and their subtypes

The current classification of diseases is based on anatomical, symptomatic and epidemiological criteria, which does not take the etiological mechanism into account. This has led to the problem of misdiagnosis due to the overlap of symptoms. Consider Parkinson's disease with dementia and Alzheimer's disease dementia. Since classification of these two neurodegenerative diseases is merely based on their clinical signs, early diagnosis of Parkinson's disease (PD) is hampered. The extent of clinical overlap between these two conditions is so much greater than chance that even some authors propose that AD and PD belong to a spectrum of neurodegenerative disorders with common disease mechanism but triggered by different etiological factors [46]. However, at the molecular level, these two conditions can be distinguished based on diagnosis between synucleinopathy from tauopathy. Thus, integrative disease modeling provides a framework for incorporation of both conventional reductionism and non-reductionism approaches and allows for re-definition of the current nosology [47].

In 2011, National Academy of Sciences committee recommended NIH to develop a new taxonomy of diseases based on their underlying molecular and environmental causes rather than on physical signs and symptoms [48]. The idea is to create a so-called Knowledge Network information system that integrates molecular data, medical histories and health outcomes of individual patients (Figure 9).

Figure 9
figure 9

Architecture of the knowledge network information system proposed by NAS committee.

In 2012, the Innovative Medicines Initiative (IMI) - Europe's largest public-private partnership aiming to improve the drug development process - launched a call that addresses the topic of developing an aetiology-based taxonomy for human diseases. The proposed methodology for this topic that was submitted by Fraunhofer SCAI and named AETIONOMY, was selected by IMI consortium in which we propose a knowledge-based approach to modeling neurodegenerative disease mechanisms underlying Alzheimer's and Parkinson's disease and signature-based classification of syndromes belonging to these disorders. The aim of AETIONOMY is to come up with a new mechanism-based classification system for AD and PD so that mixed pathologies could be unequivocally distinguished, new features or classes could be easily accommodated in the taxonomy, and using this information for drug and biomarker discovery could be facilitated.

Predictive drug re-purposing and modeling polypharmacology effect

Low productivity of drug discovery pipelines in recent years has been largely attributed to insufficient efficacy of failed drugs. Failure of drugs in phases II and III clinical trials in fact reflects the poor understanding of the mode of action of such drugs at the molecular level. It is becoming increasingly evident that approved drugs have more than one target and this ‘multi-targeting’ nature of drugs could have both positive and negative consequences in terms of increased efficiency of therapeutic effect or adverse reactions, respectively. Since approved drugs have already passed safety tests in clinical trials, they can be re-used for new therapeutic indications; however, efficacy still has to be shown for the repositioned drug. At this point, the re-purposing strategy can be linked to efficacious targeted therapy for personalized medicine applications [49]. The recent approval of crizotinib for treatment of non-small cell lung cancer (NSCLC) that was repositioned from anaplastic large-cell lymphoma therapy is a good example of linking drug re-purposing to personalized medicine. The repositioned drug is accompanied by a diagnostic test for stratification of NSCLC patient subsets [50].

Usually, novel disease indications for known drugs have been discovered by serendipity. It is only recently that computational and systems approaches have been employed for inferring and predicting drug repurposing. These approaches can be grouped into target-based (similar binding sites) or disease-based (similar mechanisms) [51]. The rational design of ligands acting on multiple targets requires that such ligands bear certain molecular properties. We have recently devised a machine learning method that successfully classified approved drugs to NDD and non-NDD drugs based on a set of ten pharmacological properties [52]. Integration of drug pharmacological data and target structural features into tissue- or organ-specific disease models will support the systematic identification of new opportunities (‘re-purposing’) for existing drugs.

Integrative modeling in support of drug development decision making

The goal of integrative disease modeling is to provide a framework for hypothesis generation. Thus, it is expected that modeling predictions provide many inputs into the decision-making process in the pharmaceutical industry [53]. Since the main reasons for drug failures in the clinic are efficacy and safety, target identification and validation are the most important steps in drug development. Disease-specific computational models, when enhanced and enriched with efficacy data, can assist scientists to construct a more robust rationale in support of the main hypothesis compared to competing hypotheses. Based on this approach, a likelihood function can be developed and applied to proof-of-concept decision making so that the late-phase drug failures caused by lack of efficacy could be prevented [54]. Such an integrative model will summarize the evidence from multiple data sources that supports competing hypotheses on the basis of observed data at any time point.

Comparative modeling of rodent and human pathobiological mechanisms

High rate of drug failures in general and recent failures of Alzheimer's therapies in phase III of clinical trials in particular, despite substantial support from animal studies, raise the concern whether animal models of chronic complex diseases reliably inform human conditions. Several factors including insufficient statistical power, inadequate internal validity of animal experiments (methodological flaws) and lack of external validity (differences between animals and humans in the pathophysiology of the disease) have been reported to be responsible for the translational failure of apparently promising animal studies [55].

Reduced external validity of animal studies, including insufficient similarity to the human condition and ignorance of comorbidities, can partially explain the mechanistic issues behind the poor efficacy of failed drugs. For example, a systematic study evaluating the mimicry precision of human inflammation by mouse models showed that inflammatory responses in murine models poorly correlated with the human conditions [56].

In the case of chronic, progressive diseases such as neurodegenerative disorders, current research has mainly relied on injury-induced models that do not recapitulate the progressive, degenerative nature of the disease in humans. Reasons for this could be attributed to fundamental differences in neural circuitry between human and rodents as well as different molecular mechanisms in man from animal models [57]. The fact that mouse models of human neurodegenerative diseases have largely failed to predict the efficacy of clinical trials calls for more rigorous validation of predictions made with the help of mouse models using alternative strategies [58]. For instance, in response to the Species Translation Challenge by sbv IMPROVER [59], we sought to systematically model and mechanistically compare neuroinflammatory pathways between human and mouse in the context of Alzheimer's disease. Differential pathway analysis on both models revealed unique and segregating pathways in human and mouse.

Integrative disease modeling could offer a powerful approach to increasing the probability of predicting efficacy of animal data by providing comparative mouse-human computational models that compute the extent to which molecular details of animal models match with the corresponding molecular mechanisms in human subjects. Ideally, such a comparative study should take the multiple biological scales from cell and tissue types to genes, proteins and pathways into account so that it clearly reveals anatomic and mechanistic differences.


Integrative in silico disease modeling is emerging as a new paradigm at the core of translational research, which prepares the ground for transitioning from descriptive to mechanistic representation of disease processes. Given the tremendous potential of integrative disease modeling in supporting translation of biomarker and drug research into clinically applicable diagnostic, preventive, prognostic and therapeutic strategies, it is anticipated that computer-readable disease models will be an indispensable part of future efforts in the P4 medicine research area.