Recognition of an Inadequate Rate of Discovery of New Therapeutic Agents

Across medical disciplines, we have been presented with a challenge that is both disheartening and provocative. Recognition of a lack of tangible, widespread progress has been openly admitted and graphically portrayed as the “Valley of Death” [1]. Despite an avalanche of data emanating from powerful new technology, the production rate of new effective medications has been steadily dropping. Although the Valley concept took root from cancer research, neurodegenerative research is no exception. The major Food and Drug Association (FDA)-approved therapeutic agents for Alzheimer’s disease and Parkinson’s disease are based on work done three to four decades ago. Current FDA-approved therapies for the primary symptoms of AD are still largely restricted to cholinergic replacement, an approach that was suggested by human brain tissue studies in the 1970s [2,3,4,5]. The only other approach, directed at blocking glutamatergic excitotoxicity (memantine/Namenda), was approved by the FDA in 2003, but the first description of excitotoxicity had been in 1957 [6], and its application to AD was first envisioned in the early 1980s. Similarly, for PD, dopaminergic replacement therapy is still the mainstay, more than 50 years after the discovery of the dopaminergic deficit [7]. Immunotherapy and agents targeting Aβ secretases are undergoing intensive clinical testing but have so far failed to produce convincing results.

Now, our most urgent question has become, “Where have we gone wrong?” The Valley of Death has been variously portrayed as a chasm between biomedical researchers and patients, or between basic science researchers in academia and applied science researchers in the pharmaceutical industry, or between basic science ideas and the hurdles they must cross to regulatory approval. What is clear is that the translation of basic science discoveries to drugs with obvious real-world benefits is becoming less, rather than more, frequent than anyone is comfortable with. Multiple reasons for the gap have been advanced, including the increasingly complex route to FDA approval, regulatory changes leading to a dramatic cost escalation of large clinical trials, the increasing dependence on for-profit corporate involvement, and the realization that many diseases are etiologically heterogeneous [8,9,10].

Another possible cause has received much less attention but may be much more important. In 2002, the National Dialogue on Cancer, convened to understand why the “war on cancer” was falling short of expectations, concluded that, of the ten most important roadblocks to finding cures for cancer, the single most critical one was inadequate availability of “high-quality, highly characterized human tissues for translational research”, and this has also been advanced as an obstacle to finding new therapeutic agents for AD and PD [11]. The comparison of appropriate diseased and control tissue has long been the most direct and most productive route to discovery of new, disease-relevant molecular pathways.

This article is based on previously conducted studies and does not involve any new studies of human or animal subjects performed by the author.

Are We Treating Too Late?

It has long been known, based on autopsy findings of “incidental” plaques, tangles and Lewy bodies in clinically-normal older people, that subjects with AD and PD must spend decades in a preclinical stage. In the past 10 years, this concept has become increasingly accepted, and characterization of the preclinical state has been seen as a crucial goal. Models of the trajectory of AD through asymptomatic, mild impairment and overt dementia have progressed from initial autopsy-based work [12,13,14] to those based on biomarkers measured in vivo [15, 16]. Combined with the repeated failure of clinical trials of subjects with AD dementia, the increased awareness and acceptance of a continuum between preclinical and manifest disease stages prompted the conclusion that therapy for AD might have a much higher chance for success if it was aimed at slowing or halting progression in the preclinical stage.

In 2011, three consensus panels, jointly sponsored by the National Institute on Aging and the Alzheimer’s Association, recommended that AD be defined and studied in terms of three clinically-defined stages, preclinical, mild cognitive impairment and dementia [17,18,19]. This was a paradigm shift in that previous expert clinician panels had always considered dementia as an essential part of the definition of AD, while leaving the existence of a preclinical stage in limbo. The panel that most directly articulated this paradigm shift was that reported on by Sperling et al. [19]. This panel emphasized that AD is likely best treated when subjects are in a presymptomatic or preclinical phase, before large-scale neurodegeneration has occurred, and that the definition of this needs to be addressed using evidence from biomarker, epidemiological, and neuropsychological studies. To formalize this shift in thinking, the panel suggested that a re-definition of AD is necessary. Whereas formerly the definition of AD included the presence of dementia, it would be more advantageous and more accurate to realize that AD is a continuous and gradual pathophysiological disease process, with a long asymptomatic period during which biomarkers might track an individual’s otherwise unseen disease progression. However, it was acknowledged that at the present time insufficient longitudinal data exist, preventing an accurate prediction of clinical course from biomarker measures. Ultimately, the concerted results from prospective studies, with measures and interpretation done within a standardized framework encompassing definitions and methodology, are what are hoped to identify biomarker profiles of subjects most likely to benefit from disease-modifying therapies.

Biomarkers as Accelerants for Clinical Trials

The shift towards trying to prevent AD dementia from happening, rather than trying to cure it after it occurs, had its roots decades ago, with many authors suggesting that AD, like myocardial and cerebral infarctions, might be prevented by appropriate therapy. Again, following this model, it has long been apparent that the measurement of appropriate biomarkers, i.e. blood pressure and blood lipid profile, have been essential to the prevention of heart attack and stroke, and that, with similarly-effective biomarkers, those at risk for AD dementia could be identified and treated prior to any overt signs or symptoms.

At this point, it is useful to briefly review the definition and evolution of biomarkers. A search of the US National Library of Medicine (PubMed) shows that the earliest appearance of the term in the title of a publication was in 1980, referring to a blood serum biomarker of breast cancer [20]. Shortly afterwards, the advantages of using biomarkers in trials of cancer [21] and heart disease [22] were widely discussed [23]. By 2001, in order to clarify developing ambiguity, the term was given a consensus definition by an NIH working group [24], stated as follows: “Biological marker (biomarker): A characteristic that is objectively measured and evaluated as an indicator of normal biological processes, pathogenic processes, or pharmacological responses to a therapeutic intervention”. Additionally, several possible uses of biomarkers were suggested, including (1) in early-phase clinical trials to establish “proof-of-concept”, (2) as a diagnostic tool for those subjects with a disease or an abnormal condition, (3) for staging of disease or classification of the extent of disease, (4) as an indicator of disease prognosis, and (5) for the prediction and monitoring of clinical response to an intervention. Perhaps the most important usage of biomarkers, as a “surrogate endpoint” in clinical trials, was extensively discussed. The usual and definitive proof of the efficacy of a medication, in the opinion of regulatory agencies such as the US FDA, has always been, and will continue to be, “clinical endpoints” that are measures of how a patient “feels, functions or survives”. Common clinical endpoints include myocardial infarction, stroke, cancer recurrence and death. A biomarker would qualify as a surrogate endpoint if it predicted clinical benefit. A major advantage of surrogate biomarkers as compared to clinical endpoints is that they are often less expensive, easier to measure and more precise [25].

With respect to AD, PD and other neurodegenerative diseases, a previously unrecognized issue has been the relatively low diagnostic accuracy for both of these conditions, particularly at early stages, using only standard neurological examination and investigation [26,27,28,29]. It is very likely that many clinical trials done over the past few decades have been underpowered because this was not factored into the predicted effect size, and the major reason for Phase 2 clinical trial failure has been low effect size [30]. It is expected that biomarkers will improve diagnostic accuracy and that their deployment to assist trial subject selection will greatly improve trial efficiency [31].

The clinical endpoints have been progression of cognitive impairment or motor dysfunction. These have been suboptimal for rapidly assessing new candidate drugs, as they are nonspecific syndromes that are produced by multiple diseases with differing molecular etiologies. Currently, autopsy studies have shown that most elderly subjects have more than one brain pathology present [32], and that most of the variability in cognitive performance is not accounted for, even by the summation of common histological lesions [33]. Additionally, the rates of change in functional measures are slow, necessitating long and costly clinical trials. These shortcomings are particularly critical for prevention trials, where clinical endpoints might take more than a decade to appear. Efforts should be made, using available clinicopathological data, to identify functional measures that correlate better with the responsible pathological lesions, giving greater disease specificity.

Neuropathological Validation is Critical

A variety of biomarkers are currently being deployed in AD prevention and/or treatment trials, including cerebrospinal fluid (CSF) Aβ, tau, phosphorylated tau and other neuronal proteins, PET tracers for Aβ, tau and glucose uptake, and MR measures of brain atrophy [34,35,36,37]. A serious shortcoming for many of these, however, is the lack of validation by neuropathology at autopsy. While the FDA has made autopsy validation a requirement for approval of new PET agents, other imaging modalities as well as biofluid markers have not been forced to meet this standard, and as a result their true accuracy is unknown as they have been validated mostly against the relatively inaccurate clinical diagnosis, or against other unvalidated biomarkers. A second-best approach is to validate new biomarkers against an autopsy-validated biomarker, as has been done by comparing CSF AD biomarkers to PET amyloid imaging [38,39,40]. There are diminishing returns to this strategy, however, as, if we assume an average 85% sensitivity and specificity of PET amyloid for neuropathologically-confirmed AD [31], then a biomarker that achieves 80% sensitivity and specificity against PET amyloid might identify only 68% of AD subjects correctly (0.85 × 0.80).

Success Story with Limitations: PET Amyloid as a Biomarker for AD

PET imaging of brain amyloid deposits has gained wide acceptance as the best modality for this role, with the confidence being primarily due to autopsy validation of the three currently-approved agents [41,42,43]. Although not explicitly approved for diagnosing dementia due to AD, its usage probably increases clinical diagnostic accuracy by at least 10% points [31, 44]. However, it has important limitations. None of the approved agents detects neuropathologically “sparse” neuritic plaque densities, and so amyloid PET may not be the optimal method to select subjects for prevention studies based on anti-amyloid therapeutics, as only those with the relatively heavier “moderate” and “frequent’ neuritic plaque densities are amyloid PET-positive. On the other hand, subjects with AD dementia have usually reached a plateau in plaque density, which may limit the ability of an anti-amyloid agent to reduce PET-demonstrated plaque loads. Additionally, until now, amyloid PET has provided mostly a “yes–no” answer to plaque presence and is not useful for disease staging, although SUVr quantitation could be employed and striatal PET amyloid deposits may demonstrate two stages of plaque regional spread, mirroring Thal histopathological plaque staging [45,46,47]. Perhaps the optimal usage of amyloid PET is for selecting amyloid-positive subjects for secondary prevention or mild cognitive impairment trials in order to exclude subjects without amyloid. The experience with the phase III solanezumab and bapineuzumab trials indicated that about 25% of included subjects were amyloid PET-negative, despite being selected with the traditional neurological criteria from the pre-amyloid imaging era [48]. Perhaps most disappointing has been the disconnect between plaque load reduction and cognitive improvement [49], thus far denying amyloid PET true surrogate status, although the Phase 1b aducanamab trial of 125 subjects offers hope of achieving this in subsequent trials [50].

The Near and Far Future

PET imaging for the demonstration of tau pathology in AD, although so far not validated by autopsy (an autopsy-validated study is currently being conducted by Avid), has the great advantage, over any other AD biomarkers, of potentially being able to provide, as would striatal PET amyloid, pathologically-based clinical staging of AD. Flortaucipir (formerly AV-1451) has been shown by several groups [51,52,53] to demonstrate cognitively-correlated cortical patterns similar to those originally described by Braak [54]. As a surrogate biomarker for clinical AD, Braak staging with tau PET would potentially be the best yet, as autopsy studies have shown that regional brain tau pathology is the strongest correlate of cognitive impairment and disease progression [55, 56]. Yet some limitations may be foreseen, including a relatively low sensitivity to non-AD tauopathy as well as low-abundance tau pathology, pretangles and other “immature” tau pathology [57,58,59].

Nascent biofluids methods currently impressing include high-sensitivity assays such as MagQu’s detection system utilizing magnetic nanoparticles [60, 61] and “seeding” and “quaking” assays that take advantage of the prion-like properties of pathological tau and alpha-synuclein [62,63,64]. Early reports of extremely high sensitivity and specificity have to be tempered against the reality that clinical diagnosis has so far been the only gold standard, reducing accuracies of 90% to less-stellar levels.

Biomarkers for the synucleinopathies have been relatively underperforming as there is as yet no PET agent [65], while CSF alpha-synuclein concentrations broadly overlap with normal [66, 67]. A concerted effort has been made by the Michael J. Fox Foundation (MJFF) to spur the development of a PET diagnostic and a peripheral tissue synuclein-based biomarker. A multi-center group sponsored by the MJFF has provided considerable clarification to the latter field, demonstrating with autopsy-validated tissue that discrepancies in earlier work have been due to idiosyncracies of single-center studies, inadequate definition of disease-specific staining patterns and insufficient training of histopathology readers [68,69,70]. A current Phase IV observational study, “S4”, is being sponsored by the MJFF and will determine the sensitivity and specificity of immunohistochemical staining for pathological alpha-synuclein in colonic, submandibular gland and skin biopsies from 60 PD subjects and 20 controls [71].

Genotype is expected by many to be an especially powerful biomarker, not only in autosomal dominantly-inherited forms of disease and individually strong risk factors but also in terms of composite possession of up to dozens of weaker risk and protective alleles [72]. In large genome-wide studies, the inaccuracy of the clinical diagnostic gold standard has been overcome by huge sample size, but it has been recognized that genomic studies utilizing subjects with a neuropathological diagnosis, because of the more precise molecular phenotyping, may be done with much smaller subject numbers and with a much richer payout in terms of clinicopathological correlations [73, 74]. Genome-wide association studies have been superb at revealing molecular risk factors, but at present it seems that, at least for neurodegenerative disease, genomic analysis may continue to lack the precision required for individualized medicine.

The eventual bogeyman to contend with, however, will be molecular neuropathological complexity, as neurodegenerative diseases most commonly occur in groups rather than as solo performers [32, 75,76,77,78], and this heterogeneity of pathology has also been demonstrated in subjects enrolled in diagnostic trials [79, 80]. Such comorbidity increases with extreme old age, but is still common in subjects in their 70s. Successfully detecting and treating one disease may not be completely effective when there are two, three or even four present.

In summary, biomarkers for neurodegenerative disease have come a long way in the past 10 years, but validation against autopsy still occurs too infrequently. Useful single-pathology methods are having an impact on clinical trials and in limited clinical settings, but combinations of pathology are likely to forestall them from reaching their full potential in the near future.