Introduction

During the 1970s and early 1980s, the primary focus of renal transplant research was the treatment of acute rejection (AR). A major breakthrough in the prevention of AR occurred with the introduction of cyclosporine A (CsA), first into experimental animal studies and then into clinical practice [1] in the late 1970s. However, the dramatic improvement in short-term transplant survival in the 1980s was not reflected in the half-life of kidneys after the first post-operative year [2], and so researchers turned their attention to the reasons for long-term kidney graft loss.

In a consensus conference on chronic rejection (a term that was later replaced with chronic allograft nephropathy) during the Fourth Alexis Carrel Conference held in Kiruna, Sweden in 1992, chronic kidney allograft rejection was defined as “gradual deterioration in graft function (during the course of at least three months post-transplantation) in the absence of any other disease and necessary graft histology to confirm the diagnosis” [3]. Chronic allograft nephropathy (CAN) is therefore a clinical and pathological description and not a diagnosis, and its pathogenesis must still be clarified. This review will examine quantitative histological and functional markers as early indicators (surrogate endpoints) of CAN, risk factors associated with chronic allograft pathology, and potential treatments.

Quantitative Histology and Functional Markers

The pathologic changes of CAN are myointimal proliferation of arteries and arterioles; interstitial fibrosis with relatively mild-to-moderate mononuclear cell infiltration; significant tubular atrophy and glomerular changes, including increased mesangial matrix, glomerular basement membrane, and Bowman capsular thickening; and segmental focal glomerulosclerosis [46]. The histological changes of chronic allograft pathology, however, are nonspecific [7]. Similar findings may be encountered in other conditions such as kidneys from old donors, hypertension, etc. Moreover, the histologic features of this pathology are also found in routine biopsies of kidney transplants, even with well-functioning grafts [8].

Although it is not clear to what extent the intensity of histological changes predict a chronic deterioration of allograft function, the typical histopathologic changes of chronic allograft pathology have been well-described by several investigators. In 1992, six quantitative histologic variables (interstitial inflammation, interstitial fibrosis, glomerular mesangial matrix expansion, glomerulosclerosis, vascular intimal proliferation, and tubular atrophy) were found to be significantly associated with decline of kidney function as measured by the level of serum creatinine in both humans [8] and a rat renal allograft model [9]. The Chronic Allograft Damage Index (CADI) was constructed from the sum of the scores for those lesions. The CADI was strongly correlated to the impairment of kidney transplant function both in rodents (r = 0.874) [10] and in human studies (r = 0.717) [11]. This association was further supported by other independent studies [6, 12] and confirmed in a protocol biopsy-controlled multicenter study, where the 12-month CADI score was found to correlate with three-year graft survival [13].

From these beginnings, attempts have been made to establish consensus criteria for histological changes associated with chronic allograft pathology. In fact, the popular term “chronic rejection” was replaced by CAN to acknowledge the lack of specificity of the histologic changes described in the original Banff classification in 1991 [14]. The Banff ‘91 criteria for grading overall intensity of chronic allograft injury included interstitial fibrosis and tubular atrophy as assessed by visual estimates, without any numerical quantitation. These criteria also omitted the glomerular and vascular changes [14] despite the fact that glomerular and vascular changes are characteristic of CAN both in humans [6, 8, 12] and experimental animals [9]. The Banff ‘97 classification [15] provided threshold values for chronic changes in renal allograft biopsies, which were largely adapted from the previous CADI thresholds, with minor alterations. In the Banff ’05 classification, the of Banff ‘97 CAN grading was replaced with a new name, “IF/TA” (“interstitial fibrosis and tubular atrophy, no evidence of any specific etiology”) [16]. These nonspecific terms, CAN or IF/TA, have been used to denote the morphological changes used as surrogate markers for chronic allograft injury.

Most studies have determined that tubular atrophy and interstitial fibrosis are strongly linked and synergistic, and thus interstitial fibrosis is sufficient as a predictor without the addition of tubular atrophy. Research has also found that isolated IF/TA is only weakly correlated with graft survival, whereas IF/TA in conjunction with other histopathological parameters, such as inflammation and vascular intimal thickening, is more strongly correlated with graft outcome [1720] suggesting that a combination of histological indicators may be superior to a limited number in predicting graft loss.

Clinical studies have revealed several clinical risk factors that are associated with chronic allograft injury [21]. Given the complex relationship between clinical risk factors and early histopathological changes, the composite and quantitative histological indices are best suited as surrogate markers for evaluating the status of the transplant [22]. As we are aware that transplant outcome is a situation where more than one pathway is critical, it is reasonable to assume that a potential surrogate marker may comprise several histological indices [23]. Therefore, the Banff ‘97 and Banff ‘05 gradings, with their current histological [15, 16] components for chronic allograft injury, may underestimate the severity of disease and could lead to inappropriate assessment of the degree of damage. In a study by our group [24], we assessed the prognostic values of two composite scores, CADI and the sum of the Banff Chronic score (ci + ct + cg + cv + mm), as well as CAN grading (mild, moderate, and severe) for predicting allograft dysfunction/graft loss in 318 consecutive 6-month surveillance biopsies. The CADI score outperformed the other variables in its ability to predict graft outcome.

A drawback to the use of early histological markers to predict chronic allograft pathology is that many centers are reluctant to perform transplant protocol biopsies, given the invasive nature of the procedure. Although the risk of significant complications following renal biopsy has decreased over the years, many transplant centers appear to use early functional biomarkers as a replacement for histopathology to predict chronic allograft dysfunction. Functional biomarkers such as serum creatinine [25] and estimated glomerular filtration rate (eGFR) [26] have the clear advantage of being inexpensive and noninvasive screening tests for chronic allograft dysfunction compared with transplant protocol biopsy. Unfortunately, the widespread adoption of serum creatinine and eGFR as surrogate markers for predicting graft outcome has occurred without rigorous evaluation. In fact, Kaplan and colleagues [27] demonstrated that serum creatinine at one year was a poor predictor of allograft loss at two years (area under the receiver operator characteristic curve [AUROC] =0.63). This result illustrates the scenario where a test result may have a good correlation with a clinical endpoint but fail as a surrogate marker. In a study from our institution [28], there was a relatively high prevalence of acute (interstitial inflammation, 30 %) and chronic (fibrosis, 50 %; tubular atrophy, 65 %; vascular intimal proliferation, 48 %; arteriolar hyalinosis, 47 %) histological changes seen in transplant protocol biopsies at 6 months from 280 patients. In contrast, renal function was relatively stable, and the average serum creatinine and eGFR levels were 122 ± 38 μmol/l and 58 ± 19.0 ml/min, respectively, which are considered normal for transplant patients. We tested the capacity of creatinine and eGFR for predicting histological changes using area under the receiver operator characteristic curve (AUROC). The values ranged from 0.51 (representing no discrimination) to 0.62, indicating that neither serum creatinine nor eGFR had predictive value for any of the individual histological parameters. Despite the obvious attractiveness of using functional markers as predictors or surrogates for chronic allograft dysfunction/graft loss, the adoption of serum creatinine and eGFR as surrogate or predictive markers of graft outcomes has been premature.

Although substantial progress has been made toward developing histological indices that will predict or act as surrogate markers for graft outcomes, more research is needed. Development of such indices may allow both earlier intervention and provide surrogate endpoints for drug development, thus reducing the time and cost involved.

Risk Factors

There are a number of both alloantigen-dependent and alloantigen-independent risk factors that appear to contribute to chronic allograft damage/chronic allograft nephropathy. CAN is a multifactorial syndrome and represents the cumulative response to injury, regardless of the type of injury [29, 21]. Considering the heterogeneity of these risk factors, chronic allograft damage probably results from a multitude of various interacting immunological and non-immunological factors with a similar final common pathway. Taken together, it appears that CAN is not a distinct entity with a unique pathogenesis, but is rather the end result of successive episodes of AR (some of which may have been self-limiting and thus not recognized), additional injury via various mechanisms of peri- and postoperative damage, and additional non-immunologic risk factors (similar to atherosclerosis risk factors).

Immunologic Risk Factors

One or more AR episodes are the strongest predictors for long-term graft survival [30, 31]. A single rejection episode after transplantation reduces the average half-life of first transplants from 45 to 25 years, and multiple rejection episodes decrease the half-life to five years [32].

The development of CAD in allogeneic grafts is a much more rapid process than is observed in syngeneic or isogeneic rat renal grafts, suggesting a role for an immune process in allogeneic grafts [33]. Our rat renal transplant model across strong histocompatibility barriers was entirely concordant: the more AR episodes, the less satisfactory the transplant function at 3 months and the more prominent the vascular and glomerular changes in transplant histology [34]. We also showed in the same animal model that there is a causal relationship between AR and chronic rejection and that the intensity and length of AR episodes, as well as their frequency, contribute to the development of chronic allograft damage [10].

Apart from frequency of AR, other alloantigen-dependent factors such as subclinical rejection [35•], antibody-mediated rejection [36], the presence of anti-HLA antibodies [37], vascular rejection [38], and HLA mismatches [39] show strong effects on the development of chronic allograft damage and thus long-term graft outcome.

Over the last 10 years, we have begun seeing major reductions in AR rates with the current immunosuppressive protocols. However, the reduced rates of rejection in two large series from Australia and New Zealand [38] and the U.S. [40] were not associated with significant improvement in graft survival, suggesting that factors other than AR are involved.

Renal Calcineurin Inhibitor (CNI) Toxicity

Calcineurin inhibitors (CNIs), such as CsA and tacrolimus, given as immunosuppressants to transplant patients have been credited for the drastic reduction in AR over the last 30 years. Since the early 1980s, CNI nephrotoxicity has been implicated as one of the major causes of chronic allograft dysfunction and late renal allograft loss, thought to be due to irreversible histological damage. More recently, however, this determination has come into question.

During the 1980s, when high doses of CsA were used, isometric vacuolization, tubular calcification, giant mitochondria, and the focal or striped form of interstitial fibrosis were considered to be diagnostic indicators of CsA nephrotoxicity [41]. More recently, hyaline arteriolar thickening has been the only morphological indicator found to distinguish patients with CNI toxicity [42, 4].Hyaline arteriolar thickening, however, is not specific to CNI toxicity. It was demonstrated in 839 non-transplant Japanese autopsy subjects that the frequency of arterial hyalinosis and arteriosclerosis increased linearly with advancing age and was associated with higher blood pressure, higher total cholesterol level, higher serum creatinine level, lower glomerular filtration rate, and proteinuria [43]. In our recently completed multicenter clinical trial with central biopsy reading, more than 400 baseline-transplant biopsies showed 34 % arteriolar hyalinosis (unpublished data). It is becoming increasingly clear that arteriolar hyalinosis is linked to a multitude of underlying conditions in addition to chronic CNI use.

One difficulty in understanding the impact of CNI treatment on kidney pathology is that many studies in this area do not have appropriate comparators. In one of the earliest reports linking CNIs to native renal damage, which studied heart transplant recipients with and without CsA treatment [44, 45], the study did not include an analysis of the biopsy findings in the non-CsA group, the sample size for the renal hemodynamic studies was very small, and the patient baseline characteristics were neither defined in detail nor stratified. Subsequent studies of heart [46, 47] and liver [48, 49] transplant patients reported striped interstitial fibrosis, nodular arterial hyalinosis, and glomerulosclerosis in renal biopsies of non-renal transplant recipients, which were attributed to CNI toxicity. However, when we examined the patient populations in those reports, we found that the majority of patients had preexisting conditions such as hepatitis C, chronic kidney disease, diabetes, and hypertension. Additionally, one frequently cited study suggesting CNI toxicity did not include a non-CNI-treated control group [50]; moreover, the 10-year death-censored graft survival for their patient population was excellent (95 %) with an immunosuppressant regimen that included CNI. Finally, a recent study found that the prevalence of arteriolar hyalinosis at five years was the same whether or not the patients were treated with CNI [51•]. Taken together, it is difficult to determine whether the histological lesions seen in CNI-treated transplant patients are a direct result of drug treatment or are secondary to other pathological processes.

Despite the concern that CNI nephrotoxicity may be responsible for a significant proportion of late graft dysfunction, most studies have shown that CNI-free immunosuppression provides no long-term benefit [52]. Based on scientific registry data, 6-month AR rates were significantly higher with sirolimus in combination with mycophenolate mofetil versus other regimes including CNI. Overall graft survival was significantly lower on sirolimus/mycophenolate mofetil relative to tacrolimus/mycophenolate mofetil or CsA/mycophenolate mofetil [53]. In a study of 566 patients who received kidney transplants, the frequency of chronic rejection was higher in patients who received less than 5 mg of CsA per kilogram per day [30]. Furthermore, patients who received higher doses of CsA had better renal function for the first 3 years after transplantation [54], and patients who had lower exposure to tacrolimus had higher chronic histological damage [55], refuting the posited chronic nephrotoxic effects of CNIs.

Two recent comprehensive reviews [56••, 57••] argue that CNI toxicity has been overstated. The current trend toward CNI avoidance, elimination, or minimization has not been based on strong evidence that CNIs can cause progressive nephropathy. While CNIs may play some role in potentiating chronic kidney injury in solid-organ transplantation, the extent of their contribution to chronic allograft injury is unknown. Moreover, the increased immunologic activity due to minimization or avoidance of CNIs may pose a greater danger to the function of renal allografts than CNI nephropathy. CNIs remain the backbone of current immunosuppression regimens and have been associated with excellent graft survival rates.

Ischemia/Reperfusion Injury

Delayed graft function and prolonged ischemia is associated with overall poor graft survival compared to grafts with immediate function in both humans [58] and animal models [59]. As a result of non-specific ischemia/reperfusion injury, the biologically uninjured transplant transforms into an inflamed-injured organ that appears to facilitate subsequent specific immunologic events; intragraft allorecognition and alloactivation by/of host T-lymphocytes result in progressive fibrosis of renal tissue [6062].

Donor and Recipient Age

Very early studies identified donor age as a risk factor for short- and long-term graft survival. More recently, donor age was significantly associated with interstitial fibrosis, tubular atrophy, and glomerulosclerosis in implant biopsies. Additionally, among donor demographics, only donor age was found to be a significant predictor of long-term death-censored graft survival [63]. In our study, which correlated a number of clinical risk factors with histologic evidence of chronic allograft damage in one-year protocol biopsies, older donor age and younger recipient age were associated with both higher CADI score and, to varying extents, individual histopathological indices [22].

Other studies confirm that younger recipient age is a risk factor for graft survival. When all deaths were censored, recipient age > 50 was associated with significantly better graft survival than age < 50 [64], and better survival of renal transplants in older recipients was also demonstrated [64]. With increasing age, there was a diminishing percentage of graft loss due to immunologic causes [65]. The protective effect of higher recipient age may be related to decreased immune responsiveness and thus decreased graft loss due to immunological causes. These age-related changes in the AR rate, especially when all deaths are censored, may explain why younger recipients have shorter long-term graft survival.

Hypertension

The precise role of hypertension in chronic allograft damage has been difficult to characterize. The cause-effect relationship between hypertension and deteriorating allograft function is confounded by “reverse causation,” i.e., impaired renal function leading to hypertension. Virtually all patients with renal disease are hypertensive by the time end-stage renal disease is reached. Although blood pressure may be well-controlled by dialysis, hypertension after renal transplantation is extremely common, occurring in 80 % of CsA-treated patients [66]. Poorly controlled blood pressure beyond one year after transplantation is associated with poor allograft survival [66].

In experimental rat aortic allografts, hypertension was associated with a significant increase in intimal thickness involving both smooth muscle cell proliferation and collagen secretion. The angiotensin-converting-enzyme inhibitor perindopril was able to decrease the systolic blood pressure by 30 % and concomitantly reduce intimal thickness by 40 % [67]. It has been observed that antihypertensive drugs decrease the extent of glomerular mesangiolysis and glomerulosclerosis in experimental rat renal transplants [68].

Hyperlipidemia

Hypercholesterolemia has been implicated in the pathogenesis of chronic allograft damage due to histological similarities between CAD and atherosclerosis. In an animal experimental model of CAD, increased cholesterol intake was found to aggravate the vascular lesions [69]. In addition, clinical studies have reported an association between lipid abnormalities and several histopathological indicators of chronic renal allograft damage [70, 71]. The presence of metabolic syndrome was associated with impaired renal allograft function beyond one year post-transplant. Among its components, only systolic blood pressure and hypertriglyceridemia were independently associated with impaired renal allograft function beyond one year post-transplant [72].

Glomerular Hyperfiltration

Glomerular hyperfiltration has been identified as a risk factor [73]. Consistent with findings in non-transplant renal disease, loss of renal mass due to immunological damage and/or non-immunologically mediated hemodynamic factors [74] lead to an increase in glomerular capillary blood flow and capillary pressure [75], and hyperfiltration in the remaining nephrons accelerates renal injury [76]. At this point, renal functional deterioration becomes a self-propagating process.

Cytomegalovirus Infection

Although there is no clear-cut evidence of the role of cytomegalovirus (CMV) infection in CAD in humans, clinical studies have shown an association between CMV infection and arteriosclerosis in heart transplants [77]. Studies in rats showed that CMV infection enhanced CAN in kidney allografts [78] and arteriosclerosis in aortic allografts [79]. In the case of rat kidney allografts, CMV infection significantly increased chronic histological lesions, which are associated with the number of AR episodes, interstitial inflammation, and the expression of intercellular adhesion molecule-1 on vascular endothelial and tubular epithelial cells [78]. When rats receiving aortic allografts were treated with ganciclovir, the enhanced intimal response was entirely eradicated [80]. Similarly, lower rates of AR were reported at six months in a group receiving valacyclovir prophylaxis compared with placebo, primarily in donor-positive, recipient-negative sero-pairs [81].

The temporal relationship between CMV infection and AR is not clearly delineated. It is possible that CMV infection may increase the risk of AR, as suggested in the studies mentioned above; conversely, the enhanced immunosuppression during the treatment of AR may predispose the recipient to CMV infection. CMV may enhance chronic rejection by the enhancement of proinflammatory and inflammatory responses, which may or may not act in concert with acute cell-mediated rejection. This hypothesis has yet to be tested.

Treatments

Animal studies [82, 83] clearly demonstrate that, once triggered, chronic rejection becomes independent of the causative event and proceeds on its own [29]. This may also be the case for human transplantation. Since no effective therapy is currently available, CAN inevitably progresses to end-stage kidney disease. In kidney transplantation, it has been known since the early 1990s that recipients having AR episodes are at increased risk for graft loss [30]. Conversely, in the absence of an AR episode, graft survival is excellent [31]. Since prevention is clearly better than cure, optimizing the immunosuppressive regimen in organ transplantation may be the optimal prophylactic strategy today, and other pharmacological agents also show promise as adjuncts to immunotherapy. Whereas avoiding or mitigating recognized clinical risk factors are also good strategies for improving graft survival, it is clear that there are other risk factors for late graft loss that have yet to be defined.

Immunosuppressive Regimens

Potent immunosuppressive combinations to reduce inflammation, diagnosed as either subclinical [84] or clinical rejection [31], result in better graft survival, as alloimmunity remains the most common mechanism leading to graft failure [85••]. As mentioned earlier, treatment with CNIs such as cyclosporine and tacrolimus has revolutionized the area of transplantation medicine, resulting in the reduction of AR events. However, several researchers are of the view that this early survival gain is countered by perceived chronic nephrotoxicity. Regardless of whether chronic nephrotoxicity is overstated, as is the view of some authors [56••, 57••], it is clear that further research is needed, and drug regimens that balance the AR-preventing effect of CNIs while minimizing dose requirement are being developed [52].

Mycophenolate mofetil, which inhibits the proliferation of T and B lymphocytes, is commonly used in combination with CNIs. A systematic review of 19 studies – including a total of 3,143 patients [86] – and a single-center study [87] showed that this drug conferred a clinical benefit and reduced chronic histological changes, respectively.

Sirolimus is another immunosuppressant and vasculoprotective agent with promising albeit not yet definitive benefits. Sirolimus is a relatively weak immunosuppressive compound [29], however, and the high doses required may expose the patient to perioperative complications such as wound dehiscence, delayed graft function, and lymphocele. Indeed, these are factors that have prompted the discontinuation of sirolimus in many cases. Based on results of recent clinical trials in the de novo transplant setting, triple therapy with sirolimus, mycophenolate mofetil, and corticosteroids is not recommended [52], as CNIs have been associated with major improvements in early rejection outcomes in the acute phase of the transplant. However, in patients with normal/stable kidney function after 3–6 months post-transplant, sirolimus-based regimens can be used for the agent’s vasculoprotective effects to prevent further organ scarring [88]. In our institution, we perform surveillance biopsies in patients with normal/stable kidney function at around six months post-transplantation. If the biopsy shows chronic histologic changes, sirolimus is added to the therapy, replacing MMF or CNI [88].

Other Pharmacological Agents

Various pharmacological agents have shown some promise in slowing the progression of renal allograft failure in animal studies of chronic rejection. These include prostaglandin inhibitors [89], the somatostatin analog angiopeptin [90, 91], low-molecular-weight heparin in combination with CsA [92], and antihypertensive drugs [68, 93]. The promising results, however, have not been reproduced in humans.

In a retrospective study, the use of aspirin as an adjuvant to current post-transplant treatment regimens was found to have functional benefits and improved allograft survival [94].

Retrospective studies measuring the effects of angiotensin-converting-enzyme (ACE) inhibitors or angiotensin receptor blockers (ARBs) in kidney transplant recipients reported conflicting results on allograft and patient survival. In an analysis of 17,209 renal and 1,744 heart transplant recipients, the use of ACE inhibitors/ARBs was not associated with an improvement in patient or transplant survival [95]. In contrast, another retrospective study of 2,031 patients found that the use of such agents was associated with improved long-term graft and patient survival [96]. A systematic review of randomized trials (21 trials with 1,549 patients) to determine the effect of the use of ACE-inhibitor or ARB following kidney transplantation concluded that there were insufficient data to determine the effect on patient or graft survival [97].

Dietary protein restriction has also been shown to improve the course of renal failure in chronic kidney allograft rejection [98], as it has on patients with chronic renal impairment of native kidney diseases [99].

Clinical Risk Factors

In addition to immunosuppressive protocols, avoiding clinical risk factors is a sound strategy for maximizing graft viability. This includes avoiding sensitization, optimizing HLA matching, taking into account cold ischemia time, minimizing ischemic injury, and matching recipient and donor age. As the average age of both donors and recipients and is rising, older donors should be allocated to older recipients in order to match the recipient’s shorter life expectancy. Conversely, younger recipients should receive organs from younger donors, who would have a larger nephron mass compared with older donors.

Despite our deepened understanding of clinical risk factors, it is clear that there are other determinants of graft loss that have yet to be identified. When we attempted to correlate a number of clinical risk factors with histologic evidence of CAN on protocol biopsies, we found that only 60 % of the variability in chronic allograft damage could be explained by clinical risk factors [22]. Therefore, other “missing” factors must play a role in the pathogenesis of CAN.

Conclusions

The significant progress over the last two decades in renal transplantation has been driven primarily by improvements in short-term graft survival, Unfortunately, there has been little parallel improvement in long-term outcomes [100]. Although the development of better histological scores are improving efforts to detect early signs of chronic allograft pathology by histological indicators, therapies to rescue declining graft function are limited and warrant further efforts.