Introduction

Heart transplantation (HT) is the treatment of choice for end-stage heart disease in the absence of contraindications [1, 2]. However, despite advances in the criteria for selection of donors and recipients and in immunosuppression, rejection is still a significant problem. Acute rejection kills 11 % of patients in the first 3 years post-HT, and injury due to immune system activity (whether acute or chronic) is an important contributor to cardiac allograft vasculopathy (CAV) and unexplained chronic graft failure, the main causes of death in the long term [3]. Other major causes of post-HT morbidity and mortality are infections, malignancies and nephrotoxicity [3], all of which are related to immunosuppression. Thus, both overimmunosuppression and underimmunosuppression can have lethal consequences after HT, but tailoring immunosuppression to the individual patient is not easy: in principle, it requires knowledge of the state of his or her immune system, the state of the graft and the patient’s susceptibility to changes in immunosuppressive therapy.

An ideal strategy for monitoring the effective level of immunosuppressive intervention and the state and cardiac effects of the immune system would be non-invasive, reliably allow discrimination between the presence and absence of rejection and detect states of overimmunosuppression. No such strategy currently exists. The best currently achievable approximation might combine (i) drug-level monitoring, (ii) assessment of endomyocardial biopsies (EMBs) both histologically and immunopathologically (including evaluation of markers such as C3d, C4d and CD68) [4, 5], (iii) graft function tests (imaging) and assays (B-type natriuretic peptide) [6], (iv) immune system monitoring (gene expression profiling (GEP) [7], T cell function tests [8] and monitoring of donor-specific antibodies) [9] and possibly (v) other promising biomarkers for rejection, such as circulating donor-derived cell-free DNA [10, 11] or micro-RNA [12, 13] (Table 1).

Table 1 Immune monitoring tools that have been investigated in heart transplant recipients

Current Monitoring Tools and Novel Strategies

Early diagnosis of rejection is one of the still-unsolved problems in HT. The diagnostic procedure currently regarded as the “gold standard” (though employed in accordance with different schedules by different centres) is evaluation of EMBs [4, 5], but this procedure is invasive (requiring intravenous catheterization), has poor sensitivity because of the necessary limitations on tissue sampling and suffers from significant inter-observer variability [14]. The development of a sensitive, objective, non-invasive technique for detection of allograft rejection is therefore a major goal for HT. Below, we indicate a number of approaches based on the measurement of putative biomarkers and predictors of rejection that have been proposed in recent years.

Gene Expression Profiling

Gene expression profiling (GEP) is a non-invasive technique aimed at assessing the risk of acute rejection by evaluation of immune system activation and leukocyte trafficking by measurement of the expression of relevant genes in peripheral blood mononuclear cells (PBMC) [7]. Its current commercial implementation is the AlloMap® molecular expression test (CareDx, Inc., Brisbane, CA), which was approved by the Food and Drug Administration (FDA) in 2008. This test assays the expression of 20 genes: 11 reflecting the probability of acute rejection and 9 for normalization and quality control [7]. An algorithm weighting the contribution of each gene affords a score in the range 0–40, low scores being associated with a low likelihood of moderate or severe cardiac allograft rejection (grade ≥3A according to the original ISHLT classification [15], grade ≥2R in the revised version [4]).

The AlloMap® test was developed in the Cardiac Allograft Rejection Gene Expression Observation (CARGO) study, a multicentre observational study of 629 HT patients at eight US centres [7]. The main objective of CARGO was to develop and evaluate a GEP test to discriminate between a quiescent state (ISHLT EMB grade 0R) and moderate or severe rejection (ISHLT EMB grade ≥2R). CARGO was carried out in three phases: (1) identification of a population of candidate genes using a combination of focused genomic and knowledge-based approaches, (2) selection of a panel of optimally informative genes using PCR assays and statistical methods and (3) validation. Validation was performed using 32 quiescent patients and 31 with acute rejection, all monitored prospectively by both GEP and EMB evaluation at the post-HT follow-up visits. All EMBs were graded by a local pathologist and by three independent core pathologists who were blind to the patients’ clinical data [7]. The rejection group had a significantly higher mean test score than the quiescent group (p = 0.0018), and a pre-defined score threshold of 20 correctly classified 84 % of the rejection group. However, scores increased with time post-HT (in association with the reduction of steroid doses that is generally effected in the first year). The main strength of this test at times greater than 1-year post-HT is claimed to be its high negative predictive value (NPV) for moderate or severe rejection, 99.6 % with a score threshold of 30. Although the positive predictive value of the test is low, 6.8 %, the high NPV may allow EMB to be dispensed with when the test is negative and there are no other indications of possible rejection either.

Since the main CARGO report appeared [7], there have been relevant papers by Bernstein et al. [16], Starling et al. [17] and Mehra et al. [18]. Bernstein et al. [16], reanalyzed CARGO data to investigate in greater detail the CARGO finding that test scores corresponding to grade 1B EMBs were similar to those associated with grades ≥3A. This was confirmed, as were the significant average score differences between grades ≥3A (32 ± 0.9) and grades 0 (25.3 ± 0.5), 1A (23.8 ± 2.1) and 2 (26.9 ± 1.5) (p = 0.00001, 0.001 and 0.01, respectively). The similarity between results for grade 1B and for grades ≥3A, which was seen not only in test scores but also in gene expression profiles, is in keeping with the difficulty, noted by Stewart et al. [4], in distinguishing histologically between grade 1B and grade 3B EMBs.

In keeping with the above-noted time dependence of AlloMap® scores and on the basis of early experience of its clinical use, Starling et al. [17] recommended that rejection surveillance with this test employs different thresholds for patients in different follow-up periods: 20 for months 3–6, 30 for months 7–12 and 34 for >12 months post-transplant. The protocol that they used at follow-up visits started with history taking, physical examination and assessment of graft function by echocardiography. If these procedures suggested rejection, heart failure or impaired left ventricular function, EMB was performed; if not, an AlloMap® test was carried out and EMB was performed only if the appropriate score threshold was exceeded.

Mehra et al. [18] carried out a nested case-control study of 104 HT patients enrolled in the CARGO study, all with a baseline EMB grade of 0 or 1A. The baseline AlloMap® scores of “controls” who subsequently remained rejection-free for 12 weeks were significantly lower than those of patients who suffered grade ≥3A rejection within 12 weeks, 23.9 ± 7.1 as against 27.4 ± 6.3 (p = 0.01), with a wider gap between non-rejectors and rejectors in the first 6 months post-HT (22.4 ± 7.5 vs 28.4 ± 4.9, p < 0.001). In this latter period, no scores from patients who went on to reject within 12 weeks were <20. Furthermore, mRNA measurements of genes not included in the AlloMap® panel suggested that pathways regulating T cell homeostasis and corticosteroid sensitivity are associated with future acute cellular rejection and are measurably active before rejection becomes histologically detectable.

Direct comparison of AlloMap-based monitoring with EMB-based monitoring among patients with low risk of rejection (clinically stable and with a left ventricular ejection fraction (LVEF) ≥45 %) was undertaken in the Invasive Monitoring Attenuation through Gene Expression trial (IMAGE) [19]. This trial randomized 602 patients who were between 6 months and 5 years post-HT at baseline to either AlloMap® or EMB monitoring (in both cases, together with clinical examination and echocardiography); median follow-up was 19 months. The 2-year cumulative incidence of the composite primary endpoint (rejection with haemodynamic compromise, graft dysfunction due to other causes, death or retransplantation) was no greater among AlloMap-monitored patients than among EMB-monitored patients (14.5 vs 15.3 %), and the all-cause death rates were likewise similar (6.3 and 5.6 %). EMB-monitored patients underwent an average 3.0 EMBs per person-year as against 0.5 for AlloMap-monitored patients; these latter prompted by clinical or echographic signs of rejection or by AlloMap® scores greater than a threshold initially set at 30 and later at 34. In 6 of the 34 episodes of rejection among AlloMap-monitored patients, the only initial indication of rejection was the AlloMap® score. The main limitation of this study was its restriction to low-risk patients, imposed not only through the stability and ejection fraction conditions but also by the exclusion of patients who had undergone HT less than 6 months previously [20]. Additionally, the change in AlloMap® threshold prevented evaluation of a well-defined value. A further weakness was the lack of blinding, which may possibly have influenced clinical decisions.

Kobashigawa et al. have recently published the results of an AlloMap-EMB comparison that complemented IMAGE to some extent by enrolling its 60 participants at times between 55 days post-HT (when GEP becomes valid) and 6 months post-HT [21]. The Early Invasive Monitoring Attenuation through Gene Expression trial (EIMAGE) was a single-centre randomized trial to assess the safety and efficacy of GEP in comparison to EMB between 55 days and 18 months post-HT. In the GEP arm, EMB was performed if indicated by clinical or echocardiographic findings or if GEP was ≥30 between 2 and 6 months or ≥34 at later times. As in IMAGE, participants were required to be clinically stable and to have normal allograft function (defined by LVEF ≥50 %). The coprimary outcome variables were (a) the composite death, retransplantation, rejection with haemodynamic compromise or graft dysfunction at any time between baseline and 18 months post-HT and (b) the change in intravascular ultrasound (IVUS)-measured maximal intimal thickness (MIT) between baseline and 12 months post-HT. The GEP and EMB groups did not differ significantly with respect to either of these outcome variables, nor with respect to cardiac index at 1 year, LVEF at 18 months or overall survival (100 % in both groups). The numbers of EMBs performed in the GEP and EMB groups were, respectively, 42 and 253, a ratio very similar to that occurring in IMAGE. The main weakness of EIMAGE was its small sample size. Like IMAGE, it was also limited to low-risk patients (only 60 % of patients evaluated for inclusion were finally included). These factors together led to their only occurring four EMB-proven cases of rejection and only eight cases of primary endpoint (a) (including five cases of haemodynamic compromise without EMB-proven rejection as first event, a reflection of the limitations of EMB).

Having observed informally that the AlloMap® scores of individual patients seemed to remain quite stable, Deng and coworkers investigated score variability and the prognostic significance of its magnitude [22, 23]. Score stability was formally corroborated in a study of 1135 GEP results from 238 patients without clinical signs or symptoms of rejection, in which within-patient score variability was significantly less than between-patient variability [22]. Then, in a sub-analysis of 369 participants in IMAGE [23], it was found that after adjusting for significant non-GEP variables (younger age, shorter time since HT and race), the probability of suffering the IMAGE composite primary outcome within the next 12 months was significantly correlated (hazard ratio 1.76) with the standard deviation of inverse-logit-transformed past AlloMap® scores but not with the scores themselves or with whether they exceeded a threshold of 34.

The usefulness of GEP for testing for antibody-mediated rejection (AMR) is not known. Cadeiras et al. [24] performed a single-centre pilot study that included 45 HT patients, five of them with clinical signs of AMR. They found that patients with and without AMR differed in the expression of 388 genes belonging to more than 30 gene ontology categories related to cell metabolism, protein synthesis, immune response, apoptosis and cell proliferation. These observations suggest that HT patients with AMR may have a specific gene expression profile in PBMC, but further studies are needed.

The utility of AlloMap® for testing for cardiac allograft vasculopathy (CAV) has also been explored. In one study [25], Yamani et al. found CAV to be associated with higher AlloMap® score among 69 HT patients with ISHLT grade 0R or 1R who were evaluated for CAV by coronary angiography a mean 35 months post-HT. After adjustment for different follow-up times and other potential confounders, the CAV group had scores of 32.2 ± 3.9 vs 26.1 ± 6.5 in the non-CAV group (p < 0.001), with 70 % of patients having a score >30 in the CAV group as against 30 % in the non-CAV group. In another study of 67 patients from whom AlloMap® scores were obtained a mean 34 months after HT, higher scores were found among 19 patients who had had evidence of early post-HT ischaemic injury [26]. These observations suggest that one cause of false-positive AlloMap® tests may be prior ischaemic injury.

Although the clinical use of GEP is not universally accepted in post-HT care, the current ISHLT Guidelines for the Care of Heart Transplant Recipients [27] consider that GEP with AlloMap® can be used to rule out the presence of acute cellular rejection of grade 2R or greater in appropriate low-risk patients between 6 months and 5 years after HT (recommendation class IIa, level of evidence C).

T Cell Function Assays

ImmuKnow

The ImmuKnow assay, also known as the Transplantation Immune Cell Function Assay (Cylex Inc., Columbia, MD, USA), is an FDA-approved commercial assay that measures ATP release from activated lymphocytes in peripheral blood and correlates with immune responsiveness. The higher the assay value, the greater the risk of rejection, and the lower the assay value, the greater the risk of infection. ATP levels are considered high if ≥525 ng/mL, moderate between 226 and 524 ng/mL and low if ≤225 ng/mL [28]; they have been investigated not only in regard to the incidence of infection, acute rejection, immunological quiescence and clinical stability, but also in relation to white blood cell count; absolute lymphocyte count; trough levels of cyclosporine, tacrolimus and mycophenolate mofetil; C-reactive protein levels; and Afro-American ethnicity [26, 28].

In a meta-analysis of 504 solid-organ transplant recipients attending ten US centres [29], statistical balance between infection and rejection was obtained at an ImmuKnow-measured ATP level of 280 ng/ml. This meta-analysis included only 86 HT patients, who together suffered only three rejection episodes and two infections. However, in a longitudinal study of 50 HT patients providing 327 samples, Israeli et al. [30] found that ATP levels during infection and EMB-proven acute rejection (median values 129 and 619 ng/mL, respectively) differed significantly from those observed during clinical quiescence (median 351 ng/mL). By contrast, in a longitudinal study of 296 HT patients with a total of 864 ImmuKnow assays performed between 2 weeks and 10 years post-HT, Kobashigawa et al. [8] observed that although ATP levels were significantly lower among patients in whom infections were diagnosed within 1 month post-test than among those with no diagnosed infection, 187 vs 280 ng/mL (p < 0.001), the levels recorded during rejection episodes did not differ significantly from those recorded during steady state; however, like the meta-analysis of Kowalski et al., this study included no more than a small number of acute rejection episodes (8, as against 38 episodes of infection). In another small study (125 patients, 182 assays, three rejection episodes, eight infection episodes), Gupta et al. [31] observed no correlation between baseline ATP levels and either infection or rejection during the following 6 months and no useful ATP threshold for either increased risk of rejection or increased risk of infection, and in a study of 56 HT patients providing 162 assays over 18 months, Heikal et al. [32] found that ATP levels were significantly lower during the 31 infection episodes within 1 month of assay (p = 0.04) but observed no episodes of acute cellular rejection allowing evaluation of ImmuKnow for its detection. Furthermore, a meta-analysis of nine studies [33] found that ImmuKnow did not perform well for diagnosis of either infection or rejection, except insofar as its NPV for rejection, 0.96, might allow it to be used in the same way as AlloMap®.

The utility of the ImmuKnow assay for monitoring paediatric HT patients is similarly questionable. Macedo et al. [34] found negative correlation between T cell ATP release and Epstein-Barr virus load, but Rossano et al. [35] found ImmuKnow results to be predictive of neither rejection nor infection.

In regard to outcomes other than infection or cellular rejection, Kobashiwaga et al. [8] reported that ATP levels were related to AMR, but Heikal et al. [32] found that IL-5, but not ATP, was associated with AMR. Heikal et al. also reported that IL-8 and tumour necrosis factor-α, but not ATP, were related to CAV, whereas Cheng et al. [36] reported that ImmuKnow values measured 2 months post-HT increased with the 2- to 12-month increase in IVUS-measured average percent stenosis, plaque volume and MIT. However, the sensitivity of the optimal ATP threshold for predicting an increase of ≥0.5 mm in MIT, 406 ng/mL, was only 66.7 %, although its specificity was 94.3 %.

Enzyme-Linked Immunospot

The enzyme-linked immunospot (ELISPOT) assay is capable of detecting cytokine secretion by individual antigen-reactive T cells; whereas an enzyme-linked immunosorbent assay (ELISA) for a cytokine determines the amount of cytokine produced, an ELISPOT assay determines the number of reactive cells [37, 38]. HT patients have been tested with ELISPOT assays in only a few small studies. In one, the activities of the direct and indirect antigen presentation pathways against irradiated and fragmented donor spleen cells, respectively, were separately quantified using ELISPOT assays for PBMC secreting interferon-γ (IFN-γ) before, during and after acute rejection episodes [39]. The direct presentation pathway was always present, the number of IFN-γ-producing cells increasing during rejection and decreasing after successful treatment. The indirect pathway also increased during rejection episodes when present but was present in only 8 of the 18 episodes investigated, 5 of these 8 occurring more than 6 months post-HT. In another study [40], evaluation of the response to donor-HLA-derived peptides by IFN-γ ELISPOT assays and of alloantibodies by flow cytometry using HLA-coated beads showed a cellular or humoral response in 17 of 32 patients with CAV as against only 4 of 33 CAV-free patients, though only in one patient, both pathways were active. Further studies are needed to determine the usefulness of such assays in everyday clinical practice.

Antibody Monitoring

Anti-HLA Antibodies

The development of anti-HLA antibodies after HT has been associated with allograft injury, cellular rejection, AMR, CAV and poor survival [9, 4144]. Solid-phase immunoassays have allowed quantification of even low levels and variation of antibodies against donor-specific HLA (donor-specific antibodies (DSA)), although there are still technical issues that must be ironed out, including questions of assay standardization [45]. As in the case of kidney grafts [46], the ability to activate complement may be important in differentiating clinically more relevant from less relevant DSA [45]: in a large retrospective study of 565 pre-HT sera from patients with negative pre-HT crossmatch to donor cells, C4d-fixing DSA were even more detrimental to survival than DSA in general [47], and in a retrospective study of 18 paediatric HT patients [48], Chin et al. found that the presence of C1q-fixing DSA immediately after transplantation achieved perfect differentiation between patients who did and did not develop AMR.

“Non-HLA” Antibodies

Antibodies against various donor antigens other than HLA can also cause injury, although information in this area is relatively scant. Non-HLA antibodies can be directed against polymorphic non-HLA antigens such as major histocompatibility complex class I chain-related proteins (MIC). These antibodies bind to endothelium and result in apoptosis but not complement-mediated injury [9]. Antibodies against endothelial cells have been associated with acute humoral rejection, CAV and poor graft survival [49]. Vimentin is the most abundant immunoreactive endothelial cell antigen, and in experimental models [50, 51] and clinical studies [52, 53], anti-vimentin antibodies have been associated with poor outcomes and CAV. Antibodies against MIC A have also been associated with poor allograft outcome in some studies [54, 55], but not in others [56]. Other antibody targets related to rejection or poor graft outcome include myosin [57, 58], angiotensin type 1 receptor [59] and endothelial cell cytoskeletal proteins (actin, tubulin and anticytokeratin) [60].

Current Consensus

In spite of the growing evidence for the importance of antibody monitoring after HT, the optimal monitoring schedule and response to results have not yet been elucidated [9]. Kobashigawa et al. [61], reporting on the ISHLT Consensus Conference on AMR in Heart Transplantation, stated that circulating DSA was no longer believed to be required for diagnosis of AMR (since, for example, DSA might be immobilized on the allograft), but that when AMR is suspected, blood should be nevertheless be tested for DSA against HLA classes I and II and also, possibly, in its absence, for non-HLA antibodies. The conference recommended testing for DSA by solid-phase and/or cell-based assays 2 weeks and 1,3, 6 and 12 months after HT and then annually and when AMR is suspected on clinical grounds, with quantification if detected, and also after major decreases in immunosuppressive medication, to determine any increase in alloimmune activation.

The antibody measurement recommendations of a recent American Heart Association scientific statement on AMR after HT are as follows [9]. Solid-phase and/or cell-based assays of DSA should be performed during the first 90 days after transplantation or when AMR is suspected (recommendation class I, level of evidence C). It is reasonable to perform solid-phase and/or cell-based assays of DSA for surveillance 3, 6 and 12 months after transplantation and annually thereafter or in accordance with the centre’s routine surveillance protocol (recommendation class IIa, level of evidence C). Determination of non-HLA antibodies against targets such as endothelial cells, vimentin, MIC A and MIC B may be considered when anti-HLA antibodies are absent and AMR is suspected (recommendation class IIb, level of evidence C).

Donor-Derived Cell-Free DNA

Cell-free DNA (cfDNA) is the name given to short lengths of double-stranded DNA released from tissues as a result of either normal physiological cell turnover or pathological apoptotic or necrotic cell death. It is found in the circulating blood plasma at levels typically around 3 ng/mL. It has, for some years, been used for non-invasive prenatal testing [62, 63], is being intensively investigated in the field of oncology [64] and is now being considered for use in transplantation [6567]. Just as its use in prenatal testing depends on the fraction of cfDNA in a mother’s circulation that is of foetal origin, so also its use to detect graft rejection depends on the fraction in a graft recipient’s circulation that originates in the graft. This donor-derived cfDNA (dd-cfDNA) was first established as detectable in recipients of sex-mismatched grafts [65]. Later work confirmed the hypothesis that dd-cfDNA correlates with graft rejection [68, 69], rising from less than 1 % to around 5 % during rejection episodes, and the development of massively parallel sequencing technology has potentially made dd-cfDNA detection and measurement available for all graft recipients, not just those with grafts from donors of the opposite sex [10, 70]. Although shotgun sequencing [11, 70] and targeted amplification [71] have the drawback of requiring both recipient and donor genotypes, more recent methods using digital droplet PCR of single-nucleotide polymorphisms with high minor allele frequencies seem promising [72].

MicroRNAs

MicroRNAs (miRNAs) are molecules of endogenous RNA some 18–25 nucleotides long that, instead of encoding proteins, act as post-transcriptional gene expression regulators. They are present not only intracellularly, but also in a highly stable form, in the bloodstream, which suggests the targeting of distant tissues. Although in principle, they have great potential as biomarkers of protein dysregulation [73], their interpretation is complicated. Nevertheless, in the past couple of years or so, there have been reports of modified levels of a number of miRNAs during rejection of transplanted livers [74], kidneys [75] and hearts [12, 13]. In the case of heart grafts, Öhman and coworkers identified seven miRNAs with higher levels during than before or after rejection [12], while Van Huyen et al. [13] reported that joint consideration of the levels of miR-10a, miR-31, miR-92a and miR-155 was able to discriminate between patients with and without rejection. It is to be hoped that these findings will not only be confirmed by subsequent studies, but will also throw light on the pathophysiology of heart graft rejection.

Conclusions

The development of sensitive, objective, non-invasive immune biomarkers of allograft rejection has being a major goal in HT. In the last years, there has been great progress in different approaches. GEP test seems to be useful for ruling out the presence of a moderate to severe acute cellular rejection in low-risk patients. Antibody monitoring is recommended at least during the first year after HT and if AMR is suspected. T cell functional assays have been investigated although its usefulness in HT remains controversial. Newer monitoring tools, like dd-cfDNA or miRNA, seem to be promising for individualizing immunosuppressive therapies and better understanding the mechanisms of rejection.