In patients with chronic heart failure (HF), the most catastrophic outcome is sudden cardiac death (SCD), particularly in a patient who is otherwise doing relatively well. SCD accounts for up to 50% of HF deaths.1 While there are a variety of etiologies, it is most often a ventricular tachycardia progressing to ventricular fibrillation (VT/VF). Numerous large prospective multicenter studies have demonstrated improved patient survival with primary preventive use of an implantable cardioverter defibrillator (ICD) that in guidelines has been assigned a Class IA indication for “primary prevention of SCD to reduce total mortality in selected patients with nonischemic [dilated cardiomyopathy] or ischemic heart disease at least 40 days post-MI with [a left ventricular rejection fraction] (LVEF) of 35% or less and NYHA class II or III symptoms on [guideline directed medical therapy], who have a reasonable expectation of meaningful survival for more than 1 year.”2 Yet, in spite of such a strong recommendation, it is widely acknowledged that basing primary prevention ICD use on a LVEF threshold, particularly 35%, is not well supported by clinical experience,3 or when findings of the randomized studies upon which guidelines are based are examined more closely.4 With current practice, >80% of primary prevention patients do not use their ICD over a duration of as long as 8 years, with the device costing as high as $235,000 per year of life saved, and with a significant potential for serious device complications.5 It is clear that a better approach to selecting HF patients for primary prevention ICD implantation is needed.

Necessity of Patient Testing Based on Underlying Arrhythmic Mechanisms

It should be self-evident that a more effective way of determining arrhythmic risk would be to use testing methods that examine the basic underlying causes of arrhythmias. Such methods need, in some way, to identify myocardial substrate abnormalities predisposing to electrical conduction heterogeneity, which in the setting of a deleterious trigger and modulating factors can produce dangerous arrhythmias.5,6 Techniques that assess cardiac autonomic systemic hormonal control or direct cardiac innervation that, if impaired in some way, has been shown to help produce an abnormal myocardial electrical conduction substrate, may produce an arrhythmic trigger and can modulate arrhythmias by either propagating or quenching them, should be effective tools for arrhythmic risk stratification.5

While measurements of serum neurohumoral markers, such as catecholamine levels, are commonly used to risk stratify HF patients, they tend to be nonspecific for arrhythmic risk, and therefore much recent work has focused on direct cardiac adrenergic innervation imaging using the norepinephrine (NE) single-photon emission computed tomographic (SPECT) radionuclide analog, iodine-123 meta-iodobenzylguanidine (123I-mIBG). Since its development approximately 35 years ago, 123I-mIBG was recognized as visualizing the “wiring” of the heart, with potential for identifying patients with autonomic neuropathies who have partial denervation predisposing to arrhythmias and SCD.7 Early basic and clinical work demonstrated the potential of 123I-mIBG imaging to evaluate arrhythmic risk. In dogs, Zipes and colleagues8 showed that artificially created sympathetic efferent denervation produces an exaggerated shortening of the effective refractory period (ERP) during NE or isoproterenol infusions, which increases vulnerability to stimulation-induced ventricular fibrillation, termed “denervation supersensitivity.” Visualization of a focal 123I-mIBG defect, which was present even in the absence of a 201Tl perfusion, increased arrhythmogenicity for a period of up to 3 weeks after reinnervation. In several small (<30 patient) human studies, post-MI regional 123I-mIBG defects were found to predispose to ventricular tachyarrhythmias.9,10

As a key, relatively large (90 patient) prognostic study by Merlet et al11 found global cardiac 123I-mIBG uptake, as measured by the heart-to-mediastinum ratio (HMR), to strongly and independently risk stratify advanced HF patients in terms of overall survival, the predominant focus shifted to planar techniques. Numerous studies thereafter showed that HMR and the global tracer washout rate (WO or WR) between initial and delayed (~4 hour) planar images robustly risk stratified advanced HF patients in terms of cardiac and overall survival, as well as various secondary HF-associated events.12 The most important investigation to date has been the prospective, rigorously performed 961-patient “AdreView Myocardial Imaging for Risk Evaluation in Heart Failure” (ADMIRE-HF) study, which demonstrated that an HMR <1.6 increased the occurrence of cardiac events, independent of commonly used prognostic variables such as LVEF and B-type natriuretic peptide (BNP).13 Of course, a patient who survives does not have an arrhythmic SCD, and as ADMIRE-HF recruited patients before ICD guidelines were established, for extended periods many patients did not have ICDs to prevent this outcome. In an early analysis of such patients without an ICD, for those with an HMR ≥1.6, there was only one arrhythmic death (at HMR = 1.6), as depicted in Figure 1.14. A later, more rigorous analysis of 777 patients who did not initially have ICDs showed that HMR added incremental prognostic value and enhanced risk reclassification (for survival), with HMR identifying the number of lives saved by ICD use per 100 treated.15 For all patients (ICD and no ICD) in ADMIRE-HF, combined “arrhythmic” events (self-limited ventricular tachycardia, resuscitated cardiac arrest, appropriate ICD discharges) were more common in subjects with HMR <1.60 (10.4%) than in those with HMR ≥1.6 (3.5%, P < 0.01).

Figure 1
figure 1

Arrhythmic events vs heart-to-mediastinum ratio (HMR) in patients without an implantable cardioverter-defibrillator (ICD). In patients with HMR ≥1.6, there was only one arrhythmic death, in a patient with HMR value of 1.6. Reprinted from Seminars in Nuclear Medicine, Vol. 41, Chirumamilla A, Travin MI. Cardiac applications of 123I-mIBG imaging.; pp. 374-87, 2011, with permission from Elsevier

Interestingly, Figure 1 illustrates that patients with the highest arrhythmic risk, i.e., those for whom an ICD would be most beneficial, may not be those with the lowest HMR, but rather those with an intermediate HMR. In accordance with this concept, a recent report by Verschure et al from a cohort of 135 stable HF patients from 13 European institutions, enrolled in an ICD implantation primary prevention trial, found that the highest occurrence of SCD or appropriate ICD therapy was in patients with HMR being between 1.4 and 2.1, as opposed to those with HMR <1.4 or >2.1.16 While part of the explanation may be a higher rate of pump failure deaths in patients with the lower HMRs, this observation likely also indicates that an intermediate HMR represents an increased substrate heterogeneity that predisposes to electrical instability.

Tomographic Imaging as a Potentially Better Tool for Assessing Substrate Electrical Heterogeneity

Of course, one would expect tomographic imaging to be better than planar imaging for assessing such heterogeneity. In accordance with the notion of global adrenergic imaging perhaps being too blunt a tool for arrhythmic risk stratification, in reviewing data from over 600 patients, Verschure et al noted that while HMR effectively risk stratified in terms of cardiac mortality, all-cause mortality, and the need for transplant, it did not independently stratify arrhythmic events (ArEs).17 Thus, the finer tool of tomographic identification of focal adrenergic abnormalities should be a better approach. In a prospective study of 116 patients receiving an ICD, Boogers et al18 found that a dysinnervation defect score above a particular threshold increased the likelihood of an appropriate discharge. However, a more recent subanalysis from ADMIRE-HF that undertook rigorous interpretation of tomographic images from patients with ischemic HF showed that although the occurrence of ArEs was significantly related to the dysinnervation defect score, the hazard ratio was <1 because arrhythmic risk was the highest in patients with intermediate defect scores as shown in Figure 2.19 While further investigation is required to confirm and establish this concept, such findings suggest that methods used in perfusion imaging to risk stratify patients, i.e., interpreting defects that are larger and more severe as increasing risk, may not apply when using nonperfusion tracers for the characterization of pathology that is not coronary disease. A different paradigm, i.e., interpreting nuclear cardiac images from a different viewpoint, appears to be required.

Figure 2
figure 2

Proportion of arrhythmic events (ArEs) and nonfatal arrhythmic events in relation to low, intermediate, and, high 17-segment 123I-mIBG scores. Note Nonfatal ArEs include only spontaneous sustained (>30 seconds) ventricular tachycardia, resuscitated cardiac arrest, and appropriate ICD activation. From Travin MI, Henzlova MJ, van Eck-Smit BLF, Jain J, Carrió I, Folks RD, Garcia EV, Jacobson AF, Verberne HJ. Assessment of 123I-mIBG and 99mTc-tetrofosmin single-photon emission computed tomographic images for the prediction of arrhythmic events in patients with ischemic heart failure. J Nucl Cardiol; 2016; 24: 377-91, with permission of Springer

Potential for Determining Arrhythmic Risk by Assessment of Tracer Washout Heterogeneity

Another widely reported aspect of 123I-mIBG image interpretation has been assessing dynamic changes over time by deriving washout (WO) measurements between early and delayed images. Henderson et al20 first described the presence of significantly higher 123I-mIBG washout in patients with cardiomyopathy compared with healthy volunteers. Later work showed high WO to be associated with worsened survival21 with potential ability to predict SCD better than many sophisticated electrocardiographic variables.22

The factors contributing to WO values are complex and multifactorial, involving uptake of tracer into the neuron, storage in vesicles, and transport in extraneuronal tissues,20 as well as competition for pre-synaptic uptake receptors from circulating catecholamines23 and augmented spillover from an accelerated sympathetic drive.24 Various ways to derive WO rate have been described. Unfortunately, global WO measurements from planar imaging can be corrupted by overlying lung tracer uptake, and assessment of tomographic images is often impaired by extremely low counts in advanced HF.

Considering these concepts, in this issue of the journal, Yamamoto et al25 present their findings from an investigation, in 73 patients who have HF with reduced ejection fraction, of the relationship between the regional variation of WO on SPECT 123I-mIBG imaging, meant to reflect potential electrical heterogeneity of myocardial substrate, and the occurrence of SCD. Although patients in their report are derived from a study performed between 1995 and 1999 for another primary purpose, i.e., investigation of the HF-treatment efficacy of various medications, patients did undergo 123I-mIBG imaging and were followed up over a mean of 7.5 years for occurrence of SCD that was defined as a “witnessed cardiac arrest or death within 1 hour after the onset of acute symptoms or unexpected or unwitnessed death in a patient known to have been well within 24 hours,” or development of sustained VT lasting ≥30 seconds on a series of periodic Holter monitors. It should be noted that the lack of proof that the deaths were arrhythmic in origin, and the fact that the patient did not necessarily have death from the VT episodes as observed on Holter monitors, does limit study conclusions.

For the characterization of SPECT regional WO abnormality, the authors used a calculated WO range defined as the difference between the maximum and the minimum WO values among 17 segments in their polar plot analyses, and the presence of “abnormal regional” WO defined as a segment having a value >2 standard deviations higher than a mean segment WO value derived from 15 “control” patients without HF. For 20.5% of the patients (N = 15) who subsequently had SCD (as defined by the authors), there was a higher WO range, and a significantly more frequent abnormal regional WO presence, as well as higher global WO, and lower delayed HMR on planar images. Survival analysis showed that the presence of both an abnormally high global WO and an abnormal regional WO was associated with an overall death rate 6-fold higher than that for either alone. Multivariate analysis showed that abnormal regional and abnormal global WO values were significant and independent predictors of SCD, with both together being associated with a significantly higher ArE rate than that for none or either variable alone. On the contrary, conventional HF parameters, including serum creatinine, sodium, or NE, and especially LVEF, did not predict SCD.

Although there are significant limitations to this study, including a low number of patients, limited heterogeneity analytic techniques, issues related to SCD characterization, with study findings therefore being far from ready to be used for guidance of ICD management, the authors have made a commendable effort to advance the technique of using cardiac adrenergic imaging to improve arrhythmic risk stratification in HF patients. Their findings further support the contention that interpretation of adrenergic images for arrhythmic risk must be approached from a viewpoint different from what has conventionally been the case for perfusion imaging. Results of this study should encourage further development of algorithms to assess regional tomographic image heterogeneity, both for tracer uptake and washout, perhaps similar to the type of analyses done for gated SPECT phase analysis,26 or maybe using a more complex technique such as fractal analysis.27 In addition, imaging techniques that provide better resolution, such as a SPECT imaging with solid state cameras, or positron emission tomographic (PET) imaging with tracers such as 11C-hydroxyephedrine (11C-HED) that can better differentiate between innervated and denervated myocardium, promise better arrhythmic risk stratification.28

Importance of Clinical Context

When using image findings to predict events and management patients, it is of utmost importance to incorporate clinical data. Recently, Levy et al showed the potential for clinical parameters from the Seattle Proportional Risk Model to effectively predict occurrence of SCD, and thus better help determine the benefit of an ICD.1 Another recent study showed that ICD use may not improve survival in most patients with nonischemic cardiomyopathies,29 in which case 123I-mIBG imaging may not be helpful at all, or conversely, that it may be a good way to identify a subset of such patients who would benefit. Thus, clinical context has to be considered when choosing the imaging technique (planar vs tomographic, SPECT vs PET) and interpreting the meaning of test findings for predicting arrhythmic risk, with more investigational data needed in this regard.

In addition, the likelihood of a particular adverse outcome varies with time.30 Risk varies at different stages of a disease process, something that must be considered when using image findings to guide patient management. As a patient’s clinical situation changes, especially if there is ischemic coronary disease, image findings may only apply for a limited period of time, after which repeat testing to update risk stratification may be necessary.

Final Thoughts

One cannot assume that techniques that have customarily been used to interpret myocardial perfusion imaging and guide management apply to other forms of radionuclide imaging. In particular, predicting arrhythmic risk with adrenergic imaging must consider underlying mechanisms, and thus one must view and interpret the images with this in mind.