Reliability of NI-RADS criteria in the interpretation of contrast-enhanced magnetic resonance imaging considering the potential role of diffusion-weighted imaging

Abstract

Objectives

To assess inter- and intrareader agreement of the Neck Imaging Reporting and Data System (NI-RADS) used in contrast-enhanced magnetic resonance imaging (MRI) including analysis of diffusion-weighted imaging (DWI), which is currently not part of the NI-RADS criteria.

Methods

This retrospective study included anonymized surveillance contrast-enhanced MRI datasets of 104 patients treated for different head and neck cancers. Three radiologists experienced in head and neck imaging reported findings for the primary site and the neck using NI-RADS criteria in a first step and evaluated DWI sequences for the primary site in a second step. Thirty randomly selected imaging datasets were again presented to the readers. Kappa statistics and observed agreement (Ao) were calculated.

Results

Interreader agreement across all MRI datasets was moderate (κFleiss = 0.53) for NI-RADS categories assigned to the primary site, substantial for NI-RADS categories of the neck (κFleiss = 0.67), and almost perfect for DWI of the primary site (κFleiss = 0.83). Interreader agreement for the primary site was particularly low in cases of cancer recurrence (κFleiss = 0.35) and when categories 2a, 2b, and 3 were combined (κFleiss = 0.30). Intrareader agreement was considerably lower for NI-RADS categories of the primary site (range Ao = 53.3–70.0%) than for NI-RADS categories of the neck (range Ao = 83.3–90.0%) and DWI of the primary site (range Ao = 93.3–100.0%).

Conclusion

Interreader agreement of NI-RADS for reporting contrast-enhanced MRI findings is acceptable for the neck but limited for the primary site. Here, DWI has the potential to serve as a reliable additional criterion.

Key Points

• NI-RADS was originally designed for contrast-enhanced computed tomography with or without positron emission tomography but can also be used for contrast-enhanced magnetic resonance imaging alone.

• Overall interreader agreement was acceptable for NI-RADS categories assigned to the neck but should be improved for the primary site, where it was inferior to DWI; similar tendencies were found for intrareader agreement.

• DWI is currently no criterion of NI-RADS, but has shown potential to improve its reliability, especially for categories 2a, 2b, and 3 of the primary site.

Introduction

In the 2018 worldwide cancer statistics, head and neck (HN) cancer including malignancies of the thyroid gland accounted for 8% (1,454,892 patients) of all new cases and 5.2% (494,378) of all deaths [1]. The term HN cancer denotes a heterogeneous group of entities, among which squamous cell carcinoma (SCC) is the largest subgroup [2]. After definitive treatment of HN cancer including surgery, radiotherapy, and chemotherapy, patients are usually put on surveillance imaging programs including contrast-enhanced computed tomography (CECT) or magnetic resonance imaging (CEMRI), possibly combined with positron emission tomography (PET-CT and PET-MRI). Currently, non-standardized reports are common practice. While such reports allow radiologists to individually weight their findings, they can also cause problems [3]. The already complicated head and neck anatomy is even more difficult to interpret in posttreatment situations [4]; therefore, findings, conclusions, and recommendations in radiological reports tend to be heterogeneous, which might reduce their acceptance among surgeons. In 2014, the American College of Radiology (ACR) called upon radiologists to become “patient-centric, data-driven, and outcomes-based” and published the first version of a lexicon and standardized reporting system in 2016 calling it “Neck Imaging Reporting and Data System” (NI-RADS) [5, 6]. NI-RADS thus joins a number of RADS already implemented in clinical routine, which started with the release of the by now well-established BI-RADS in 1997 [7, 8]. NI-RADS reports provide two separate numerical categories between 1 and 4 that stratify the risk of HN cancer recurrence for the primary site and the neck (i.e., cervical lymph nodes) [9]. NI-RADS also addresses the ACR’s imaging 3.0 strategy, which seeks to include clinical decision support directly from radiological reports by linking each category with a concrete recommendation for further patient handling such as shortening the interval until the next imaging examination or performing a biopsy [10]. In this way, NI-RADS standardizes the entire surveillance program. While originally designed for CECT with or without PET, NI-RADS—according to its authors—can also be applied to MRI [10]. Diffusion-weighted imaging, which has already established itself as a standard sequence in head and neck imaging, is not yet a criterion in NI-RADS [11].

The objective of our study was to assess the inter- and intrareader agreement of NI-RADS used in the interpretation of surveillance CEMRI after definitive treatment of head and neck cancer and to gain first insights into the potential of simultaneously acquired DWI data, which are currently not part of the NI-RADS criteria.

Materials and methods

Study design and patient population

Our institutional review board approved this retrospective study, and written informed consent for use of their data for scientific purposes was obtained from all patients. We searched our department’s database for suitable patients from all three sites of our hospital using the following inclusion criteria: (1) consecutive patients with available CEMRI datasets of the neck acquired for surveillance after curative treatment of HN cancer; (2) CEMRI datasets including (a) axial contrast-enhanced and fat-suppressed T1-weighted (T1w CE FS) sequences, (b) coronal and fat-suppressed T2-weighted (T2w FS) sequences, (c) axial diffusion-weighted images with corresponding apparent diffusion coefficient (ADC) maps; (3) pretreatment or prior surveillance imaging datasets (CEMRI or CECT) available; (4) availability of reports on either subsequent histopathology or subsequent surveillance imaging no earlier than 6 months later. Datasets with typical MRI artifacts were deliberately not excluded but part of the reporting process as described in the “Image interpretation” paragraph.

The CEMRI datasets included were generated on eight different MRI scanners. Scanner details as well as specific acquisition parameters of the sequences included are presented in Tables 1, 2, 3, and 4 of Electronic Supplementary Material.

Image interpretation

Three radiologists (A: 15 years; B: 7 years; and C: 6 years of experience in head and neck imaging) contributed reports to this study. Readers A and C are members of the institutional weekly interdisciplinary head and neck tumor board team. The NI-RADS system was well understood as each reader had gained prior experience from interpretation of over 200 CECT and CEMRI datasets using NI-RADS.

All CEMRI datasets included were anonymized and scrubbed of all identifying information and then randomly reordered by using a Mersenne Twister pseudorandom number generator embedded in Microsoft Excel (Version 16.16.10, Microsoft Corporation). The resulting CEMRI datasets were presented on a radiology workstation with two diagnostic monitors and an administrative monitor using Visage Client (Version 7.1.11, Visage Imaging Inc.). The left diagnostic monitor was used to show the current CEMRI dataset, while an earlier imaging dataset (obtained before or after treatment) was presented on the right monitor for comparison. Not until after finishing a NI-RADS report by accessing the axial T1w CE FS and coronal T2w FS sequences intended were readers allowed to view DWI sequences. Readers were provided with a short medical report containing information on the cancer entity, its localization (primary site), and the type of curative therapy (resection, radiation therapy, chemotherapy) to simulate a clinical reporting situation.

For each imaging study interpreted using NI-RADS, two categories are assigned—one for the primary site and another for the neck [9]. Categories range from 1 to 4 and reflect rising probabilities of cancer recurrence. A category of 0 is assigned only if a new baseline study is being interpreted and a prior imaging study exists but is not available at the time of reading. Therefore, a category of 0 could not be assigned in our study setting. A NI-RADS 1 category for the primary site should be assigned when expected posttreatment changes such as diffuse submucosal edema or superficial linear enhancement and non-mass-like distortion are present. NI-RADS 1 for the neck indicates absence of residual abnormal, new or enlarging lymph nodes. NI-RADS 2 for the primary site is divided into 2a for focal mucosal but non-mass-like or nodular enhancement and 2b for deep, ill-defined but not discrete enhancement. NI-RADS 2 for the neck stands for new or enlarging lymph nodes without morphologically abnormal features (central necrosis or extranodular extension). NI-RADS 3 for the primary site is assigned for new or enlarging discrete nodules or masses with intense enhancement. NI-RADS 3 for the neck indicates presence of new or enlarging lymph nodes with central necrosis or extranodular extension. NI-RADS 4 represents clinically or radiologically definite cancer recurrence either at the primary site or in the neck.

Additionally, readers were asked to evaluate the DWI sequences for the primary site by choosing one of two possible options: clear diffusion restriction (positive) and no or ambiguous diffusion restriction (negative). In addition to NI-RADS categories for the primary site and the neck and DWI evaluation for the primary site, readers were asked to also rate image quality for both target sites and for the DWI sequences on a 4-point scale (1 = excellent, 2 = good, 3 = sufficient, 4 = insufficient). More specifically, readers should assign a score of 4 for insufficient image quality when adequate interpretation of the primary site, the neck, or both was prevented by the presence of artifacts.

Three months after completion of retrospective reading of the datasets, readers were asked again to provide NI-RADS categories and evaluate DWI sequences following the same scheme for a subset of 30 randomly selected CEMRI datasets (cases in which an image quality category of 4 was assigned by at least one reader in the first session were excluded). The second session was intended to calculate intrareader agreement.

Data analysis

“RStudio” (Version 1.1.383, RStudio) was used both for statistical data analysis (“irr” package) and for creating the heatmap (Fig. 1, “gplots” package).

Fig. 1
figure1

Category distribution chart for all 104 patients. Category counts are coded as shades of blue. The results of the confirmation studies are coded in green and red. Numbers to the left and to the right side indicate figures providing examples for illustration

Fleiss’ kappa (κFleiss) and Conger’s kappa (κConger) were calculated to test interreader agreement between all three readers. Intrareader agreement was tested by calculation of Cohen’s kappa (κCohen). Additionally, the observed agreement (AO, i.e., all three readers assigned the same category) was calculated.

κFleiss, κConger, and κCohen were interpreted as recommended by Landis and Koch: 0.01–0.20 = “slight agreement,” 0.21–0.40 = “fair agreement,” 0.41–0.60 = “moderate agreement,” 0.61–0.80 = “substantial agreement,” and 0.81–1.0 = “(almost) perfect agreement” [12].

Subgroups were formed according to the results of the confirmation study (disease-free versus cancer recurrence) and the NI-RADS categories assigned by the majority of the readers (2a/2b/3 versus 1/4 merged; in case each reader provided a different category for a target, the category assigned by the most experienced reader A was decisive). As a secondary outcome, recurrence rates for all patients and for each NI-RADS category were calculated.

Results

A total of 104 patients (35 female, 69 male) were finally accepted for retrospective analysis. They had a median age of 60 years with a range of 18–89 years. The median follow-up interval for imaging confirmation studies was 219 days (range 184–450 days). Table 1 provides the distribution of HN cancer entities included in our study along with subsites and treatment plans. Pretreatment or prior surveillance imaging datasets were acquired by CEMRI in 64 cases and CECT in 40 cases.

Table 1 Distribution of cancer entities, sites, and treatment plans

With regard to image quality, NI-RADS categories of 101 datasets for the primary site (3 ratings of 4 = insufficient) and 103 datasets for the neck (1 rating of 4 = insufficient) as well as evaluations of 87 datasets for DWI (17 ratings of 4 = insufficient) were accepted for further analysis. Rating distributions for all 104 patients as well as the results of the confirmation studies are presented in a heatmap in Fig. 1. Examples illustrating high and low interreader agreement for NI-RADS categories assigned to the primary site and the neck are shown in Figs. 2, 3, 4, and 5.

Fig. 2
figure2

a-d Example case illustrating high agreement for the primary site. Axial T1w CE FS images obtained before (a) and 6 months after total parotidectomy and adjuvant radiochemotherapy (b) in a patient diagnosed with adenocarcinoma located in the right parotid gland (white arrows). Image b again shows nodular enhancement at the primary site (white arrows), for which all three readers assigned a category of 4. All three readers also identified corresponding diffusion restriction in DWI/ADC maps (c and d, white arrows). Histopathology confirmed cancer recurrence

Fig. 3
figure3

a-d Example case illustrating low agreement for the primary site. Axial T1w CE FS images obtained 12 (a) and 24 months (b) after definitive radiochemotherapy in a patient diagnosed with of SCC located in the right root of the tongue. Image b shows new enhancement at the primary site (white arrows), for which readers assigned categories of 2a, 2b, and 3. All three readers identified corresponding diffusion restriction in DWI/ADC maps despite severe geometric distortion (c and d, white arrows). Histopathology revealed cancer recurrence

Fig. 4
figure4

a, b Example case illustrating high agreement for the neck. Axial T1w CE FS images obtained before (a) and 6 months (b) after surgical resection in a patient diagnosed with SCC located in the right palatine tonsil (white arrows in a). Posttreatment image b shows a new right retropharyngeal lymph node (white arrows), for which all three readers assigned a category of 3, 3, and 3. Histopathology confirmed cancer recurrence

Fig. 5
figure5

a, b Example case illustrating low agreement for the neck. Magnified section of axial T1w CE FS images obtained 6 (a) and 12 months (b) after lateral parotidectomy and radiotherapy in a patient diagnosed with adenoid cystic carcinoma located in the right parotid gland. Image b shows a slightly enlarged lymph node at level Ib on the left side, for which readers assigned categories of 1, 2, and 2. The subsequent MRI obtained 6 months later showed no further increase in size

The results for interreader agreement are compiled in Table 2. Interreader agreement was moderate for NI-RADS categories of the primary site (κFleiss = 0.53, κConger = 0.54, Ao = 55.5%), substantial for NI-RADS categories of the neck (κFleiss = 0.67, κConger = 0.68, Ao = 75.7%), and almost perfect for evaluation of DWI of the primary site (κFleiss = 0.83, κConger = 0.83, Ao = 92.0%). Regarding assignment of NI-RADS categories to the primary site, interreader agreement was particularly low in cases of cancer recurrence (κFleiss = 0.35, κConger = 0.38, Ao = 38.9%) and when categories 2a, 2b, and 3 were combined (κFleiss = 0.30, κConger = 0.34, Ao = 24.5). Conversely, interreader agreement for DWI was perfect (κFleiss = 1.00, κConger = 1.00, Ao = 100.0%) or almost perfect (κFleiss = 0.83, κConger = 0.83, Ao = 89.1%) in those subgroups.

Table 2 Interreader agreement for all three readers. κFleiss = Fleiss’ kappa, κConger = Conger’s kappa, Ao = observed agreement (in %)

Results for intrareader agreement determined in 30 cases are shown in Table 3. Overall intrareader agreement was considerably lower for NI-RADS categories of the primary site (range κCohen = 0.76–0.84, range Ao = 53.3–70.0%) than for NI-RADS categories of the neck (range κCohen = 0.85–0.90, range Ao = 83.3–90.0%) and DWI of the primary site (range κCohen = 0.63–1.00, range Ao = 93.3–100.0%).

Table 3 Intrareader agreement for all three readers. κCohen = Cohen’s kappa, Ao = observed agreement (in %)

Table 4 summarizes recurrence rates for all datasets and for the individual NI-RADS categories calculated from the results of confirmation studies—histopathological examinations or further follow-up imaging.

Table 4 Category counts, recurrence counts, and recurrence rates by NI-RADS category and total based on majority decisions (the most experienced reader A was decisive in case of tied category counts). For comparison, the last two columns list the recurrence rates reported in the two previously published studies investigated the validity of NI-RADS [19, 22]

Discussion

This study was conducted to obtain first results on the reliability of NI-RADS in the interpretation of CEMRI including DWI, which is currently no criterion of NI-RADS. Overall, our results show acceptable interreader agreement for NI-RADS categories assigned to the neck, while further improvement is desirable for primary cancer sites. In particular, interreader agreement was low in recurrent cancer and for the combination of categories 2a, 2b, and 3. Overall intrareader agreement showed similar tendencies but was acceptable even for NI-RADS categories assigned to the primary site. For DWI, we found excellent interreader agreement.

So far, a single other study on the reliability of NI-RADS has been published; however, this study investigated surveillance CECT datasets after treatment of oral and oropharyngeal SCC, and categories were assigned by four instead of three readers [13]. While not directly comparable with our design, this study found moderate overall interreader agreement for both the primary site (κFleiss = 0.48) and the neck (κFleiss = 0.50). Intrareader agreement was almost perfect (κCohen = 0.85–0.96) for both the primary site and the neck (0.89–0.95). A major difference to our MRI study is that agreement for NI-RADS categories of the primary site was higher in patients with confirmed cancer recurrence than patients without recurrence. This discrepancy could be due to a higher proportion of less clear cases in our MRI study or reader-dependent differences in the interpretation of abnormal findings. Finally, compared with the CECT study, our patient population better reflects the claim of NI-RADS to be applicable to any HN cancer.

Since experience with NI-RADS is still limited, comparison with published data on the reliability of other RADS is worthwhile. Two recent studies are of particular interest here because they arrived at very different results. For discussion and comparison of the overall agreements reported by these studies and quoted below, we use the same interpretations of the statistical parameters as defined by us, though they were not used by the authors of these studies. The first study investigated the current version of BI-RADS and found moderate interreader agreement (κFleiss = 0.57) while intrareader agreement was substantial or almost perfect (range of κCohen = 0.72–0.81) [14]. In contrast, the second study, which investigated the current version of PI-RADS, showed fair interreader agreement (κFleiss = 0.24) and a moderate intrareader agreement range of Cohen’s kappa of 0.43–0.67 [15].

We measured inter- and intrareader agreement using two approaches—calculation of observed agreement (the proportion of perfect agreements between readers) and calculation of the widely used kappa statistics (which takes agreement expected by chance into account). Kappa statistics are suitable for binary (DWI) and categorical or ordinal data. Fleiss and Congers kappa, unlike Cohen’s kappa, can be used to calculate agreement between more than two readers. For interpretation of DWI results in cases confirmed as disease-free, the kappa statistic is very low (κFleiss = 0.23) and does not match the observed agreement (Ao = 87.3%). This paradoxical situation can occur when most of the subjects examined are assigned to the same category [16, 17]. In our study, we observed this paradox in the DWI disease-free subgroup, where the three readers combined reported a diffusion restriction (positive) in 11 instances and no restriction (negative) in 199 instances. This proportion was much more balanced across all patients who underwent DWI with 59 positive versus 202 negative instances.

Another issue to be addressed is poor image quality. In our study, image quality was considered insufficient for three datasets for the primary site and one dataset for the neck. To report such cases in a standardized manner as well, the existing categories might be supplemented by a category for non-diagnostic image quality (e.g., NI-RADS N/A).

In discussions with the readers after the reading sessions, two issues emerged. First, handling of the NI-RADS 2 category for the neck was unclear in some cases with measurable increases in size of unclear relevance. Figure 5 illustrates such a case, in which a very subtle enlargement of a contralateral lymph node (not a typical draining lymph node) is apparent. According to the RECIST criteria, only a short axis diameter larger than 10 mm would be considered relevant [18]. This problem was already noted in the above-quoted study investigating NI-RADS reliability in CECT [13]. We think that a statement regarding this issue would be a valuable supplement to the NI-RADS criteria. The second issue is that one reader reconsidered the NI-RADS category assigned in light of the recommended clinical management that would automatically follow from the category. While this might indicate a potential bias in this reader, it is exactly what is likely to happen in clinical practice [19].

The full spectrum of NI-RADS criteria can only be used with PET, which is likely to result in higher interreader agreement for the primary site. However, given the high interreader agreement we found for the evaluation of DWI, its integration into NI-RADS could have the same effect as additional consideration of PET findings. This is of interest in light of the high cost of PET [20, 21]. Differentiation between categories 2b and 3 would probably benefit most from additional DWI. An example illustrating this situation is given in Fig. 3; in this case, DWI was highly suspicious for primary cancer recurrence. The readers in our study unanimously agree that, with its higher soft tissue contrast, CEMRI might be particularly susceptible to discrepant interpretation in categories 2a, 2b, and 3. Another possible reason for disagreement is that there is an inherent problem in differentiating category 3 and 4, both for the primary site and the neck: while category 3 has clearly defined criteria, category 4 leaves room for interpretation. The recurrence rates calculated as secondary outcome in our study support the high discriminatory power of NI-RADS categories already demonstrated in previous studies [19, 22] with one notable exception. For patients assigned category 3 for the neck, these two studies report higher recurrence rates of 70.0% and 80.0% compared with the rate of 58.3% in our study. Possible explanations are the small sample size for the validity check and an overinterpretation of the “abnormal features.” If these trends are confirmed, a feature-specific investigation could provide additional insights.

Our study has several limitations. MRI datasets acquired on eight different MRI scanners meeting the inclusion criteria were analyzed. Even if this realistically reflects the clinical setting, where radiologists also have to deal with images acquired on different machines, a potential bias remains because consistent interpretation of different image impressions by the readers cannot be taken for granted. A coronal T2-weighted FS sequence is standard in many MRI examinations of the neck and was therefore included in our analysis, although it is not specifically mentioned in the NI-RADS criteria. In this respect, the methodology used in our single-center study may differ from approaches used by other investigators in figure studies. The subgroup of non-SCC entities was very heterogeneous in our study, as was the distribution of primary cancer sites. A subanalysis was therefore not reasonable, but could provide important results in future adequately powered for such an analysis.

In conclusion, the results of our study suggest that NI-RADS categories can be statistically reliable when used in the interpretation of CEMRI. While interreader agreement for NI-RADS categories of the neck is acceptable, there is still need for improving assessment of the primary site, especially for NI-RADS categories of 2a, 2b, and 3. This limitation could be overcome by supplementary DWI, for which we found high interreader agreement.

Abbreviations

ADC:

Apparent diffusion coefficient

CE:

Contrast-enhanced

DWI:

Diffusion-weighted imaging

FS:

Fat-suppressed

HN:

Head and neck

NI-RADS:

Neck Imaging Reporting and Data System

SCC:

Squamous cell carcinoma

T1w:

T1-weighted

T2w:

T2-weighted

References

  1. 1.

    Bray F, Ferlay J, Soerjomataram I et al (2018) Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin 68:394–424

    PubMed  PubMed Central  Google Scholar 

  2. 2.

    Cohen N, Fedewa S, Chen AY (2018) Epidemiology and demographics of the head and neck cancer population. Oral Maxillofac Surg Clin North Am 30:381–395

    Article  Google Scholar 

  3. 3.

    Faggioni L, Coppola F, Ferrari R et al (2017) Usage of structured reporting in radiological practice: results from an Italian online survey. Eur Radiol 27:1934–1943

    Article  Google Scholar 

  4. 4.

    Syed F, Spector ME, Cornelius R et al (2016) Head and neck reconstructive surgery: what the radiologist needs to know. Eur Radiol 26:3345–3352

    Article  Google Scholar 

  5. 5.

    Dodd GD 3rd, Allen B Jr, Birzniek D et al (2015) Reengineering the radiology enterprise: a summary of the 2014 Intersociety Committee Summer Conference. J Am Coll Radiol 12:228–234

    Article  Google Scholar 

  6. 6.

    Aiken AH, Farley A, Baugnon KL et al (2016) Implementation of a novel surveillance template for head and neck cancer: Neck Imaging Reporting and Data System (NI-RADS). J Am Coll Radiol 13:743–746.e741

    Article  Google Scholar 

  7. 7.

    An JY, Unsdorfer KML, Weinreb JC (2019) BI-RADS, C-RADS, CAD-RADS, LI-RADS, Lung-RADS, NI-RADS, O-RADS, PI-RADS, TI-RADS: reporting and data systems. Radiographics 39:1435–1436

    Article  Google Scholar 

  8. 8.

    D’Orsi CJ, Kopans DB (1997) Mammography interpretation: the BI-RADS method. Am Fam Physician 55(1548-1550):1552

    Google Scholar 

  9. 9.

    Aiken AH, Rath TJ, Anzai Y et al (2018) ACR Neck Imaging Reporting and Data Systems (NI-RADS): a white paper of the ACR NI-RADS Committee. J Am Coll Radiol 15:1097–1108

    Article  Google Scholar 

  10. 10.

    Juliano AF, Aiken AH (2019) NI-RADS for head and neck cancer surveillance imaging: what, why, and how. Cancer Cytopathol. https://doi.org/10.1002/cncy.22214

  11. 11.

    Ailianou A, Mundada P, De Perrot T et al (2018) MRI with DWI for the detection of posttreatment head and neck squamous cell carcinoma: why morphologic MRI criteria matter. AJNR Am J Neuroradiol 39:748–755

    CAS  Article  Google Scholar 

  12. 12.

    Landis JR, Koch GG (1977) The measurement of observer agreement for categorical data. Biometrics 33:159–174

    CAS  Article  Google Scholar 

  13. 13.

    Elsholtz FHJ, Ro SR, Shnayien S et al (2020) Inter- and intrareader agreement of NI-RADS in the interpretation of surveillance contrast-enhanced CT after treatment of oral cavity and oropharyngeal squamous cell carcinoma. AJNR Am J Neuroradiol 41:859–865

    CAS  Article  Google Scholar 

  14. 14.

    Irshad A, Leddy R, Ackerman S et al (2016) Effects of changes in BI-RADS density assessment guidelines (fourth versus fifth edition) on breast density assessment: intra- and interreader agreements and density distribution. AJR Am J Roentgenol 207:1366–1371

    Article  Google Scholar 

  15. 15.

    Smith CP, Harmon SA, Barrett T et al (2019) Intra- and interreader reproducibility of PI-RADSv2: a multireader study. J Magn Reson Imaging 49:1694–1703

    Article  Google Scholar 

  16. 16.

    Viera AJ, Garrett JM (2005) Understanding interobserver agreement: the kappa statistic. Fam Med 37:360–363

    PubMed  Google Scholar 

  17. 17.

    Feinstein AR, Cicchetti DV (1990) High agreement but low kappa: I. The problems of two paradoxes. J Clin Epidemiol 43:543–549

    CAS  Article  Google Scholar 

  18. 18.

    Eisenhauer EA, Therasse P, Bogaerts J et al (2009) New response evaluation criteria in solid tumours: revised RECIST guideline (version 1.1). Eur J Cancer 45:228–247

    CAS  Article  Google Scholar 

  19. 19.

    Krieger DA, Hudgins PA, Nayak GK et al (2017) Initial performance of NI-RADS to predict residual or recurrent head and neck squamous cell carcinoma. AJNR Am J Neuroradiol 38:1193–1199

    CAS  Article  Google Scholar 

  20. 20.

    Sheikhbahaei S, Taghipour M, Ahmad R et al (2015) Diagnostic accuracy of follow-up FDG PET or PET/CT in patients with head and neck cancer after definitive treatment: a systematic review and meta-analysis. AJR Am J Roentgenol 205:629–639

    Article  Google Scholar 

  21. 21.

    Ho AS, Tsao GJ, Chen FW et al (2013) Impact of positron emission tomography/computed tomography surveillance at 12 and 24 months for detecting head and neck cancer recurrence. Cancer 119:1349–1356

    Article  Google Scholar 

  22. 22.

    Dinkelborg P, Ro SR, Shnayien S et al (2020) Retrospective evaluation of NI-RADS for detecting post-surgical recurrence of oral squamous cell carcinoma on surveillance CT or MRI. AJR Am J Roentgenol. https://doi.org/10.2214/AJR.20.24209

Download references

Funding

The authors state that this work has not received any funding.

Author information

Affiliations

Authors

Corresponding author

Correspondence to Fabian Henry Jürgen Elsholtz.

Ethics declarations

Guarantor

The scientific guarantor of this publication is Fabian Henry Jürgen Elsholtz.

Conflict of interest

Bernd Hamm declares relationships with the following companies: Abbott, Actelion Pharmaceuticals, Bayer Schering Pharma, Bayer Vital, BRACCO Group, Bristol Myers Squibb, Charité Research Organisation GmbH, Deutsche Krebshilfe, Dt. Stiftung für Herzforschung, Essex Pharma, EU Programmes (Past President of ESR), Fibrex Medical Inc., Focused Ultrasound Surgery Foundation, Fraunhofer Gesellschaft, Guerbet, INC Research, InSightec Ltd., IPSEN Pharma, Kendle/MorphoSys AG, Lilly GmbH, Lundbeck GmbH, MeVis Medical Solutions AG, Nexus Oncology, Novartis, Parexel CRO Service, Perceptive, Pfizer GmbH, Phillips, Sanofi-Aventis S.A., Siemens, Spectranetics GmbH, Terumo Medical Corporation, TNS Healthcare GmbH, Toshiba, UCB Pharma, Wyeth Pharma, Zukunftsfonds Berlin (TSB).

Stefan Markus Niehues declares relationships with the following companies: Vital Images, Canon Medical Systems, Guerbet, Bracco Imaging, Teleflex/Vidacare.

The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.

Statistics and biometry

No complex statistical methods were necessary for this paper.

Informed consent

Written informed consent was obtained from all patients in this study.

Ethical approval

Institutional Review Board approval was obtained.

Methodology

• retrospective

• diagnostic or prognostic study

• performed at one institution

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

ESM 1

(DOCX 20 kb)

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Elsholtz, F.H.J., Erxleben, C., Bauknecht, HC. et al. Reliability of NI-RADS criteria in the interpretation of contrast-enhanced magnetic resonance imaging considering the potential role of diffusion-weighted imaging. Eur Radiol (2021). https://doi.org/10.1007/s00330-021-07693-4

Download citation

Keywords

  • Head and neck cancer
  • Magnetic resonance imaging
  • Surveillance
  • Reliability of results