Deep learning in interstitial lung disease—how long until daily practice

Abstract

Interstitial lung diseases are a diverse group of disorders that involve inflammation and fibrosis of interstitium, with clinical, radiological, and pathological overlapping features. These are an important cause of morbidity and mortality among lung diseases. This review describes computer-aided diagnosis systems centered on deep learning approaches that improve the diagnostic of interstitial lung diseases. We highlighted the challenges and the implementation of important daily practice, especially in the early diagnosis of idiopathic pulmonary fibrosis (IPF). Developing a convolutional neuronal network (CNN) that could be deployed on any computer station and be accessible to non-academic centers is the next frontier that needs to be crossed. In the future, early diagnosis of IPF should be possible. CNN might not only spare the human resources but also will reduce the costs spent on all the social and healthcare aspects of this deadly disease.

Key Points

• Deep learning algorithms are used in pattern recognition of different interstitial lung diseases.

• High-resolution computed tomography plays a central role in the diagnosis and in the management of all interstitial lung diseases, especially fibrotic lung disease.

• Developing an accessible algorithm that could be deployed on any computer station and be used in non-academic centers is the next frontier in the early diagnosis of idiopathic pulmonary fibrosis.

Introduction

Interstitial lung diseases (ILDs) denote over 200 diverse lung disorders that involve inflammation and fibrosis of interstitium, with overlapping clinical, radiological, and pathological features, representing an important morbidity and mortality cause [1].

High-resolution computed tomography (HRCT) is the main method in ILD diagnosis, due to the lung tissue–specific radiation attenuation properties and maximum spatial resolution. The imaging data are evaluated by various textural patterns in the lung window extent and distribution. The assessment focuses on the image’s gray tones and geometrical structures, effectively a repetitive pattern matching problem, creating the perfect context for using computer-aided diagnosis (CAD) systems.

CAD enables medical practitioners to understand and utilize various imagistic investigations, with the help of information technology (IT) techniques [2]. The aim is to improve the diagnosis time and accuracy, with IT acting as support or even as an independent diagnostic option [3]. The CAD algorithms belong to artificial intelligence (AI), emulating human thinking [4]. The ILD diagnosis is essentially an algorithm with the following workflow: in comprehensive history, if physical examination findings and paraclinical investigations (chest X-ray, measurements of lung function, usual and specific blood tests) present a suspicion for ILD, a HRCT is performed [5]. The human factor intervenes next, by verifying the resulted data validity and, if no problems/artifacts are found, searching for patterns in specific locations. If there are clear findings, a diagnosis can be formulated, but if the results are inconclusive, the outcome is a list of possibilities requiring further discussions and more complex, invasive investigations. The individual performing this algorithms is therefore critical to accurate and speedy diagnosis, injecting an overall intrinsic variability. Since CAD can emulate the algorithm, it would make it an ideal choice for this stage, eliminating variances.

This paper aims to offer an in-depth analysis of the way virtual AI improves ILD diagnosis, with an emphasis on convolutional neural networks (CNNs).

Computer-aided diagnosis history

AI component’s virtual subclass is machine learning (Fig. 1), comprising mathematical algorithms used by computer systems in order to learn a specific task through experiences, without specific human instructions [6].

Fig. 1
figure1

Artificial intelligence progression diagram (AI, artificial intelligence; CNN, convolutional neural network)

The advancement of that is deep learning (DL), consisting of a multi-layer representation learning architecture. The representation activates the first layer of neurons through a sensor, which, in turn, activates the next layer by complex connections. Each layer processes the representation in a non-linear way, creating an increasingly complex schema, diverging from the general machine learning task-specific algorithm [7, 8].

DL’s major advantage is that it can improve autonomously, without human input. From a usage standpoint, it can perform arbitrary parallel computation more efficiently than other algorithms [9, 10]. DL is used in visual object recognition [7], speech recognition [11], driving assistance [12], and language classification [10] among others.

The first algorithm successfully used for pattern recognition was neocognitron, in 1980, which integrated neurophysiological architecture [8, 13, 14]. The key for successful feature extraction is in creating an appropriate network architecture, as shown by the apparition of backpropagation technique, in 1989, which allowed handwritten digit recognition, becoming a landmark reference [15].

CNNs require large, balanced datasets and advanced algorithms, reflecting into processing power and storage capacity requirements [8, 9]. Krizhevsky et al developed a CNN model named Alex Net by gathering the biggest database for training: 1.2 million images. The algorithm was able to classify the images into 1000 nature categories with the smallest possible error rate [16], making it the state-of-the-art database for training CNNs.

Performance wise, the first neural network to achieve superhuman performance in visual pattern recognition (http://people.idsia.ch/~juergen/superhumanpatternrecognition.html) appeared in 2011, when Ciresan et al used a deep neural network on graphics processing unit to recognize traffic signs images. In the last decade, graphics processing unit improvements made possible shorter computation time for complex operations in a common setting, flourishing CNN development [8, 17].

Each CNN has a complex architecture with an initial image input as pixel array from a receptive field with several hidden computational connection layers afterward [6, 7] as detailed in Fig. 2.

Fig. 2
figure2

Convolutional neural network architecture

The convolutional layer, CNN’s main component, consists of multiple, weighted individual filters [7]. The multiple filter sets detect different patterns in images. The detection progresses from small patterns such as corners, lines, and edges to shapes and objects.

For a CNN to perform, it requires multiple layer types and transmission between them [6]. The final stage predicts the image category probabilities, determining the strongest and the most relevant active feature class [16].

The earliest CNN application in the Healthcare field dates into the early 1990s. Lo et al applied the CNN algorithm for lung nodule detection on chest X-rays, reaching a true-positive detection rate of 80% [18]. Sahiner et al used CNNs to classify mass from normal breast tissue on mammograms, with a 90% positive value [19]. In 2008, CNNs successfully detected hippocampal sclerosis on brain MRI [20]. These first studies’ high positive rates are biased due to the small database and easily detected lesion types.

There are multiple challenges in acquiring medical images for using in deep learning:

  1. (i)

    They are difficult and costly to obtain, compared with common images.

  2. (ii)

    To validly annotate bio-images, specialists must be used.

  3. (iii)

    The medical database volume is generally insufficient, while the state-of-the-art image analysis datasets (ImageNet, AlexNet, GoogLeNet, VGGNet) contain thousands or even millions of natural image instances.

A workaround could be transfer-learning, where weights from a trained CNN on a nature dataset are conveyed to another CNN with a different dataset [6, 21].

Even though natural images and medical images are greatly different, the former being colored, whereas X-rays, CT images, and MRI images are gray-scaled, they all share the same descriptors. Histograms of oriented gradient and scale-invariant feature transform have been successfully applied to medical image segmentation and detection. Bar et al validated this in chest pathology by applying CNNs trained on non-medical image datasets to 93 chest X-ray images [22]. The area under the curve (AUC) was 0.93 for right pleural effusion detection, 0.89 for heart enlargement detection, and 0.79 for classifying normal versus abnormal chest X-rays. All values are well above 0.5, showing excellent model skill.

Starting in 2014, CNN applications in bio-imaging research flourished in segmentation, detection, and classification applications like lung nodule detection and classification, colon polyp detection, coronary calcification detection [23, 24], skin cancer classification [25], knee cartilage segmentation [26], brain tumor segmentation [27], and breast lesions classification [28, 29]. At present, the intention is to achieve higher accuracy and better performance than that of a human counterpart.

Although AI research has been predominantly directed to neurology [30], oncology, and cardiovascular diseases [6], as the first three major death causes, the chest imaging field also presents interest for lung nodule detection and classification [31, 32], tuberculosis lesion classification [33], lesion detection and classification [34], and parenchyma pulmonary disease classification [35].

Interstitial lung disease–specific CAD

Typical ILD patterns in (HR) CT images are reticulation (RE), honeycombing (HC), ground-glass opacity (GGO), consolidation (CD), micro-nodules (MN), emphysema (EM), or combinations of the above. The difficulty appears when the results are combined or inconclusive (Fig. 3)

Fig. 3
figure3

Difficult ILD patterns. a NL in subject 1. b EM in subject 2. c RE in subject 3. d HC in subject 4. e Mixed HC/RE in subject 5. f Mixed RE/GGO in subject 6. Source: “Victor Babes” Database

CNNs need large image samples because normal lung or different tissue categories could exhibit similar appearances (Fig. 3a, b or c, d), while significant variations might be seen between different subjects for the same tissue class (Fig. 3c–e).

Anthimopoulos M. et al [36] designed and trained one of the first CNN to classify the most common ILD patterns, achieving an 85.5% classification performance, therefore demonstrating the DL recognition capacity for lung tissue idiosyncrasy. Experienced radiologists annotated 120 HRCT by excluding ambivalent lung areas and the bronco-vascular tree which were used afterwards for training and testing the CNN. The proposed algorithm achieved superior performance compared with the state-of-the-art methods at that time (Alex Net, VGG-Net-D), mainly because of the hyper-parameters developed for the ILD pattern characterization. However, a misclassification rate was found between HC and RE due to their common fibrotic nature as shown in Annex 1 in the Supplementary Material. The combination between GGO/RE and individual GGO and RE pattern had also a high misclassification rate due to the overlapping appearance. Clinically, this reflects the significance in differentiating idiopathic pulmonary fibrosis (IPF) vs. non-specific interstitial pneumonia, as an accurate description of texture apart of gray-scale intensity value.

Dealing with these challenges, Christodoulidis S. et al [21] presented a CNN architecture that could extract the textural variability of ILD patterns. Using transfer learning from multiple non-medical source databases, they achieved an unsatisfactory increase of only 2% in the CNN performance. The big downfall in [21] is the usage of CT images, instead of HRCT.

Applications that use HRCT are only a few, like [37,38,39], all using CNNs for categorization. Even if Li et al [38] and Li et al [39] go a step further by using a custom architecture in an unsupervised algorithm, their performance is not as good as other options (e.g., [21, 40]).

How complex the CNN should be?

Kim et al [40] compared shallow learning (SL) with DL in pattern classification. In their study, they used 4 convolutional layers and 2 fully connected layers for the CNN architecture that proved to have a significantly better accuracy from 81.27 to 95.12% by only increasing the number of convolutional layers. This lowered the misclassification rate between ambiguous cases such as HC/RE (Fig. 3e) or NL/EM and emphasized that a higher complexity of DL methods should be applied for a better ILD diagnostic.

Distinguishing between different lung tissue patterns on HRCT images is a challenging task, especially when using small samples in region of interest (ROI)–based classification. This could lead to mismatches since the lung tissue may have the same appearance between different tissue categories with great variation in the same category for different patients. Image processing consists not only in the gray-scale differences but also in object detection, regardless of variant parameters. Considering that patient movement during scanning and different types of breathing might affect lung volume size, Wang Q. et al [41] proposed a multi-scale rotation-invariant CNN algorithm to overcome this bias. This approach uses a Gabor filter, which analyzes the specific frequency and directions in a localized region, making it more like the human visual cortex. The performance accuracy achieved by this algorithm in classifying all the ILD patterns was greater than 85%, up to 90% for N, GGO, and MN patterns. The error rate has decreased by increasing the number of the CNN layers, similar to the previous study [40]. The downfall of this approach is the exponential complexity in the Gabor filter implementation, which requires significant resources to compute.

All other studies [21, 36, 40, 41] that focused on classifying ILD patterns employed a patch-based image representation method. Their pitfall consisted in the very small image sections (~ 31 × 31pixels), which resulted in fine detail loss. Furthermore, the image-patch needed to be manually annotated, creating an arduous process for radiologists.

By recognizing the impediment in manual identification of ROIs for automated pulmonary CAD systems, Gao M. et al [42] tried a different approach in ILD pattern classification. They proposed a holistic image recognition method, based on the gray-scale level, similar to emphysema quantification [43], but with greater autonomy. This perspective captured more details and used slice level image labels or tags without outlining the ILD regions. Rescaling the CT image in Hounsfield unit (HU), the method expresses three different attenuation scales in regard to the lung ILD pattern: low attenuation pattern (HU = − 1400 and − 950) such as EM, high attenuation pattern (HU = − 160 and 240) for CD, and normal lung attenuation (HU = − 1400 and 200). The holistic approach achieved 68.6% accuracy, less than the patch-based classification with 87.9% accuracy. However, the overall results are misleading: the holistic method classified EM perfectly; its difficulties were in separating normal lung (NL), MN, and CD pattern. The perfect EM classification warrants further exploration, maybe in a mixed method approach.

The dataset remains the Achilles heel in all of these approaches. An interesting approach was proposed by Bae HJ. et al [44], by making an infinite number of arbitrary different ILD patterns from 2D HRCT images which increased the accuracy of CNN in classifying lung tissue patterns. The algorithm prevented over-fitting, stabilizing accuracy loss for the validation set and providing a diverse mix of ILD patterns. Accuracy on specific region of interest or on the whole lung was 89.5%, higher than the conventional CNN data augmentation (82.1%), close to the human’s capacity. Best results were obtained for NL, GGO, RE, and EM. One of the algorithm’s drawback scan be randomization of ILD patterns which cannot mathematically insure a hypothetical perfect accuracy. Also, the iterative nature of the algorithm requires considerable computation resources, unfit for the normal computer.

Idiopathic pulmonary fibrosis—the challenge of all ILDs

HRCT plays a central part both in the diagnosis and in management of all interstitial lung diseases, in particular in fibrotic lung disease. In an appropriate clinical context, idiopathic pulmonary fibrosis (IPF) diagnosis can be made without surgical lung biopsy, when the HRCT features are of usual interstitial pneumonia (UIP) [45]. Based on growing evidence, a statement from the Fleischner Society expanded this recommendation to include patients with features of probable UIP [46, 47]. Despite this framework, the ILD radiological assessment is still a challenge, due to substantial inter-observer variability, even between experienced radiologists [48, 49].

In clinical practice, this can be a challenge because imaging expertise is not always available, specifically in non-academic centers. This could lead to diagnosis delay and unnecessary interventional investigations, like surgical lung biopsy, which might pose unacceptable risks, especially for the older patients with advanced disease.

To overcome these limitations, Walsh et al [35] proposed a CAD that could easily be deployed on standard computing equipment.

A total of 1157 HRCT scans underwent a pre-processing step to create a maximum of 500 unique four-slice montages (concatenations) per CT scan resulting in a multiplied image dataset of 420,096 unique montages for the training algorithm and 40,490 for the validation set. The specific neural network architecture used in this study was the convolutional neural network Inception-ResNet-v2 [50, 51]. Each HRCT was classified by an experienced thoracic radiologist in one of the three categories: UIP, possible UIP, or inconsistent with UIP [45], with the specific diagnostic prediction outcome. A specific optimization algorithm was used to adjust the network’s internal parameters and to reduce the scan errors, making the neural network training an interactive process. The algorithm accuracy was 76.4%, tested on 139 HRCT (68,093 unique test montages), with 92.7% of diagnoses within one category. The algorithm needed 2.31 s to evaluate 150 four-slice montages.

This algorithm [35] was clinically tested on a second 150 HRCT scan cohort with fibrotic lung diseases. Numerous patients with a multidisciplinary team diagnosis of IPF, chronic fibrotic hypersensitivity pneumonitis, or connective tissue disease–related fibrotic interstitial lung disease were evaluated by 91 thoracic radiologists (not involved in the training process). The performance against the radiologists’ opinion was an average of 73.3% (93.3% within one category). The median accuracy of the thoracic radiologists on this cohort was 70.7%. By providing reproducible, almost instantaneous reporting with human-level accuracy, this algorithm stands as an important diagnosing tool for IPF.

Since the UIP pattern is known to be related to high mortality rates in ILD, the labeling of UIP vs. non-UIP is very important. The algorithm and the radiologists’ majority opinion provided equally prognostic discrimination (p = 0.62) between these two groups [35]. When Fleischner criteria for IPF diagnostic were taken into consideration, a good inter-observer agreement between algorithm and the radiologists was noticed. Although CNN is not trained to recognize basal honeycombing as a distinguishing feature of UIP, it seems that it can learn how to recognize it. This autonomous behavior might provide a framework for discovery novel image biomarker in fibrotic lung disease. The problem with this algorithm is that it considers only one tissue subtype in one slice, eliminating all the mixed patterns. The huge advantage, however, is its availability and low local resource requirements.

Discussion

One major conundrum in developing high-accuracy deep learning algorithms for fibrotic lung disease diagnostics is the lack of large imaging datasets for training. To overcome this problem, a centralized imaging repository needs to be created through an international collaborative effort. Not only that, but the images need to be unified under a desirable format, HRCT. There are too few algorithms that deal with these image types like [35, 37,38,39] and since the 1-mm-thick slices can show lesions otherwise omitted, this format is a necessity. Since the images will, more than likely, be gathered from multiple sources, the resolution, gray-scale, and annotations need a unified format, also.

A handle-able CNN that could be deployed on any computer station and accessible to non-academic centers is the “Next Generation” of AI in clinical practice. Namely, in the field of ILDs, AI can help to differentiate and early diagnose the patients with the most severe forms, i.e., IPF. Early diagnosis in IPF will lead to targeted antifibrotic treatment which substantially prolongs the survival and reduces acute exacerbations which are not only deadly but also costly [52,53,54,55,56]. In order to reach such high expectations, a hybrid algorithm should be developed. Since some algorithms show exquisite accuracy in some areas, like [42], a combination of multiple CNN can be the answer to reducing the costs spent on the social and healthcare aspects [57, 58]. The CNN combination could have different configurations and start forming the same input or, on the contrary, have similar structures and different inputs. Even more interesting would be a combination of different AI techniques, like CNN with clusterization and classification algorithms, maybe in different stages, parallel or subsequent. No matter what the approach is, the purpose is to keep local area computing to a minimum, leaning therefore towards a cloud architecture style. In this case, legal aspects, like data privacy and security, begin to play an important role and the communication between the local and computational nodes could be obstructed by security protocols.

Conclusions

In this review, we describe the deep learning algorithm developments and their implication in the medical field, especially in the ILD diagnostic. We highlighted the challenges, but also the implementation options that would, one day, lead to daily practice, with a clinical implication in early diagnosis of ILDs. Developing a CAD that could be deployed on any computer station and be accessible to non-academic centers is the next frontier in the early diagnosis of IPF.

Abbreviations

AI:

Artificial intelligence

AUC:

Area under the curve

CAD:

Computer-aided diagnosis

CD:

Consolidation

CNN:

Convolutional neuronal network

CT:

Computer tomography

DL:

Deep learning

EM:

Emphysema

GGO:

Ground-glass opacity

HC:

Honeycombing

HRCT:

High-resolution computed tomography

HU:

Hounsfield unit

ILD:

Interstitial lung diseases

IPF:

Idiopathic pulmonary fibrosis

IT:

Information technology

ML:

Machine learning

MN:

Micro-nodules

MRI:

Magnetic resonance imaging

NL:

Normal lung

RE:

Reticulation

ROI:

Region of interest

UIP:

Usual interstitial pneumonia

References

  1. 1.

    Raghu G, Weycker D, Edelsberg J, Bradford WZ, Oster G (2006) Incidence and prevalence of idiopathic pulmonary fibrosis. Am J Respir Crit Care Med 174(7):810–816

  2. 2.

    Doi K (2007) Computer-aided diagnosis in medical imaging: historical review, current status and future potential. Comput Med Imaging Graph 31(4-5):198–211

    Article  Google Scholar 

  3. 3.

    Takahashi R, Kajikawa Y (2017) Computer-aided diagnosis: a survey with bibliometric analysis. Int J Med Inform 101:58–67

    Article  Google Scholar 

  4. 4.

    Hamet P, Tremblay J (2017) Artificial intelligence in medicine. Metabolism 69S:S36–S40. https://doi.org/10.1016/j.metabol.2017.01.011

    CAS  Article  PubMed  Google Scholar 

  5. 5.

    Meye KC (2014) Diagnosis and management of interstitial lung disease. Transl Respir Med. https://doi.org/10.1186/2213-0802-2-4

  6. 6.

    Soffer S, Ben-Cohen A, Shimon O, Amitai MM, Greenspan H, Klang E (2019) Convolutional neural networks for radiologic images: a radiologist’s guide. Radiology 290(3):590–606

  7. 7.

    LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444

    CAS  Article  Google Scholar 

  8. 8.

    Schmidhuber J (2015) Deep learning in neural networks: an overview. Neural Netw 61:85–117

    Article  Google Scholar 

  9. 9.

    Murphy A, Skalski M, Gaillard F (2018) The utilisation of convolutional neural networks in detecting pulmonary nodules: a review. Br J Radiol 91(1090):20180028

    Article  Google Scholar 

  10. 10.

    Sutskever I, Vinyals O, Le QV (2014) Sequence to sequence learning with neural networks. Proceedings of the 27th International Conference on Neural Information Processing Systems 2:3104–3112

  11. 11.

    Mikolov T, Deoras A, Povey D, Burget L, Cernocky J (2011) Strategies for training large scale neural network language models. 2011 IEEE Workshop on Automatic Speech Recognition & Understanding 196–201

  12. 12.

    Cireşan D, Meier U, Masci J, Schmidhuber S (2012) Multi-column deep neural network for traffic sign classification. Neural Netw 32:333–338

  13. 13.

    Hubel DH, Wiesel TN (1962) Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex. J Physiol 160(1):106–154.2

    CAS  Article  Google Scholar 

  14. 14.

    Fukushima K (1980) Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybern 36(4):193–202

    CAS  Article  Google Scholar 

  15. 15.

    LeCun Y, Boser B, Denker JS et al (1990) Handwritten digit recognition with a back-propagation network. Adv Neural Inf Process Syst 2:396–404

  16. 16.

    Krizhevsky A, Sutskever I, Hintonet GE (2011) ImageNet classification with deep convolutional neural networks. NIPS’12 Proceedings of the 25th International Conference on Neural Information Processing Systems 1:1097–1105

  17. 17.

    Lee JG, Jun S, Cho YW et al (2017) Deep learning in medical imaging: general overview. Korean J Radiol 18(4):570–584

    Article  Google Scholar 

  18. 18.

    Lo SCB, Lin JS, Freedman MT, Mun SK (1993) Computer-assisted diagnosis of lung nodule detection using an artificial convolution neural network. Proc SPIE Int Soc Opt Eng 1898:859–869

  19. 19.

    Sahiner B, Chan HP, Petrick N et al (1996) Classification of mass and normal breast tissue: a convolution neural network classifier with spatial domain and texture images. IEEE Trans Med Imaging 15(5):598–610

    CAS  Article  Google Scholar 

  20. 20.

    Döhler F, Mormann F, Weber B, Elger CE, Lehnertz K (2008) A cellular neural network based method for classification of magnetic resonance images: towards an automated detection of hippocampal sclerosis. J Neurosci Methods 170(2):324–331

  21. 21.

    Christodoulidis S, Anthimopoulos M, Ebner L et al (2017) Multi-source transfer learning with convolutional neural networks for lung pattern analysis. IEEE J Biomed Health Inform 21(1):76–84

    Article  Google Scholar 

  22. 22.

    Bar Y, Diamant I, Wolf L et al (2015) Deep learning with non-medical training used for chest pathology identification. Proc. SPIE 9414, Medical Imaging 2015: Computer-Aided Diagnosis

  23. 23.

    Shin HC, Roth HR, Gao M et al (2016) Deep convolutional neural networks for computer aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging 35(5):1285–1298

    Article  Google Scholar 

  24. 24.

    Suzuki K (2017) Overview of deep learning in medical imaging. Radiol Phys Technol 10(3):257–273

    Article  Google Scholar 

  25. 25.

    Esteva A, Kuprel B, Novoa RA et al (2017) Dermatologist-level classification of skin cancer with deep neural networks. Nature 542(7639):115–118

    CAS  Article  Google Scholar 

  26. 26.

    Prasoon A, Petersen K, Igel C, Lauze F, Dam E, Nielsen M (2013) Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network. Med Image Comput Comput Assist Interv 16(Pt 2):246–253

  27. 27.

    Pereira S, Pinto A, Alves V, Silva CA (2016) Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans Med Imaging 35(5):1240–1251

  28. 28.

    Han S, Kang HK, Jeong JY et al (2017) A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys Med Biol 62(19):7714–7728

    Article  Google Scholar 

  29. 29.

    Kooi T, Litjens G, van Ginneken B et al (2017) Large scale deep learning for computer aided detection of mammographic lesions. Med Image Anal 35:303–312

    Article  Google Scholar 

  30. 30.

    Jiang F, Jiang Y, Zhi H et al (2017) Artificial intelligence in healthcare: past, present and future. Stroke Vasc Neurol 2(4):230–243

    Article  Google Scholar 

  31. 31.

    Ciompi F, de Hoop B, van Riel SJ et al (2015) Automatic classification of pulmonary peri-fissural nodules in computed tomography using an ensemble of 2D views and a convolutional neural network out-of-the-box. Med Image Anal 26(1):195–202

    Article  Google Scholar 

  32. 32.

    Tu X, Xie M, Gao J et al (2017) Automatic categorization and scoring of solid, part-solid and non-solid pulmonary nodules in CT images with convolutional neural network. Sci Rep 7(1):8533

    Article  Google Scholar 

  33. 33.

    Lopes UK, Valiati JF (2017) Pre-trained convolutional neural networks as feature extractors for tuberculosis detection. Comput Biol Med 89:135–143

    CAS  Article  Google Scholar 

  34. 34.

    Cicero M, Bilbily A, Colak E et al (2017) Training and validating a deep convolutional neural network for computer-aided detection and classification of abnormalities on frontal chest radiographs. Invest Radiol 52(5):281–287

    Article  Google Scholar 

  35. 35.

    Walsh SLF, Calandriello L, Silva M, Sverzellati N (2018) Deep learning for classifying fibrotic lung disease on high-resolution computed tomography: a case-cohort study. Lancet Respir Med. https://doi.org/10.1016/S2213-2600(18)3

  36. 36.

    Anthimopoulos M, Christodoulidis S, Ebner L, Christe A, Mougiakakou S (2016) Lung pattern classification for interstitial lung diseases using a deep convolutional neural network. IEEE Trans Med Imaging 35(5):1207–1216

  37. 37.

    Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. Comput Vis Pattern Recognit arXiv:1409–1556

  38. 38.

    Li Q, Cai W, Wang X, Zhou Y, Feng DD, Chen M (2014) Medical image classification with convolutional neural network. IEEE. 13th International Conference on Control Automation Robotics & Vision (ICARCV) pp 844–848

  39. 39.

    Li Q, Cai W, Feng DD (2013) Lung image patch classification with automatic feature learning. Conf Proc IEEE Eng Med Biol Soc 2013:6079–6082

  40. 40.

    Kim GB, Jung KH, Lee Y et al (2018) Comparison of shallow and deep learning methods on classifying the regional pattern of diffuse lung disease. J Digit Imaging 31(4):415–424

    Article  Google Scholar 

  41. 41.

    Wang Q, Zheng Y, Yang G, Jin W, Chen X, Yin Y (2018) Multiscale rotation-invariant convolutional neural networks for lung texture classification. IEEE J Biomed Health Inform 22(1):184–195

  42. 42.

    Gao M, Bagci U, Lu L et al (2018) Holistic classification of CT attenuation patterns for interstitial lung diseases via deep convolutional neural networks. Comput Methods Biomech Biomed Eng Imaging Vis 6(1):1–6

    Article  Google Scholar 

  43. 43.

    Wang Z, Gu S, Leader JK et al (2013) Optimal threshold in CT quantification of emphysema. Eur Radiol 23(4):975–984

    Article  Google Scholar 

  44. 44.

    Bae HJ, Kim CW, Kim N et al (2018) A perlin noise-based augmentation strategy for deep learning with small data samples of HRCT images. Sci Rep 8(1):17687

    CAS  Article  Google Scholar 

  45. 45.

    Raghu G, Collard HR, Egan JJ et al (2011) An official ATS/ERS/JRS/ALAT statement: idiopathic pulmonary fibrosis: evidence-based guidelines for diagnosis and management. Am J Respir Crit Care Med 183:788–824

    Article  Google Scholar 

  46. 46.

    Lynch DA, Sverzellati N, Travis WD et al (2018) Diagnostic criteria for idiopathic pulmonary fibrosis: a Fleischner Society White Paper. Lancet Respir Med 6:138–153

    Article  Google Scholar 

  47. 47.

    Gruden JF (2016) CT in idiopathic pulmonary fibrosis: diagnosis and beyond. AJR Am J Roentgenol 206:495–507

    Article  Google Scholar 

  48. 48.

    Walsh SL, Calandriello L, Sverzellati N, Wells AU, Hansell DM, UIP Observer Consort (2016) Interobserver agreement for the ATS/ERS/JRS/ALAT criteria for a UIP pattern on CT. Thorax 71:45–51

  49. 49.

    Walsh SLF, Wells AU, Desai SR et al (2016) Multicentre evaluation of multidisciplinary team meeting agreement on diagnosis in diffuse parenchymal lung disease: a case-cohort study. Lancet Respir Med 4:557–565

    Article  Google Scholar 

  50. 50.

    Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, Inception-ResNet and the impact of residual connections on learning. Thirty-First AAAI Conference on Artificial Intelligence; San Francisco, CA

  51. 51.

    Kaiming H, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV; June 27–30, 2016, pp 770–78

  52. 52.

    Costabel U, Albera C, Lancaster LH et al (2017) An open-label study of the long-term safety of pirfenidone in patients with idiopathic pulmonary fibrosis (RECAP). Respiration 94(5):408–415

    CAS  Article  Google Scholar 

  53. 53.

    Richeldi L, Du Bois RM, Raghu G et al (2014) Efficacy and safety of nintedanib in idiopathic pulmonary fibrosis. N Engl J Med 370(22):2071–2082

    Article  Google Scholar 

  54. 54.

    Crestani B, Huggins JT, Kaye M et al (2019) Long-term safety and tolerability of nintedanib in patients with idiopathic pulmonary fibrosis: results from the open-label extension study, INPULSIS-ON. Lancet Respir Med 7(1):60–68

    CAS  Article  Google Scholar 

  55. 55.

    Costabel U, Albera C, Glassberg MK et al (2019) Effect of pirfenidone in patients with more advanced idiopathic pulmonary fibrosis. Respir Res 20(1):55

    Article  Google Scholar 

  56. 56.

    Raghu G, van den Blink B, Hamblin MJ et al (2019) Long-term treatment with recombinant human pentraxin 2 protein in patients with idiopathic pulmonary fibrosis: an open-label extension study. Lancet Respir Med. https://doi.org/10.1016/S2213-2600(19)3017

  57. 57.

    Ley B, Bradford WZ, Weycker D, Vittinghoff E, du Bois RM, Collard HR (2015) Unified baseline and longitudinal mortality prediction in idiopathic pulmonary fibrosis. Eur Respir J 45:1374–1381

  58. 58.

    Raimundo K, Chang E, Broder MS, Alexander K, Zazzali J, Swigris JJ (2016) Clinical and economic burden of idiopathic pulmonary fibrosis: a retrospective cohort study. BMC Pulm Med 16:2

Download references

Funding

The authors state that this work has not received any funding.

Author information

Affiliations

Authors

Corresponding author

Correspondence to Diana Manolescu.

Ethics declarations

Guarantor

None.

Conflict of interest

The authors of this manuscript declare no relationships with any companies whose products or services may be related to the subject matter of the article.

Statistics and biometry

No complex statistical methods were necessary for this paper.

Informed consent

Written informed consent was not required for this study.

Ethical approval

Institutional Review Board approval was not required.

Methodology

• Review article

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

ESM 1

(DOCX 19 kb)

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Trusculescu, A.A., Manolescu, D., Tudorache, E. et al. Deep learning in interstitial lung disease—how long until daily practice. Eur Radiol (2020). https://doi.org/10.1007/s00330-020-06986-4

Download citation

Keywords

  • Interstitial lung disease
  • Idiopathic pulmonary fibrosis
  • Deep learning