Previous Article in Journal
Immune Checkpoint Blockade Response in Mucinous Tubular and Spindle Cell Carcinoma
Previous Article in Special Issue
Deep-Learning-Based Reconstruction of Single-Breath-Hold 3 mm HASTE Improves Abdominal Image Quality and Reduces Acquisition Time: A Quantitative Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Deep Learning-Based (3D U-Net Model) Automated Pulmonary Nodule Detection Tool for CT Imaging

1
Department of Imaging, The Clatterbridge Cancer Centre NHS Foundation Trust, Liverpool L7 8YA, UK
2
Faculty of Health and Life Sciences, University of Liverpool, Liverpool L69 3BX, UK
3
Department of Radiodiagnosis and Imaging, Tata Memorial Hospital, Homi Bhabha National Institute, Mumbai 400012, India
4
Endimension Technology Pvt Ltd., Maharashtra 400076, India
5
Department of Surgical Oncology, Tata Memorial Hospital, Mumbai 400012, India
6
Department of Medical Oncology, Tata Memorial Hospital, Mumbai 400012, India
7
Department of Pathology, Tata Memorial Hospital, Mumbai 400012, India
*
Author to whom correspondence should be addressed.
Curr. Oncol. 2025, 32(2), 95; https://doi.org/10.3390/curroncol32020095 (registering DOI)
Submission received: 24 November 2024 / Revised: 29 January 2025 / Accepted: 29 January 2025 / Published: 8 February 2025

Abstract

:
Background: Precise detection and characterization of pulmonary nodules on computed tomography (CT) is crucial for early diagnosis and management. Objectives: In this study, we propose the use of a deep learning-based algorithm to automatically detect pulmonary nodules in computed tomography (CT) scans. We evaluated the performance of the algorithm against the interpretation of radiologists to analyze the effectiveness of the algorithm. Materials and Methods: The study was conducted in collaboration with a tertiary cancer center. We used a collection of public (LUNA) and private (tertiary cancer center) datasets to train our deep learning models. The sensitivity, the number of false positives per scan, and the FROC curve along with the CPM score were used to assess the performance of the deep learning algorithm by comparing the deep learning algorithm and the radiology predictions. Results: We evaluated 491 scans consisting of 5669 pulmonary nodules annotated by a radiologist from our hospital; our algorithm showed a sensitivity of 90% and with only 0.3 false positives per scan with a CPM score of 0.85. Apart from the nodule-wise performance, we also assessed the algorithm for the detection of patients containing true nodules where it achieved a sensitivity of 0.95 and specificity of 1.0 over 491 scans in the test cohort. Conclusions: Our multi-institutional validated deep learning-based algorithm can aid radiologists in confirming the detection of pulmonary nodules through computed tomography (CT) scans and identifying further abnormalities and can be used as an assistive tool. This will be helpful in national lung screening programs guiding early diagnosis and appropriate management.

1. Introduction

Lung cancer is one of the most frequent cancers in the world, being the most commonly diagnosed cancer and the leading cause of cancer-related mortality. It is the most regularly diagnosed cancer in the male population in India and the fifth most usually diagnosed cancer in the female population [1,2]. The outcome of lung cancer remains the poorest among all the 5-year survival rates of about 10–20%, the rates being largely dependent on the stage of detection ranging from 0 to 97% from later stages to the earliest ones. The proposed reason for its low survival is its aggressive nature and detection in advanced stages [3]. Therefore, early detection gives an important chance to improve overall survival; however, its use has had limited success with less sensitivity. Many studies have used low-dose MDCT and shown significant benefits compared to the rest of the available methods like chest radiography and sputum cytology [4,5]. The detection and characterization of the pulmonary nodules are well established and performed by trained authorized radiologists in a clinical setup. However, in the recent ongoing trend of mass lung cancer screening trials or in a setup with a high patient load, the already limited number of trained radiologists can be a limiting factor in detecting these pulmonary lesions among hundreds of scans. This can lead to an increase in the rates of missed findings due to multiple reasons including nodule characteristics, image quality, or perception error by the radiologist, which could be due to scanning, recognition, or decision-making errors in addition to reasons like reading conditions, fatigue, or distraction [6,7,8]. This highlights the need for technological advancements to reduce workload and improve nodule screening sensitivity while minimizing false positives [9]. Deep learning-based computer-aided detection (CAD) systems have been developed and tested to assist radiologists by automatically detecting and locating pulmonary nodules, offering promising AI-driven methods for image analysis [10]. By exploring these new pathways, radiologists are likely to play a leading role in medical applications of AI [11]. AI encompasses complex processes like hypothesis generation, data analysis, testing, and program creation through machine learning (ML), neural networks, and deep learning. ML, a subset of AI, enables computers to learn from data without explicit programming and is widely used in medical imaging.
Deep learning algorithms are new in AI research and do not require explicit feature definition, representing a fundamentally different paradigm in machine learning [12]. With ample data and computational power, deep learning algorithms have advanced significantly, learning directly from data without predefined features. Among various architectures, convolutional neural networks (CNNs) are the most widely used in medical imaging [13].
Rafael Wiemker from Philips Research highlighted CAD’s potential in early cancer detection and nodule diagnosis based on morphology and volume changes, noting that improved spatial and temporal CT resolution enhances algorithm performance in pulmonary nodule detection [14]. Conventional CAD systems showed low sensitivity and high false positives, prompting a shift to advanced deep learning models. These systems provide secondary predictions to aid final decision-making [15,16,17]. Our study aims at developing such an algorithm using a 3D deep convolutional neural network and further testing to validate its performance in a given dataset to increase its diagnostic performance in routine practice.

2. Materials and Methods

The study was conducted in collaboration with a tertiary cancer center and the study was approved by the Institutional Ethics Committee. The need for obtaining informed consent was waived, and data collection and storage were performed in accordance with the local guidelines. Data were anonymized to keep the personal information of patients confidential. The study was performed in accordance with the ethical guidelines outlined in the Declaration of Helsinki, Good Clinical Practice guidelines, and the Indian Council of Medical Research guidelines. The study is not registered in the clinical trial registry. No funding was used for the purpose of this study.

2.1. Variables

To measure the performance of our deep learning model, we used measures like specificity, sensitivity, FROC curve, and CPM score.

2.2. Definitions

Specificity: Fraction of positive ground truths covered by the predictions.
Sensitivity: Fraction of negative ground truths covered by the predictions.
FROC curve: Free-response receiver operating characteristic (FROC) analysis [18]. In the FROC curve, sensitivity is plotted as a function of the average number of false positives per scan (FPs/scan).
CPM score: The competition performance metric (CPM) is defined as the average sensitivity at seven predefined FPs/scan rates: 1/8, 1/4, 1/2, 1,2,4, 8.

2.3. Study Methodology

We used the standard procedure for evaluating deep learning-based model systems in this study. We used a dataset provided by our institute and a publicly available dataset. The entire AI pipeline consists of different stages and training models with different types of loss functions.
Our entire AI pipeline consists of two stages of processing the computed tomography (CT) scans. The first stage predicts and segments the pulmonary nodules using a deep convolutional neural network (DCNN). The DCNN architecture with which we decided to segment the pulmonary nodules is a variant of U-Net [17], a state-of-the-art deep convolutional network. The first stage is referred to as the detection stage. The second stage uses a similar U-Net-like deep convolutional network as the first stage, but it is coupled with a classification output to determine whether the nodule identified by the first stage is an actual nodule. The second stage is referred to as the classification stage. This network takes predictions from the first-stage network and confirms or rejects the nodule predictions.
Out of the total dataset provided by our institute, the scans were divided into the training (176 scans), validation (44 scans), and test (491 scans) cohorts. The final evaluation was performed on the test cohort while we used the training and validation cohorts for fine-tuning our algorithm.

2.4. Datasets

In general, deep learning models require a large amount of data for them to learn and generalize to the task. Public datasets play a huge role in the development of deep learning algorithms, and, hence, we collaborated with a tertiary cancer care center to form a private dataset. Available datasets were used to train the models at different stages of the AI pipeline.
The data used in this study were taken from two sources. A total of 1599 patients were selected for the study. Out of these, 711(44.7%) patients were taken from the tertiary cancer care center and the remaining 888 (55.3%) patients were part of the LIDC-IDRI dataset. The Lung Image Database Consortium image collection (LIDC-IDRI) consists of diagnostic and lung cancer screening thoracic computed tomography (CT) scans with marked-up annotated pulmonary nodules [18]. Table 1 shows the distribution of cases among the different cohorts.
The 491 test cases were solely part of the 711 cases from the tertiary cancer care center. We observed that the patients provided by our hospital had a higher number of pulmonary nodules per patient as compared to the LIDC-IDRI dataset. The pulmonary nodule distribution is given in Table 2. The X axis is divided into three sections: data received from the institution, data from LUNA16 which is an open-source dataset, and total cases. Each section has two columns: the number of nodules present in the dataset and the number of unique patients with nodules considered in the study. The Y axis shows that during model training, the data are divided mainly into three sections. The column denotes the sample sizes used in the training, validation and testing phases of the model development. The nodule diameter in this study ranges from 3 mm to 30 mm. The annotations provided by radiologists and the LIDC-IDRI competition were in the form of a CSV containing x, y, and z center coordinates of the nodule along with the diameter of a nodule in mm. We decided to train a segmentation network to detect the nodules following participation in the Lung Nodule Analysis (LUNA16) challenge [19]. We converted the annotations into a segmentation mask by drawing a 3D sphere with the provided diameter and center coordinates. The pulmonary nodules in the training and validation cohorts of the hospital dataset were annotated by a team of radiologists with over six years of experience. The annotations were conducted using medical-grade BARCO monitors, which were routinely calibrated to ensure optimal display quality. All cases were independently reviewed and annotated by the team; however, granular details regarding the number of annotations performed by an individual were not available. The team members were blinded to the information from the other imaging modalities to minimize bias. Annotations were made by providing them with a tool to draw bounding boxes around the center slice of the nodule. Later, using the bounding box as a reference, we extracted similar information to the publicly available dataset to obtain the segmentation mask. Similarly, the test cohort was annotated by a different set of radiologists from our institute which was then used for the evaluation of the model.

2.5. Stages of AI Pipeline

First-Stage Network: The first-stage network was trained to detect and segment the pulmonary nodules present in the dataset. We used a deep convolutional neural network (DCNN) architecture known as U-Net. The network performed downsampling until it reached a 4 × 4 × 4 shape, resulting in 4 group blocks in the downsampling path and 4 blocks in upsampling path. Each block had a residual connection making the network called the Residual U-Net [20] as shown in Figure 1. Each group block consisted of 2 convolutional layers with Batch Norm. We used the Relu activation function. Downsampling was performed with Maxpooling for each group of blocks with a 2 × 2 × 2 kernel. Upsampling was performed with the help of nn. Upsampling was performed with a scale factor of 2. The network took a 3D image as an input and then output the pixel-wise segmentation mask of the same size. The segmentation network was chosen to exploit the ability to segment each voxel of the nodule, providing the precise shape, centroid, and volume of the nodule. This information provides a wide range of characteristics that could help to better diagnose the nodules. We used dice loss to train the first stage. Since we did not use an object detection network to predict the bounding box coordinates, the term “detection” used here is in reference to the ability to determine the location of each nodule based on the blobs predicted by the segmentation network.
Second-Stage Network: We trained the second-stage network as a classification network to reduce the false-positive nodule predictions coming from the first stage. The network takes the nodule predicted by the first stage as the input and predicts whether to classify the input nodule as a true positive or false positive. Classifying the nodule as a true positive will keep the prediction made by the first stage in the final output; otherwise, it will reject it. The network architecture is similar to that of the first stage which predicts two outputs, which are a segmentation mask for the input and a classification of the input.
The network was trained to optimize the average of dice loss for training the segmentation output and cross-entropy loss for the classification output. As this was a classification task, more weightage was given to classification loss.

2.6. Training

Convolutional neural networks cannot interpret voxel spacings natively, which is why we preprocessed our dataset by resampling all cases to a common median voxel spacing of [0.703125 × 0.703125 × 1.25] [20,21,22]. As the dataset consisted of a wide range of voxel spacing, we resized the training data into median spacing in order to provide more standardized input data for network training, which leads to better generalization. Typically, while training the networks, constraints like GPU memory might limit the size of the input that can be processed by a deep convolutional neural network. Hence, for the first stage, we trained the network with a 128 × 128 ×128 dimension which, in our experience, was able to retain enough spatial information to detect the nodule while optimally using the resources. We trained the first-stage network for 30 epochs starting with a 1e-3 learning rate and a batch size of 4 using the Adam optimizer. Figure 2 indicates the training and validation curves for stage 1. Training the network took 4 h per epoch on Nvidia GeForce RTX 2080Ti GPU.
We used the same preprocessing methods for training the second-stage network. The input to the network was taken by cropping the area containing the detected nodule out of the CT scan and resizing it to the input size of 36 × 36 × 20. The network, similar to the first stage, was trained with Adam optimizer and a learning rate of 1 × 10−3. The training took 15 min per epoch and used 25 epochs. The training and validation curves are shown in Figure 3. The validation loss for the second-stage training does not appear stable, which could be the result of the smaller number of total nodules included in the validation, leading to each nodule gaining a higher weight during the loss calculation, hence creating unstable loss.

3. Results

We evaluated our deep learning model with 491 chest computed tomography (CT) scans provided by our hospital. Out of these, 401 scans contained 5669 nodules and the remaining 90 scans did not show any nodule.

3.1. Per-Lesion Performance

The sensitivity for pulmonary nodule detections predicted by the 3D U-Net model was calculated using the annotations provided by an experienced radiologist. We also used free-response receiver operating characteristic (FROC) analysis. In the FROC curve, sensitivity was plotted as a function of the average number of false positives per scan. Figure 4 shows the FROC curve of our results. We used the FROC curve to decide on the threshold for the classification network which optimizes between the sensitivity and the number of false positives per scan. Our network achieved 0.90 sensitivity at only 0.3 false positives per scan as shown in Table 3. The X axis, the column labeled “Per-lesion performance”, denotes the performance of the total number of nodules included in the study across multiple metrics. Similarly, “Patient-wise performance” denotes the performance of the total number of unique patients involved in the study across multiple metrics.
The Y axis consists of various metrics to gauge the performance of the model on nodules and patients, namely, true positive (TP), false negative (FN), false positive (FP), true negative (TN), sensitivity, and specificity.
We also analyzed the nodule results based on the number of nodules present in a patient’s scan to explore the network’s prediction on different parameters. Table 4 shows a comparison between the patients’ scans which contain a nodule count between 1 and 6 as well as a higher number of nodules. As Table 4 suggests, our method achieved higher sensitivity irrespective of the number of nodules present in a patient’s scan. Our model was not able to perform well with nodules less than 3 mm in size, which according to the LUNA16 challenge [19] was not considered relevant for lung cancer screening protocols.
An overall score was calculated by using the CPM value. The CPM value was defined as the average of the sensitivity at seven predefined false-positive rates: 1/8, 1/4, 1/2, 1, 2, 4, and 8 FPs/scan, as shown in Table 5. Our network had a CPM score of 0.85.

3.2. Patient-Wise Performance

We analyzed the model performance based on patients to give us a clearer idea about the performance for patients with no nodules. We considered a patient as a positive scan when it had at least one nodule present; otherwise, it was negative. Table 3 shows that our model performed well, missing only 13 patients out of 401 scans giving a sensitivity of 0.95. Our model was able to identify all the negative scans without producing any false positives, achieving a specificity of 1.0.
Out of those thirteen patients, where our network did not predict nodules, nine patients had one nodule, three patients had two nodules, and one patient had three nodules. We also calculated the performance for patients who had nodule counts higher than six. If we predicted a number greater than six nodules in any of the scans, we considered that scan as a true positive. If the predicted nodule count was lower than seven, then it was considered a false negative. On a total of two hundred forty-nine patients, our model was able to achieve a patient-wise 0.99 sensitivity with two hundred forty-seven patients as true positives and two as false negatives.

4. Discussion

Our study is based on a subset of the 3D convolutional neural network called the U-Net, in which, to our knowledge, we have used for the first time a large number of nodules from a public dataset as well as from a tertiary cancer care center, which provides a varied spectrum of lung nodules to evaluate the performance of the algorithm in a better way, nullifying the biases relating to nodule characteristics. Our proposed AI pipeline achieved 90% sensitivity at only 0.3 false positives per scan over 491 CT scans provided by our institute. Since we have used a novel dataset, and given the fact that most of the previously published studies used only the LIDC-IDRI dataset, an exact comparison with the other studies would not be accurate or possible due to the unmatched datasets. Nonetheless, our model performed better with a better sensitivity and a significantly lower number of false positives per scan as compared to most of the published studies (Appendix A). This successful application demonstrates the effectiveness of using the 3D CNN U-Net for lung nodule candidate detection. The performance of the proposed scheme is comparable with that of other methods in the literature and is better than many of those. Among the usage of private datasets, Guo et al. [23] and Liu et al. [24] used small datasets comprising 29 scans (34 true nodules) and 32 scans (33 solitary nodules), respectively, for the evaluation of their proposed systems while Murphy et al. [25] used a larger private dataset of 813 scans for the evaluation of their proposed system and achieved a sensitivity of 80% with 4.2 FPs/scan, in which they used local image features and the k-nearest-neighbor classification. Their system’s sensitivity was lower than ours, despite their enormous dataset. In actual settings with a greater range of nodule types encountered in clinical scans, it is expected that systems with fewer datasets may perform poorly. Qin et al. [26] used a similar 3D CNN model as the backbone for their region proposal network (RPN), namely the U-Net architecture. Their research had a sensitivity of 98.2 percent with just four FPs/scan; but we had a significantly lower number of false positives with a substantially higher number of nodules. Our study has found better sensitivity and accuracy even with a higher number of nodules per patient, which is a special advantage in non-screening settings, for example, in an oncology setup dealing with various spectra of lung nodules in both primary cancer and metastases. Apart from this, in the testing dataset, our model was shown to perform better in cases with no nodules with a specificity of 100% signifying its huge impact for patients with negative findings, reducing the burden of additional investigations. We also found out that when using the private dataset, the performance of the model increases more than when just training with the publicly available datasets, as indicated in Table 6 and Table 7.

Retraining Without Public Dataset

The network that was trained without the public dataset (just using data from our institute) achieved 0.83 sensitivity on the test data. For the same FPs/scan, training after adding the public dataset achieved 0.9 sensitivity. The CPM score also reduced with just training with the publicly available dataset, achieving a score of 0.787, while the network trained with the data provided by our institute achieved a score of 0.85. The highest sensitivity with eight false positives per scan was 0.857 without using a dataset from the tertiary cancer care center, shown in Figure 5. Our algorithm precisely provides the number of nodules present in the scan with the location in terms of the x, y, and z coordinates and the diameter of each nodule. This assists the radiologist in easily assessing the nodule and helps identify the potential risks.
In Figure 6, we present some cases of false positives, false negatives, and true positives predicted by our AI pipeline, respectively. This example also provides a peek into the reasoning of our deep learning model in predicting the nodule in a CT scan. The red markings shown in Figure 6 represent annotations by radiologists from our institute and the green markings show the segmentation mask provided by the AI pipeline. It is apparent that our AI pipeline is making mistakes in detecting nodules in some cases. Most of the false positives are coming from fibrotic bands, and veins in the chest CT scans which create an appearance like a nodule.
The main limitation of our study is that it is restricted to detection of the nodule; however, most of the recently performed studies focus on detection and characterization, which forms the future perspective of the model. Another limitation is that we have used a two-stage detector which is slower and can take up more memory than a single-stage detector; however, it comes with the tradeoff of having higher accuracy than single-stage detectors. This forms a future challenge for us: to achieve similar accuracy with a faster single-stage detector. To improve upon the current proposed study, additional contextual information about the nodules, such as linkages with nearby blood vessels, and information about the patient, such as the medical history report, can be added. In the next work, we will concentrate on making even more enhancements in order to better serve the medical community.

5. Conclusions

The findings show that our AI pipeline can easily aid the radiologist in better detection of pulmonary nodules and their characteristics in chest computed tomography (CT) scans. Automated accurate detection of these nodules will help in applications in mass screening programs, which will be beneficial in directing early diagnosis and the appropriate management of these nodules. The limitation of this study is its restriction to nodule detection, along with the other limitation of having used a two-stage detector which is slower and can take up more memory than a single-stage detector. These limitations form a future challenge for us: to expand beyond nodule detection and to achieve similar accuracy with a faster single-stage detector.

Author Contributions

Conceptualization, A.M. (Abhishek Mahajan) and R.A.; formal analysis, B.K., A.M. (Apparao Madiraju) and V.P. (Vivek Punia); investigation, R.V.; methodology, A.M. (Abhishek Mahajan) and R.A.; resources, P.B.; software, B.K., A.M. (Apparao Madiraju) and V.P. (Vivek Pawar); supervision, V.N., K.P., P.C. and S.G.; validation, B.K., A.M. (Apparao Madiraju), V.P. (Vivek Pawar) and V.P. (Vivek Punia); visualization, R.M.A., V.M.P. and S.G.; writing—original draft, A.M. (Abhishek Mahajan) and R.A.; writing—review and editing, R.A., U.A., R.M.A., R.V., V.M.P., V.N., A.J., N.M., K.P., P.C. and S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of Tata Memorial Hospital (protocol code 900635 and date of approval: 2 March 2021).

Informed Consent Statement

Informed consent to participate in the study was obtained from all participants. Informed consent to publication was obtained from relevant participants.

Data Availability Statement

The datasets generated in this study are available on request to Abhishek Mahajan ([email protected]).

Conflicts of Interest

Authors Bharadwaj Komaravolu, Apparao Madiraju, Vivek Pawar, and Vivek Punia were employed by the company Endimension Technology Pvt Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Appendix A

Table A1. Comparison among different CAD schemes.
Table A1. Comparison among different CAD schemes.
FP per ScanSensitivity (%)DatasetNo. of NodulesTotal ScansYearSource
0.390LIDC, Private768315992021Our Study
493.6LIDC11868882019Fu et al. [27]
490.7LIDC11868882017Chen et al. [28]
488.9LIDC6985022017Ma et al. [29]
487.9LIDC11868882016Arinda et al. [30]
3.185.2LIDC631982015Lu et al. [31]
6.897.5LIDC148842014Choi et al. [32]
275LIDC681082014Brown et al. [33]
4.280LIDC103842013Teramoto et al. [34]
6.197LIDC148842012Cascio et al. [27]
487.5LIDC801252011Tan et al. [35]
4.280Private15188132009Murphy et al. [25]
5.676LIDC, Private241852009Sahiner et al. [36]
NA94.7Private34292009Guo et al. [23]
4.693.7Private33322009Liu et l. [24]
0.4848Private33NA2010Sousa et al. [37]
4.880.3Private121632003Suzuki et al. [38]
880LIDC17499492015Torres et al. [39]
297PrivateNA122012Mabrouk et al. [40]
2.2795.2LIDC151582013Choi et al. [32]
NA95.3LIDC50472016Akram et al. [41]
395.7LIDC67310102020Wang et al. [42]
490.1LIDC11868882016Setio et al. [30]
1080SPIE-AAPM LUNGNA672016Anirudh et al. [43]
494.4LIDC11868882017Ding et al. [44]
0.789.2LIDC11868882018Gruetzemac-her et al. [45]
295.2LIDC11668882019Kim et al. [46]
498.2LIDC11868882019Qin et al. [26]

References

  1. Barta, J.A.; Powell, C.A.; Wisnivesky, J.P. Global Epidemiology of Lung Cancer. Ann. Glob. Health 2019, 85, 8. [Google Scholar] [CrossRef]
  2. Prabhash, K.; Noronha, V.; Dikshit, R.; Raut, N.; Joshi, A.; Pramesh, C.; George, K.; Agarwal, J.; Munshi, A. Epidemiology of lung cancer in India: Focus on the differences between non-smokers and smokers: A single-centre experience. Indian J. Cancer 2012, 49, 74–81. [Google Scholar] [CrossRef] [PubMed]
  3. Mok, T.; Camidge, D.; Gadgeel, S.; Rosell, R.; Dziadziuszko, R.; Kim, D.-W.; Pérol, M.; Ou, S.-H.; Ahn, J.; Shaw, A.; et al. Updated overall survival and final progression-free survival data for patients with treatment-naive advanced ALK-positive non-small-cell lung cancer in the ALEX study. Ann. Oncol. 2020, 31, 1056–1064. [Google Scholar] [CrossRef]
  4. National Lung Screening Trial Research Team; Aberle, D.R.; Adams, A.M.; Berg, C.D.; Black, W.C.; Clapp, J.D.; Fagerstrom, R.M.; Gareen, I.F.; Gatsonis, C.; Marcus, P.M.; et al. Reduced Lung-Cancer Mortality with Low-Dose Computed Tomographic Screening. N. Engl. J. Med. 2011, 365, 395–409. [Google Scholar] [CrossRef]
  5. Manser, R.L.; Irving, L.B.; Byrnes, G.; Abramson, M.J.; A Stone, C.; A Campbell, D. Screening for lung cancer: A systematic review and meta-analysis of controlled trials. Thorax 2003, 58, 784–789. [Google Scholar] [CrossRef]
  6. Pinsky, P.F. Lung cancer screening with low-dose CT: A world-wide view. Transl. Lung Cancer Res. 2018, 7, 234–242. [Google Scholar] [CrossRef] [PubMed]
  7. Gierada, D.S.; Black, W.C.; Chiles, C.; Pinsky, P.F.; Yankelevitz, D.F. Low-Dose CT Screening for Lung Cancer: Evidence from 2 Decades of Study. Radiol. Imaging Cancer 2020, 2, e190058. [Google Scholar] [CrossRef]
  8. Liang, M.; Tang, W.; Xu, D.M.; Jirapatnakul, A.C.; Reeves, A.P.; Henschke, C.I.; Yankelevitz, D. Low-Dose CT Screening for Lung Cancer: Computer-aided Detection of Missed Lung Cancers. Radiology 2016, 281, 279–288. [Google Scholar] [CrossRef] [PubMed]
  9. Christe, A.; Leidolt, L.; Huber, A.; Steiger, P.; Szucs-Farkas, Z.; Roos, J.; Heverhagen, J.; Ebner, L. Lung cancer screening with CT: Evaluation of radiologists and different computer assisted detection software (CAD) as first and second readers for lung nodule detection at different dose levels. Eur. J. Radiol. 2013, 82, e873–e878. [Google Scholar] [CrossRef]
  10. Thrall, J.H.; Li, X.; Li, Q.; Cruz, C.; Do, S.; Dreyer, K.; Brink, J. Artificial Intelligence and Machine Learning in Radiology: Opportunities, Challenges, Pitfalls, and Criteria for Success. J. Am. Coll. Radiol. 2018, 15, 504–508. [Google Scholar] [CrossRef] [PubMed]
  11. Frontiers | Human, All Too Human? An All-Around Appraisal of the “Artificial Intelligence Revolution” in Medical Imaging. Available online: https://www.frontiersin.org/journals/psychology/articles/10.3389/fpsyg.2021.710982/full (accessed on 12 January 2025).
  12. Miotto, R.; Wang, F.; Wang, S.; Jiang, X.; Dudley, J.T. Deep learning for healthcare: Review, opportunities and challenges. Briefings Bioinform. 2017, 19, 1236–1246. [Google Scholar] [CrossRef] [PubMed]
  13. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed]
  14. Gilbert, F.J.; Lemke, H. Computer-Aided Diagnosis. Br. J. Radiol. 2005, 78, S1–S2. [Google Scholar] [CrossRef]
  15. Jin, H.; Li, Z.; Tong, R.; Lin, L. A deep 3D residual CNN for false-positive reduction in pulmonary nodule detection. Med. Phys. 2018, 45, 2097–2107. [Google Scholar] [CrossRef] [PubMed]
  16. Yuan, H.; Fan, Z.; Wu, Y.; Cheng, J. An efficient multi-path 3D convolutional neural network for false-positive reduction of pulmonary nodule detection. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 2269–2277. [Google Scholar] [CrossRef]
  17. Polat, H.; Danaei Mehr, H. Classification of Pulmonary CT Images by Using Hybrid 3D-Deep Convolutional Neural Network Architecture. Appl. Sci. 2019, 9, 940. [Google Scholar] [CrossRef]
  18. Zhao, Y.; Zhao, L.; Yan, Z.; Wolf, M.; Zhan, Y. A Deep-Learning Based Automatic Pulmonary Nodule Detection System. In Medical Imaging 2018: Computer-Aided Diagnosis; SPIE: Houston, TX, USA, 2018. [Google Scholar] [CrossRef]
  19. Schilham, A.M.; van Ginneken, B.; Loog, M. A computer-aided diagnosis system for detection of lung nodules in chest radiographs with an evaluation on a public database. Med. Image Anal. 2005, 10, 247–258. [Google Scholar] [CrossRef]
  20. Setio, A.A.A.; Traverso, A.; de Bel, T.; Berens, M.S.; Bogaard, C.v.D.; Cerello, P.; Chen, H.; Dou, Q.; Fantacci, M.E.; Geurts, B.; et al. Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: The LUNA16 challenge. Med. Image Anal. 2016, 42, 1–13. [Google Scholar] [CrossRef] [PubMed]
  21. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  22. Pehrson, L.M.; Nielsen, M.B.; Lauridsen, C.A. Automatic Pulmonary Nodule Detection Applying Deep Learning or Machine Learning Algorithms to the LIDC-IDRI Database: A Systematic Review. Diagnostics 2019, 9, 29. [Google Scholar] [CrossRef] [PubMed]
  23. Orozco, H.M.; Villegas, O.O.V.; Domínguez, H.d.J.O.; Sánchez, V.G.C. Lung Nodule Classification in CT Thorax Images Using Support Vector Machines. In Proceedings of the 2013 12th Mexican International Conference on Artificial Intelligence, Mexico City, Mexico, 24–30 November 2013; pp. 277–283. [Google Scholar]
  24. Lu, L.; Liu, W. A Method of Pulmonary Nodules Detection with Support Vector Machines. In Proceedings of the 2008 Eighth International Conference on Intelligent Systems Design and Applications 2008, Taiwan, China, 26–28 November 2008; Volume 1, p. 35. [Google Scholar]
  25. Murphy, K.; van Ginneken, B.; Schilham, A.; de Hoop, B.; Gietema, H.; Prokop, M. A large-scale evaluation of automatic pulmonary nodule detection in chest CT using local image features and k-nearest-neighbour classification. Med. Image Anal. 2009, 13, 757–770. [Google Scholar] [CrossRef]
  26. Qin, Y.; Zheng, H.; Zhu, Y.-M.; Yang, J. Simultaneous Accurate Detection of Pulmonary Nodules and False Positive Reduction Using 3D CNNs. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 1005–1009. [Google Scholar]
  27. Cascio, D.; Magro, R.; Fauci, F.; Iacomi, M.; Raso, G. Automatic detection of lung nodules in CT datasets based on stable 3D mass–spring models. Comput. Biol. Med. 2012, 42, 1098–1109. [Google Scholar] [CrossRef] [PubMed]
  28. Dou, Q.; Chen, H.; Yu, L.; Qin, J.; Heng, P.-A. Multilevel Contextual 3-D CNNs for False Positive Reduction in Pulmonary Nodule Detection. IEEE Trans. Biomed. Eng. 2016, 64, 1558–1567. [Google Scholar] [CrossRef]
  29. Ma, J.; Zhou, Z.; Ren, Y.; Xiong, J.; Fu, L.; Wang, Q.; Zhao, J. Computerized detection of lung nodules through radiomics. Med. Phys. 2017, 44, 4148–4158. [Google Scholar] [CrossRef]
  30. Setio, A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; van Riel, S.J.; Wille, M.M.; Naqibullah, M.; Sanchez, C.I.; van Ginneken, B. Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef] [PubMed]
  31. Lu, L.; Tan, Y.; Schwartz, L.H.; Zhao, B. Hybrid detection of lung nodules on CT scan images. Med. Phys. 2015, 42, 5042–5054. [Google Scholar] [CrossRef] [PubMed]
  32. Choi, W.-J.; Choi, T.-S. Automated pulmonary nodule detection based on three-dimensional shape-based feature descriptor. Comput. Methods Programs Biomed. 2014, 113, 37–54. [Google Scholar] [CrossRef]
  33. Brown, M.S.; Lo, P.; Goldin, J.G.; Barnoy, E.; Kim, G.H.J.; McNitt-Gray, M.F.; Aberle, D.R. Toward clinically usable CAD for lung cancer screening with computed tomography. Eur. Radiol. 2014, 24, 2719–2728. [Google Scholar] [CrossRef]
  34. Teramoto, A.; Fujita, H. Fast lung nodule detection in chest CT images using cylindrical nodule-enhancement filter. Int. J. Comput. Assist. Radiol. Surg. 2012, 8, 193–205. [Google Scholar] [CrossRef] [PubMed]
  35. Tan, M.; Deklerck, R.; Jansen, B.; Bister, M.; Cornelis, J. A novel computer-aided lung nodule detection system for CT images. Med. Phys. 2011, 38, 5630–5645. [Google Scholar] [CrossRef]
  36. Sahiner, B.; Chan, H.-P.; Hadjiiski, L.M.; Cascade, P.N.; Kazerooni, E.A.; Chughtai, A.R.; Poopat, C.; Song, T.; Frank, L.; Stojanovska, J.; et al. Effect of CAD on Radiologists’ Detection of Lung Nodules on Thoracic CT Scans: Analysis of an Observer Performance Study by Nodule Size. Acad. Radiol. 2009, 16, 1518–1530. [Google Scholar] [CrossRef]
  37. Sousa, J.R.F.d.S.; Silva, A.C.; de Paiva, A.C.; Nunes, R.A. Methodology for automatic detection of lung nodules in computerized tomography images. Comput. Methods Programs Biomed. 2009, 98, 1–14. [Google Scholar] [CrossRef]
  38. Suzuki, K.; Armato, S.G.; Li, F.; Sone, S.; Doi, K. Massive training artificial neural network (MTANN) for reduction of false positives in computerized detection of lung nodules in low-dose computed tomography. Med. Phys. 2003, 30, 1602–1617. [Google Scholar] [CrossRef]
  39. Torres, E.L.; Fiorina, E.; Pennazio, F.; Peroni, C.; Saletta, M.; Camarlinghi, N.; Fantacci, M.E.; Cerello, P. Large scale validation of the M5L lung CAD on heterogeneous CT datasets. Med. Phys. 2015, 42, 1477–1489. [Google Scholar] [CrossRef]
  40. Mabrouk, M.; Karrar, A.; Sharawy, A. Computer Aided Detection of Large Lung Nodules Using Chest Computer Tomography Images. IJAIS 2012, 3, 12–18. [Google Scholar]
  41. Akram, S.; Javed, M.Y.; Akram, M.U.; Qamar, U.; Hassan, A. Pulmonary Nodules Detection and Classification Using Hybrid Features from Computerized Tomographic Images. J. Med. Imaging Health Inform. 2016, 6, 252–259. [Google Scholar] [CrossRef]
  42. Wang, B.; Si, S.; Zhao, H.; Zhu, H.; Dou, S. False positive reduction in pulmonary nodule classification using 3D texture and edge feature in CT images. Technol. Health Care 2021, 29, 1071–1088. [Google Scholar] [CrossRef]
  43. Anirudh, R.; Thiagarajan, J.J.; Bremer, T.; Kim, H. Lung nodule detection using 3D convolutional neural networks trained on weakly labeled data. In Proceedings of the Medical Imaging 2016: Computer-Aided Diagnosis, San Diego, CA, USA, 24 March 2016. [Google Scholar] [CrossRef]
  44. Ding, J.; Li, A.; Hu, Z.; Wang, L. Accurate Pulmonary Nodule Detection in Computed Tomography Images Using Deep Convolutional Neural Networks. In Medical Image Computing and Computer Assisted Intervention−MICCAI 2017; Descoteaux, M., Maier-Hein, L., Franz, A., Jannin, P., Collins, D., Duchesne, S., Eds.; Springer International Publishing: Cham, Switzerland, 2017; Volume 10435. [Google Scholar] [CrossRef]
  45. Gruetzemacher, R.; Gupta, A.; Paradice, D. 3D deep learning for detecting pulmonary nodules in CT scans. J. Am. Med. Inform. Assoc. 2018, 25, 1301–1310. [Google Scholar] [CrossRef]
  46. Kim, B.-C.; Yoon, J.S.; Choi, J.-S.; Suk, H.-I. Multi-scale gradual integration CNN for false positive reduction in pulmonary nodule detection. Neural Netw. 2019, 115, 1–10. [Google Scholar] [CrossRef]
Figure 1. Overview of the two-stage network pipeline.
Figure 1. Overview of the two-stage network pipeline.
Curroncol 32 00095 g001
Figure 2. Training and validation curves for the first stage.
Figure 2. Training and validation curves for the first stage.
Curroncol 32 00095 g002
Figure 3. Training and validation curves for the second stage.
Figure 3. Training and validation curves for the second stage.
Curroncol 32 00095 g003
Figure 4. Test FROC curve.
Figure 4. Test FROC curve.
Curroncol 32 00095 g004
Figure 5. FROC curve for the experimental dataset which excluded the private cohort.
Figure 5. FROC curve for the experimental dataset which excluded the private cohort.
Curroncol 32 00095 g005
Figure 6. (a) False-positive examples, (b) false-negative example, and (c) true-positive examples.
Figure 6. (a) False-positive examples, (b) false-negative example, and (c) true-positive examples.
Curroncol 32 00095 g006
Table 1. Breakdown of all the patients into training, validation, and test cohorts.
Table 1. Breakdown of all the patients into training, validation, and test cohorts.
OUR INSTITUTELUNA16Total Cases
Training 176710886
Validation 44178222
Test491-491
Total7118881599
Table 2. The number of nodules amongst patients in different cohorts.
Table 2. The number of nodules amongst patients in different cohorts.
OUR INSTITUTELUNA16Total Cases
#Nodules#Patients with Nodules#Nodules#Patients with Nodules#Nodules#Patient with Nodules
Training7431449294841672628
Validation8522257117342139
Test5669401--5669401
Table 3. Per-lesion and patient-wise performance of the network on the test set.
Table 3. Per-lesion and patient-wise performance of the network on the test set.
Per-Lesion PerformancePatient-Wise Performance
Total5669491
True Positive5079388
False Negative59013
False Positive1480
True Negative-90
FP per scan0.3-
Sensitivity0.90.95
Specificity-1
Table 4. Summary of the results of the number of nodules per scan.
Table 4. Summary of the results of the number of nodules per scan.
Nodule CountNumber of
Patients
SensitivityFPFNTP
1–61520.872458396
7–82490.91245324683
9–102270.8844126932
11–19870.88681471040
20–29480.89131281020
30–3980.92222261
40–4990.93729383
50–5970.89843338
60–5940.94416246
70–7930.88226192
80–8960.95024480
90–9910.8501690
≥10010.8801397
Table 5. False positives per scan vs. sensitivity on the test and validation sets.
Table 5. False positives per scan vs. sensitivity on the test and validation sets.
-0.1250.250.51248Mean
Test0.580.840.90.90.910.920.920.85
Validation0.630.720.830.860.870.870.870.807
Table 6. Per-lesion performance of the network which was trained without the inclusion of the public dataset.
Table 6. Per-lesion performance of the network which was trained without the inclusion of the public dataset.
Total True Nodules5669
True Positive4696
False Negative973
False Positive150
Sensitivity0.828
FP per scan0.3
Table 7. False positives per scan vs. sensitivity on the test set.
Table 7. False positives per scan vs. sensitivity on the test set.
-0.1250.250.51248Mean
Test0.530.770.830.8360.840.850.8570.787
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mahajan, A.; Agarwal, R.; Agarwal, U.; Ashtekar, R.M.; Komaravolu, B.; Madiraju, A.; Vaish, R.; Pawar, V.; Punia, V.; Patil, V.M.; et al. A Novel Deep Learning-Based (3D U-Net Model) Automated Pulmonary Nodule Detection Tool for CT Imaging. Curr. Oncol. 2025, 32, 95. https://doi.org/10.3390/curroncol32020095

AMA Style

Mahajan A, Agarwal R, Agarwal U, Ashtekar RM, Komaravolu B, Madiraju A, Vaish R, Pawar V, Punia V, Patil VM, et al. A Novel Deep Learning-Based (3D U-Net Model) Automated Pulmonary Nodule Detection Tool for CT Imaging. Current Oncology. 2025; 32(2):95. https://doi.org/10.3390/curroncol32020095

Chicago/Turabian Style

Mahajan, Abhishek, Rajat Agarwal, Ujjwal Agarwal, Renuka M. Ashtekar, Bharadwaj Komaravolu, Apparao Madiraju, Richa Vaish, Vivek Pawar, Vivek Punia, Vijay Maruti Patil, and et al. 2025. "A Novel Deep Learning-Based (3D U-Net Model) Automated Pulmonary Nodule Detection Tool for CT Imaging" Current Oncology 32, no. 2: 95. https://doi.org/10.3390/curroncol32020095

APA Style

Mahajan, A., Agarwal, R., Agarwal, U., Ashtekar, R. M., Komaravolu, B., Madiraju, A., Vaish, R., Pawar, V., Punia, V., Patil, V. M., Noronha, V., Joshi, A., Menon, N., Prabhash, K., Chaturvedi, P., Rane, S., Banwar, P., & Gupta, S. (2025). A Novel Deep Learning-Based (3D U-Net Model) Automated Pulmonary Nodule Detection Tool for CT Imaging. Current Oncology, 32(2), 95. https://doi.org/10.3390/curroncol32020095

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop