Next Article in Journal
Analysis of Quality Indicators of the Pre-Analytical Phase on Blood Gas Analyzers, Point-Of-Care Analyzer in the Period of the COVID-19 Pandemic
Previous Article in Journal
Age Estimation in 0–8-Year-Old Children in France: Comparison of One Skeletal and Five Dental Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Leveraging Deep Learning Decision-Support System in Specialized Oncology Center: A Multi-Reader Retrospective Study on Detection of Pulmonary Lesions in Chest X-ray Images

1
Carebot, Ltd., 128 00 Prague, Czech Republic
2
Faculty of Nuclear Sciences and Physical Engineering, Czech Technical University, 115 19 Prague, Czech Republic
3
Department of Radiology, Masaryk Memorial Cancer Institute, 602 00 Brno, Czech Republic
4
Faculty of Mathematics and Physics, Charles University, 121 16 Prague, Czech Republic
5
Faculty of Electrical Engineering, Czech Technical University, 166 36 Prague, Czech Republic
6
Department of Artificial Intelligence, Daeyang AI Center, Sejong University, Seoul 050 06, Republic of Korea
7
Department of Imaging Methods, Motol University Hospital, 150 06 Prague, Czech Republic
8
Department of Radiodiagnosis, Podripska City Hospital, 413 01 Roudnice nad Labem, Czech Republic
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(6), 1043; https://doi.org/10.3390/diagnostics13061043
Submission received: 2 February 2023 / Revised: 27 February 2023 / Accepted: 7 March 2023 / Published: 9 March 2023
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
Chest X-ray (CXR) is considered to be the most widely used modality for detecting and monitoring various thoracic findings, including lung carcinoma and other pulmonary lesions. However, X-ray imaging shows particular limitations when detecting primary and secondary tumors and is prone to reading errors due to limited resolution and disagreement between radiologists. To address these issues, we developed a deep-learning-based automatic detection algorithm (DLAD) to automatically detect and localize suspicious lesions on CXRs. Five radiologists were invited to retrospectively evaluate 300 CXR images from a specialized oncology center, and the performance of individual radiologists was subsequently compared with that of DLAD. The proposed DLAD achieved significantly higher sensitivity (0.910 (0.854–0.966)) than that of all assessed radiologists (RAD 10.290 (0.201–0.379), p < 0.001, RAD 20.450 (0.352–0.548), p < 0.001, RAD 30.670 (0.578–0.762), p < 0.001, RAD 40.810 (0.733–0.887), p = 0.025, RAD 50.700 (0.610–0.790), p < 0.001). The DLAD specificity (0.775 (0.717–0.833)) was significantly lower than for all assessed radiologists (RAD 11.000 (0.984–1.000), p < 0.001, RAD 20.970 (0.946–1.000), p < 0.001, RAD 30.980 (0.961–1.000), p < 0.001, RAD 40.975 (0.953–0.997), p < 0.001, RAD 50.995 (0.985–1.000), p < 0.001). The study results demonstrate that the proposed DLAD could be utilized as a decision-support system to reduce radiologists’ false negative rate.

1. Introduction

Chest X-ray (CXR) is one of the most common diagnostic imaging tests. It is often used to identify and monitor various chest findings, including pulmonary lesions that may be indicative of, among other pathologies, lung cancer. However, the effectiveness of X-ray imaging in detecting both primary and secondary tumors is not always reliable [1]. In addition, CXR is prone to reading errors and low interobserver and intraobserver agreements due to its limited spatial resolution and variability in human anatomic structures.
A pulmonary lesion can usually be observed as a rounded, well-defined opacity on CXR. The margins may be smooth or convoluted, and it is surrounded by normally aerated lung tissue. A lesion is defined by its size, location, margins, and density. This description plays an important role in the differential diagnosis of the particular patient, as well as the information about the number of lesions present since the background of a solitary pulmonary nodule, pulmonary mass, and multiple lesions may vary. Generally, if the lesion is less than 30 mm in diameter, it is considered a nodule; if more than 30 mm, it is considered a mass [2,3]. Occasionally, nodules as small as 5-6 mm in size can be visible on CXR [4]. Despite this, even larger lesions can remain undetected by a well-experienced radiologist; e.g., in [5], repeated CXR was used as a screening method for the detection of lung cancer. It ws found that 45 out of 50 carcinomas detected by the screening, all of which were more than 10 mm in diameter, could also be seen on the previous CXR images, which were originally misdiagnosed as normal CXRs.
Although most solitary pulmonary nodules are benign findings, when a lesion is discovered, the probability of its malignancy must be assessed using clinical and radiological features [6]. Clinical factors, which may increase the probability of malignity, include the size of the pulmonary lesion, age of the patient, history of the oncological disease, chronic tobacco smoking, asbestos exposure, or presence of other diseases such as chronic obstructive pulmonary disease [7]. A solitary pulmonary nodule less than 7 mm in diameter has a risk of malignant potential less than 1%, an 8–20 mm nodule’s risk is 18%, and nodules of 20–30 mm have a malignancy risk of 50% [8]. Moreover, a tumor of the lung does not always appear as a nodule or mass on a CXR. Tumors causing bronchial obstruction with atelectasis are difficult to distinguish from obstructive atelectasis due to other causes [9,10]. Centrally located tumors often cause mediastinal lymphadenopathy, and bronchial obstruction with atelecatsis, may spread directly into the mediastinum, and often have seeding infection that may mask the lesion [11]. In such cases, a tumor is suspected only on repeat imaging after pneumonia has healed [12].
A solitary pulmonary nodule has a wide differential diagnosis. Causes other than malignant tumors include benign tumors, inflammatory lesions (granuloma, abscess, rheumatoid nodule, inflammatory pseudotumor, round pneumonia), congenital processes (arteriovenous malformation, pulmonary cyst, bronchial atresia with mucous plug), and others (pulmonary infarction, intrapulmonary lymphatic nodule, mucous plug, pulmonary hematoma, pulmonary amyloidosis, confluention of the vessels). The diagnostic process is focused primarily on the differentiation between benign and malignant causes [13]. If the lesion is assessed as benign, the approach depends mainly on the patient’s clinical symptoms. Some pathologies only require to be observed by follow-up CXRs or CTs, and the pathology origin needs to be clarified in detail to correctly assess the potential risk to the patient and to set up the most beneficial treatment plan [14]. Pulmonary masses differ in size from pulmonary nodules, the arbitrary threshold being 30 mm. Primary lung tumors are the most common cause; a solitary lesion larger than 30 mm should be considered malignant unless proven otherwise [15]. However, as with other lesions, the differential diagnosis is very extensive, and the most common benign cause is granuloma (caused by sarcoidosis, infection, vasculitis, or rheumatoid disease).
If a suspicion of lung cancer arises from the CXR, a contrast-enhanced chest CT is used for further assessment of the pathology’s nature and origin, and, in the case of a malignant tumor, for staging (Figure 1). CT shows the detailed location of the tumor, the pattern of its spread, and eventually, satellite foci, mediastinal and hilar lymphadenopathy, and contralateral lung metastases; if the epigastrium is also imaged, metastases to the liver or adrenal glands can be assessed. Recently, the role of CT has also been applied in the screening of selected patient groups: a lung cancer screening program has been running in the Czech Republic since January 2022.
The aim of this study is to demonstrate the effectiveness of the proposed DLAD for detecting pulmonary lesions on CXR images and to compare its performance with that of radiologists with different levels of experience in a simulated clinical setting. For this purpose, we performed a multi-reader, single-site, retrospective study.

2. Related Works

Detection of pulmonary lesions on CXRs using DLAD is an active area of research in both commercial applications (Figure 2) and academic projects [17,18,19]. Due to the known limitations in transparency [20] and the lack of clinical relevance of publicly available datasets [21], we decided to exclude studies that leveraged these data to train or evaluate their DLAD in our analysis.
In [22], a commercial DLAD for detecting pulmonary lesions (Lunit INSIGHT CXR) was developed and tested on a dataset of CXRs. The DLAD performance was compared to that of 18 doctors. The investigated DLAD demonstrated a specificity (Sp) of 0.952 while preserving a high sensitivity (Se) of 0.807, whereas the average Se of the radiologists was 0.704. Ref. [23] evaluated the effectiveness of a commercial DLAD (Riverain ClearRead Xray) in improving the detection of pulmonary nodules on CXRs by radiologists. The DLAD was used to process 300 CXRs and generated bone-suppressed images (BSIs). Five radiologists and three residents evaluated the CXRs with and without the algorithm. The results showed that the DLAD achieved a Se of 0.74 and significantly improved the average performance of the radiologists, as measured by the area under the curve (AUC), from 0.812 to 0.841. The DLAD also detected 127 of the 239 nodules that were missed by the radiologists. In [24], the standalone performance of a commercial DLAD (Arterys Chest AI) was evaluated. The DLAD was tested on CXR data from a single hospital, and its performance in detecting opacities, pleural effusions, pneumothorax, nodules, and fractures was assessed. The study found that the algorithm achieved good Se and Sp in detecting opacities, pleural effusions, and pneumothorax but showed limited performance in detecting nodules and fractures due to the small number of these abnormalities in the dataset (24 nodules and 4 fractures). The AUC for the different pathologies ranged from 0.773 for pulmonary nodules to 0.991 for pneumothorax. Ref. [25] evaluated the effectiveness of commercial DLAD (Samsung Electronics Auto Lung Nodule Detection) in improving the detection of pulmonary lesions on CXRs by radiologists. The DLAD was tested on a dataset of 600 CXRs containing lung cancer and 200 normal CXRs from four medical centers. The performance of the DLAD was compared to that of 12 radiologists. The results showed that the DLAD had a significantly higher Se in detecting pulmonary lesions than the radiologists alone (0.86 versus 0.79) and also had a significantly higher AUC (0.88 versus 0.82). Ref. [26] evaluated the effectiveness of commercial DLAD (Siemens Healthineers AI-Rad Companion) in detecting pulmonary nodules on CXRs. The study included 100 CXR images from two centers in Germany and USA. The images were selected to represent nodules with different levels of detection difficulty. Nine radiologists from Germany and the US reviewed all of the images in two sessions, one with the DLAD-aided interpretation and one without, with a washout period in between. The study found that the mean detection accuracy for the radiologists improved by 6.4% with DLAD-aided interpretation compared to unaided interpretation. Junior radiologists saw a greater improvement in Se for nodule detection with AI-aided interpretation compared to senior radiologists, while senior radiologists experienced a similar improvement in Sp to junior radiologists. Ref. [27] leveraged a commercial DLAD (Qure AI qXR) in detecting malignant nodules on CXR images. The researchers compared the accuracy of DLAD to that of two radiologists in detecting nodules on 894 preselected CXRs. The results showed that DLAD had a high AUC of 0.99 and a Se of 1 at the operating threshold, indicating a high level of accuracy.

3. Materials and Methods

3.1. Software

Deep learning (DL) algorithms are a type of artificial intelligence (AI) that uses multi-layered artificial neural networks (ANN) to analyze and recognize patterns in data. Computer vision algorithms involve training robust convolutional neural networks (CNN) on large numbers of images. The CNNs are able to learn and make predictions based on the visual data and ground truth labels provided and have been shown to be particularly effective in the field of medical imaging [28,29].
The proposed DLAD (Carebot AI CXR v2.00) utilizes various deep learning techniques to identify and localize suspicious lesions in CXR images. It is designed to assist radiologists in interpreting CXRs in posterior–anterior (PA) or anterior–posterior (AP) projections. To maintain the existing user workflow, the DLAD was implemented into the picture-archiving and communication system (PACS) for potentially more efficient clinical deployment. The DLAD leverages DL algorithms to automatically detect abnormalities based on visual patterns for the following abnormalities: atelectasis (ATE), consolidation (CON), cardiomegaly (CMG), pneumothorax (PNO), pulmonary lesion (LES), pleural effusion (EFF), and subcutaneous emphysema (SCE). This study investigates solely the software module for identifying suspicious pulmonary lesions. Standardized descriptions of individual findings were determined through discussion with the radiologists involved in the development of the proposed DLAD. Detailed information can be found in the user manual provided to collaborating hospitals.

3.1.1. Training Data

The development of the proposed DLAD involved 25,374 anonymized X-ray images with established ground truth from sites across Europe, Asia, and North America. The inclusion of CXRs from multiple hospitals was intended to account for the variation in screening quality [30] and local population characteristics [31]. The dataset excluded images of poor quality or incorrect projection, as well as those from patients under 18 years of age. A team of 27 radiologists, with varying levels of experience, annotated the collected CXR images. The final training dataset contained 12,149 images of patients with visually confirmed pathological findings and 13,225 images with either no or insignificant findings.

3.1.2. Model Architecture

To develop the LES model for detecting suspicious lesions on CXR, the YOLOv5 architecture (Figure 3) was used. The main advantage of using an object-detection approach is the ability to accurately detect and localize even smaller pulmonary nodules that conventional convolutional neural networks may struggle with. Ref. [32] introduced YOLO, a novel approach to object detection, in 2016. The basic idea is dividing the input image into a grid of cells and predicting bounding boxes and class probabilities for each cell [33]. The bounding box coordinates are predicted as offsets relative to a set of anchor boxes, which are pre-defined default boxes with different aspect ratios. The YOLO model consists of three main components: a feature extractor, a predictor, and a postprocessor. The feature extractor is responsible for extracting relevant features from the input image, which are then passed through the predictor to produce the bounding box predictions. The postprocessor filters the predicted bounding boxes using non-max suppression to remove overlapping boxes and low-confidence detections.

3.1.3. Communication Protocol

The proposed DLAD takes the form of a web-based application entity (AE) for the analysis of image documentation (DICOM CR, DX). It uses the DICOMweb™ standard to communicate with the connected PACS. The software takes the form of a prediction algorithm with application peripherals (web-based communication tools, DICOM file conversion, storage, and reporting libraries in proprietary DICOM Basic Text Structured Report and DICOM Presentation State formats).

3.2. Data Collection

To collect the data for this retrospective study, we addressed a specialized oncology center. Masaryk Memorial Cancer Institute is a leading cancer research and treatment hospital in the Czech Republic. A total of 9276 CXR images were retrieved from the period between June 2020 and July 2021. Data excluded images of children under 18 years of age, scans with technical issues, and images taken in lateral projection.

3.3. Ground Truth

To determine the ground truth for the presence or absence of a pulmonary lesion on CXR, we proposed that three radiologists with different levels of experience would be assessed for the initial reading. The evaluation included (A) a radiologist with less than 5 years of experience, (B) a board-certified radiologist with more than 5 years of experience, and (C) a head of the radiology department with more than 10 years of experience. At least 2 out of 3 radiologists needed to agree on the presence of a pulmonary lesion in order to establish the ground truth (Table 1). If there was a further disagreement between the assessed radiologists, the CXR was not included in the assessment. Subsequent pixel-level annotation was carried out by a board-certified head of the radiology department of a municipal hospital.

3.4. Assessment

A total of 300 CXR images were randomly selected for the experiment using the random.choice() function, of which 100 (55 ♀, 45 ♂) were from patients with a total of 385 confirmed pulmonary lesions (LES+ Abnormal), 100 (68 ♀, 32 ♂) were from patients with various confirmed findings other than pulmonary lesions (LES− Abnormal), and 100 (79 ♀, 21 ♂) were without any pathological findings (Normal). The prevalence of individual findings can be found in Table 2. Five independent radiologists were invited to evaluate the presence or absence of pulmonary lesions in the selected CXR images without being given any information about the patient’s history or previous or follow-up examinations. This was in order to objectively compare the results with those produced by the DLAD software. The experience levels of the radiologists were as follows: RAD 1 and RAD 2 were junior radiologists with less than 2 years of experience, RAD 3 was a radiologist with more than 2 years but less than 5 years of experience, and RAD 4 and RAD 5 were board-certified radiologists with more than 5 years of experience. Evaluation of CXR was conducted in Carebot Label App software, with each radiologist identifying if the CXR showed patterns for any of the 13 preselected abnormalities (atelectasis, consolidation, cardiomegaly, mediastinal widening, pneumoperitoneum, pneumothorax, pulmonary edema, pulmonary lesion, bone fracture, hilar enlargement, subcutaneous emphysema, and pleural effusion). The study was designed as blinded since the assessed radiologists were not aware of the ratio of lesion-positive or negative CXRs.

3.5. Statistical Analysis

The standard approach to measuring the performance in the detection of pulmonary lesions on CXR images is to use the measures of balanced accuracy (BA), sensitivity (Se), specificity (Sp), and positive (PLR) and negative likelihood ratio (NLR). Se is the proportion of actual positive cases that are correctly identified as positive by the diagnostic test; e.g., it measures the proportion of true positive (TP) and false negative (FN) results. On the other hand, Sp is the proportion of actual negative cases that are correctly identified as negative by the diagnostic test; e.g., it measures the proportion of true negative (TN) and false positive (FP) results. The mutual relations are expressed by PLR = Se/(1−Sp) and NLR = (1−Se)/Sp. The likelihood ratios (LRs) depend only on Se and Sp and are equivalent to the relative risk. Higher PLR and lower NLR are desirable.
A paired design was applied to the data; i.e., each CXR was evaluated by DLAD and all assessed radiologists, and their evaluations were compared to the ground truth. To compare the Se and Sp of the DLAD with that of individual radiologists, we calculated and compared statistical parameters using confidence intervals (CI) and p-values. The null hypothesis test ( H 0 ) was conducted to determine if there were significant differences between the DLAD and the radiologists. The procedure to compare the statistics consisted of (i) solving H 0 to an α error, calculating the Wald χ 2 test (e.g., H 0 : ( S e 1 = S e 2 and S p 1 = S p 2 ) vs. H 1 : ( S e 1 S e 2 and/or S p 1 S p 2 ) ), and (ii) if H 0 was rejected, conducting alternative hypothesis test ( H 1 ) with a multiple comparison method (e.g., McNemar with continuity correction for Se and Sp and the Holm method for LRs) to an α error. Forest plots were used to visualize the differences in Se and Sp between individual radiologists and DLAD. CIs were constructed at a two-tailed 95% confidence level.

4. Results

A total of 300 images with established ground truth were evaluated, of which 100 (55 ♀, 45 ♂) were from patients with one or more confirmed pulmonary lesions (LES+ Abnormal), 100 (68 ♀, 32 ♂) were from patients with various confirmed findings other than pulmonary lesions (LES− Abnormal), and 100 (79 ♀, 21 ♂) were without any pathological findings (Normal). The proposed deep-learning-based automatic detection algorithm (DLAD) correctly identified 91 out of 100 CXRs (Se of 0.910 (0.854–0.966), Figure 4) with pulmonary lesions (LES+ Abnormal) and 165 out of 200 CXRs (Sp of 0.815 (0.761–0.900)) as either without pulmonary lesions (LES− Abnormal) or without any abnormalities (Normal). There were 45 (15%) images that were incorrectly classified as having pulmonary lesions, even though they did not, 26 of which were originally determined as Abnormal LES− and 19 as Normal, indicating that DLAD was slightly more likely to encounter difficulties with comorbid patients than with patients without any abnormalities. In particular, rib summation and more prominent pulmonary vascular markings were among the problematic regions that indicated multiple false positives (FP). The higher false positive rate of the proposed DLAD was expected because the threshold was set to classify even suspect findings as abnormal. Additionally, nine (3%) images with pulmonary lesions were incorrectly classified as not having any lesions (FN). The false-negative CXR images with respected ground truth pixel-level annotations and confusion matrices for the DLAD and individual radiologists can be found in the Appendix A (Figure A1 and Figure A2).
For all assessed radiologists, the proposed DLAD achieved a statistically significantly higher Se (0.910 (0.854–0.966)) than that of radiologists (RAD 1: 0.290 (0.201–0.379), p < 0.001; RAD 2: 0.450 (0.352–0.548), p < 0.001; RAD 3: 0.670 (0.578–0.762), p < 0.001; RAD 4: 0.810 (0.733–0.887), p = 0.025; RAD 5: 0.700 (0.610–0.790), p < 0.001), and would therefore help to identify patients with pulmonary lesions that the radiologists evaluated as being without the suspicious findings (Table 3, Figure 5). However, the DLAD achieved a lower Sp (0.775 (0.717–0.833)) than all assessed radiologists (RAD 1: 1.000 (0.984–1.000), p < 0.001; RAD 2: 0.970 (0.946–1.000), p < 0.001; RAD 3: 0.980 (0.961–1.000), p < 0.001; RAD 4: 0.975 (0.953–0.997), p < 0.001; RAD 5: 0.995 (0.985–1.000), p < 0.001), and the difference was statistically significant. As no images were evaluated as FP by RAD 1, the CI was calculated using the Clopper–Pearson method.
The DLAD PLR (4.044 (3.104–5.269)) was significantly lower (i.e., worse) than that of radiologists (RAD 1: N/A (no CXR was evaluated as FP), RAD 2: 15.000 (6.624–33.966), p = 0.002; RAD 333.500 (12.575–89.245), p < 0.001; RAD 4: 32.400 (13.564–77.389), p < 0.001; RAD 5: 140.000 (19.734–993.174), p < 0.001); the NLR (0.116 (0.062–0.218)) was lower (i.e., better) than that of radiologists (RAD 1: 0.710 (0.626–0.804), p < 0.001; RAD 2: 0.567 (0.474–0.678), p < 0.001; RAD 3: 0.337 (0.254–0.445), p < 0.001; RAD 4: 0.194 (0.130–0.292), p = 0.132; RAD 5: 0.301 (0.223–0.406), p = 0.003); and the difference was statistically significant, with the exception of RAD 4, where no statistical difference was found for this parameter (p > 0.05). (Table 4, Figure 5).

5. Discussion

The present retrospective study investigated the performance of a deep learning-based automatic detection algorithm (DLAD) for the detection of pulmonary lesions on chest X-ray (CXR) data from a specialized oncology center. The proposed DLAD demonstrated higher sensitivity (Se 0.910 (0.854–0.966)) than all five assessed radiologists: RAD 1: 0.290 (0.201–0.379), p < 0.001; RAD 2: 0.450 (0.352–0.548), p < 0.001; RAD 3: 0.670 (0.578–0.762), p < 0.001; RAD 4: 0.810 (0.733–0.887), p = 0.025; RAD 5: 0.700 (0.610–0.790), p < 0.001). The effectiveness was indirectly demonstrated by its ability to alert to the presence of a pulmonary lesion that a doctor might have missed. While the proposed DLAD may produce more false positive results (Sp 0.775 (0.717–0.833) than radiologists (RAD 11.000 (0.984–1.000), p < 0.001, RAD 20.970 (0.946–1.000), p < 0.001, RAD 30.980 (0.961–1.000), p < 0.001, RAD 40.975 (0.953–0.997), p < 0.001, RAD 50.995 (0.985–1.000), p < 0.001), its ability to identify overlooked lesions makes it a useful decision-support tool in clinical practice. For individual radiologists, there were 62, 46, 24, 10, and 21 CXR images with missed suspicious lesions that the assessed radiologists would re-evaluate and/or seek guidance on from experienced colleagues.
Compared with related studies that applied a similar methodology, the proposed DLAD model achieved high sensitivity on a relatively large and well-described dataset. The trade-off between Se and Sp was expected, as the DLAD was designed to serve as a decision-support system. The results of this study are largely unique because they investigated detection performance on retrospectively collected data from a specialized cancer center where the prevalence of patients with suspicious lung lesions may be higher. Comparable results were presented in [22], where the performance of a commercial DLAD was investigated in a multi-centric study. The investigated DLAD demonstrated Sp of 0.952 while preserving high Se of 0.807, whereas the averaged Se of the physicians was 0.704. Ref. [24] evaluated the performance of a commercial DLAD to detect five findings on CXRs including pulmonary nodules, but the test data contained only 24 images with suspicious lesions. The proposed DLAD reached an AUC of 0.773 for pulmonary lesion detection, which, given the similar ground truth approach as in our study, is a considerably worse score. An interesting approach was proposed in [25], which simulated clinical practice with a high prevalence of patients with suspicious lesions (600 CXRs with nodules, 200 normal CXRs). The DLAD was compared to 12 radiologists and achieved Se of 0.86 compared to the radiologists’ 0.79. However, it must be stated that the prevalence of patients with suspicious lesions in clinical practice is significantly lower [35,36]. The comparison with other related studies was impossible due to the different evaluation methodologies: in [23], radiologists evaluated original CXRs and BSI CXRs, [26] did not present Se and Sp values but only an increase in accuracy, and [27] did not compare performance against physicians.

Limitations

While DLADs can provide valuable assistance to radiologists in detecting pulmonary lesions on CXR images, their use also carries potential risks and harm to patients. One of our main concerns with the use of the proposed DLAD in medical diagnosis is the potential for incorrect or misleading results. Since the algorithm is based on statistical models, it can make mistakes or produce false positives and negatives, leading to incorrect diagnoses. This can have consequences for patients, who may be given unnecessary treatments or miss out on vital care [37]. The false-positive rate of the proposed DLAD was higher than that of the comparison radiologists. However, it should be added that differential diagnosis or confusion with normal human-body factors played a role. Furthermore, there is a risk that DLAD may be trained on a dataset that is not representative of the general population, leading to unequal or unfair treatment of certain groups of patients [38]. This can have profound implications for patient care and can result in discrimination and unequal access to medical services. Since this study was conducted retrospectively and the assessed radiologists did not have access to patients’ clinical information, it may not accurately reflect real-world conditions. These factors can be minimized by carrying out a prospective multi-centric study or by retrospectively evaluating large prospectively collected multi-institutional datasets in simulated clinical practice. In addition, the option of supporting ground truth with other diagnostic methods, such as biopsy or longer-term follow-up of the lesion, should be explored. While this study showed promising results achieved with a relatively large sample of patients, further clinical validation process needs to be conducted to determine the applicability of the proposed DLAD.

6. Conclusions

The purpose of this study was to analyze the performance of a deep learning-based automatic detection algorithm (DLAD, Carebot AI CXR v2.00) in identifying suspicious pulmonary lesions on CXR, and to compare its accuracy to that of five individual radiologists. The proposed DLAD demonstrated improved detection performance compared to existing conventional imaging-based diagnostics, as it showed a significantly lower false-negative rate while also providing relatively high specificity.

Author Contributions

Conceptualization, D.K., A.C. and E.J.; methodology, D.K., A.C. and K.K.; software, D.K., A.C., R.H., M.B. and M.P.; validation, D.K., A.C., R.H. and K.K.; data curation, A.C., E.J. and S.S.; writing—original draft preparation, D.K., A.C. and P.P.; writing—review and editing, D.K. and M.A.A.-a.; visualization, D.K. and R.H.; supervision, D.K. and M.A.A.-a.; project administration, D.K.; funding acquisition, D.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Carebot, Ltd.

Institutional Review Board Statement

This study was enabled by contract for the transfer of X-ray images for medical research purposes (Smlouva o předání rentgenových snímků pro účely medicínského výzkumu) between Masaryk Memorial Cancer Institute and Carebot, Ltd., signed on 3 January 2022 in Brno, Czech Republic. The Ethics Committee approved the transfer of anonymized chest X-ray images for the research project.

Informed Consent Statement

Patient consent was waived due to Regulation 2016/679 of the European Parliament and of the Council of 27 April 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and the repealing of Directive 95/46/EC (General Data Protection Regulation); i.e., the CXR images were stripped of all direct or indirect identifiers without the possibility of retrospective patient identification.

Data Availability Statement

Data from this study can be provided by Carebot, Ltd., to independent researchers. Please contact the author for more information, if required.

Conflicts of Interest

In relation to this study, we declare the following conflicts of interest: The study was funded by Carebot, Ltd. D.K., A.C., R.H., M.B., K.K., and M.P. are employees of Carebot, Ltd.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
CIConfidence Interval
CNNConvolutional Neural Network
CXRChest X-ray
DLDeep Learning
DLADDeep Learning–based Automatic Detection Algorithm
SeSensitivity
SpSpecificity
BABalanced Accuracy
TPTrue Positive
FPFalse Positive
TNTrue Negative
FNFalse Negative
LRLikelihood Ratio
PLRPositive Likelihood Ratio
NLRNegative Likelihood Ratio
PVPredictive Value
PPVPositive Predictive Value
NPVNegative Predictive Value

Appendix A

Figure A1. False negative (FN) images incorrectly classified by DLAD software during the retrospective study as true negative (TN) and the respected ground truth pixel-level annotation. Assessed radiologist RAD 1 also incorrectly classified CXR images #1, #2, #4, #5, #6, #7, #8 and #9. Assessed radiologist RAD 2 also incorrectly classified CXR images #1, #2, #3, #4, #5, #6 and #8. Assessed radiologist RAD 3 also incorrectly classified CXR images #1, #2, #3, #4, #5, #6 and #8. Assessed radiologist RAD 4 also incorrectly classified CXR images #1, #2, #4, and #8. Assessed radiologist RAD 5 also incorrectly classified CXR images #2, #4, #5, #6 and #9.
Figure A1. False negative (FN) images incorrectly classified by DLAD software during the retrospective study as true negative (TN) and the respected ground truth pixel-level annotation. Assessed radiologist RAD 1 also incorrectly classified CXR images #1, #2, #4, #5, #6, #7, #8 and #9. Assessed radiologist RAD 2 also incorrectly classified CXR images #1, #2, #3, #4, #5, #6 and #8. Assessed radiologist RAD 3 also incorrectly classified CXR images #1, #2, #3, #4, #5, #6 and #8. Assessed radiologist RAD 4 also incorrectly classified CXR images #1, #2, #4, and #8. Assessed radiologist RAD 5 also incorrectly classified CXR images #2, #4, #5, #6 and #9.
Diagnostics 13 01043 g0a1
Figure A2. Confusion matrix showing the performance of the proposed DLAD and individual radiologists. All of them incorrectly evaluated some assessed LES+ Abnormal CXRs as without suspicious lesions. For individual radiologists, it was 62, 46, 24, 10 and 21 images, which the assessed radiologists would check again and/or consult with a more experienced colleague.
Figure A2. Confusion matrix showing the performance of the proposed DLAD and individual radiologists. All of them incorrectly evaluated some assessed LES+ Abnormal CXRs as without suspicious lesions. For individual radiologists, it was 62, 46, 24, 10 and 21 images, which the assessed radiologists would check again and/or consult with a more experienced colleague.
Diagnostics 13 01043 g0a2

References

  1. Sone, S.; Takashima, S.; Li, F.; Yang, Z.; Honda, T.; Maruyama, Y.; Hasegawa, M.; Yam, A.T.; Kubo, K.; Hanamura, K.; et al. Mass screening for lung cancer with mobile spiral computed tomography scanner. Lancet 1998, 351, 1242–1245. [Google Scholar] [CrossRef]
  2. Hansell, D.M.; Bankier, A.A.; MacMahon, H.; McLoud, T.C.; Müller, N.L.; Remy, J. Fleischner Society: Glossary of terms for thoracic imaging. Radiology 2008, 246, 697–722. [Google Scholar] [CrossRef] [Green Version]
  3. Herring, W. Learning Radiology: Recognizing the Basics; Elsevier Health Sciences: Oxford, UK, 2019. [Google Scholar]
  4. Gamboa, A.; Ethun, C.; Switchenko, J.; Lipscomb, J.; Poultsides, G.; Grignol, V.; Howard, J.; Gamblin, T.; Roggin, K.; Votanopoulos, K.; et al. Lung surveillance strategy for high-grade soft tissue sarcomas: Chest X-ray or CT scan? J. Am. Coll. Surg. 2019, 229, 449–457. [Google Scholar] [CrossRef] [PubMed]
  5. Muhm, J.R.; Miller, W.E.; Fontana, R.S.S.; Erson, D.R.; Uhlenhopp, M.A. Lung cancer detected during a screening program using four-month chest radiographs. Radiology 1983, 148, 609–615. [Google Scholar] [CrossRef]
  6. Albert, R.; Russell, J. Evaluation of the solitary pulmonary nodule. Am. Fam. Physician 2009, 80, 827–831. [Google Scholar]
  7. Goo, J.M.; Park, C.M.; Lee, H.J. Ground-glass nodules on chest CT as imaging biomarkers in the management of lung adenocarcinoma. AJR Am. J. Roentgenol. 2011, 196, 533–543. [Google Scholar] [CrossRef] [PubMed]
  8. Tang, A.W.; Moss, H.A.; Robertson, R.J. The solitary pulmonary nodule. Eur. J. Radiol. 2003, 45, 69–77. [Google Scholar] [CrossRef]
  9. Molina, P.; Hiken, J.; Glazer, H. Imaging evaluation of obstructive atelectasis. J. Thorac. Imaging 1996, 11, 176–186. [Google Scholar] [CrossRef]
  10. Vaaler, A.; Forrester, J.; Lesar, M.; Edison, M.; Johnson, B.; Venzon, D. Obstructive atelectasis in patients with small cell lung cancer: Incidence and response to treatment. Chest 1997, 111, 115–120. [Google Scholar] [CrossRef]
  11. Strollo, D.; Christenson, L.; Jett, J. Primary mediastinal tumors: Part ll. Tumors of the middle and posterior mediastinum. Chest 1997, 112, 1344–1357. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Chastre, J.; Trouillet, J.; Vuagnat, A.; Joly-Guillou, M.; Clavier, H.; Dombret, M.; Gibert, C. Nosocomial pneumonia in patients with acute respiratory distress syndrome. Am. J. Respir. Crit. Care Med. 1998, 157, 1165–1172. [Google Scholar] [CrossRef] [PubMed]
  13. Seemann, M.; Staebler, A.; Beinert, T.; Dienemann, H.; Obst, B.; Matzko, M.; Pistitsch, C.; Reiser, M. Usefulness of morphological characteristics for the differentiation of benign from malignant solitary pulmonary lesions using HRCT. Eur. Radiol. 1999, 9, 409–417. [Google Scholar] [CrossRef]
  14. Janzen, D.; Padley, S.; Adler, B.; Müller, N. Acute pulmonary complications in immunocompromised non-AIDS patients: Comparison of diagnostic accuracy of CT and chest radiography. Clin. Radiol. 1993, 47, 159–165. [Google Scholar] [CrossRef] [PubMed]
  15. Okada, M.; Nishio, W.; Sakamoto, T.; Uchino, K.; Yuki, T.; Nakagawa, A.; Tsubota, N. Effect of tumor size on prognosis in patients with non–small cell lung cancer: The role of segmentectomy as a type of lesser resection. J. Thorac. Cardiovasc. Surg. 2005, 129, 87–93. [Google Scholar] [CrossRef] [Green Version]
  16. Kim, Y.; Cho, Y.; Wu, C.; Park, S.; Jung, K.; Seo, J.; Lee, H.; Hwang, H.; Lee, S.; Kim, N. Short-term reproducibility of pulmonary nodule and mass detection in chest radiographs: Comparison among radiologists and four different computer-aided detections with convolutional neural net. Sci. Rep. 2019, 9, 18738. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Ausawalaithong, W.; Thirach, A.; Marukatat, S.; Wilaiprasitporn, T. Automatic lung cancer prediction from chest X-ray images using the deep learning approach. In Proceedings of the 2018 11th Biomedical Engineering International Conference (BMEiCON), Chaing Mai, Thailand, 21–24 November 2018; pp. 1–5. [Google Scholar]
  18. Li, X.; Shen, L.; Xie, X.; Huang, S.; Xie, Z.; Hong, X.; Yu, J. Multi-resolution convolutional networks for chest X-ray radiograph based lung nodule detection. Artif. Intell. Med. 2020, 103, 101744. [Google Scholar] [CrossRef]
  19. Nasrullah, N.; Sang, J.; Alam, M.; Mateen, M.; Cai, B.; Hu, H. Automated lung nodule detection and classification using deep learning combined with multiple strategies. Sensors 2019, 19, 3722. [Google Scholar] [CrossRef] [Green Version]
  20. Hryniewska, W.; Bombiński, P.; Szatkowski, P.; Tomaszewska, P.; Przelaskowski, A.; Biecek, P. Checklist for responsible deep learning modeling of medical images based on COVID-19 detection studies. Pattern Recognit. 2021, 118, 108035. [Google Scholar] [CrossRef]
  21. Oakden-Rayner, L. Exploring large-scale public medical image datasets. Acad. Radiol. 2020, 27, 106–112. [Google Scholar] [CrossRef] [Green Version]
  22. Nam, J.; Park, S.; Hwang, E.; Lee, J.; Jin, K.; Lim, K.; Vu, T.; Sohn, J.; Hwang, S.; Goo, J.; et al. Development and validation of deep learning–based automatic detection algorithm for malignant pulmonary nodules on chest radiographs. Radiology 2019, 290, 218–228. [Google Scholar] [CrossRef] [Green Version]
  23. Schalekamp, S.; Ginneken, B.; Koedam, E.; Snoeren, M.; Tiehuis, A.; Wittenberg, R.; Karssemeijer, N.; Schaefer-Prokop, C. Computer-aided detection improves detection of pulmonary nodules in chest radiographs beyond the support by bone-suppressed images. Radiology 2014, 272, 252–261. [Google Scholar] [CrossRef] [Green Version]
  24. Arterys Retrospective Study X-ray Chest AI Whitepaper. Arterys. 2020. Available online: https://www.arterys.com/retrospective-study-x-ray-chest-ai-wp (accessed on 15 January 2023).
  25. Sim, Y.; Chung, M.; Kotter, E.; Yune, S.; Kim, M.; Do, S.; Han, K.; Kim, H.; Yang, S.; Lee, D.; et al. Deep convolutional neural network–based software improves radiologist detection of malignant lung nodules on chest radiographs. Radiology 2020, 294, 199–209. [Google Scholar] [CrossRef] [PubMed]
  26. Homayounieh, F.; Digumarthy, S.; Ebrahimian, S.; Rueckel, J.; Hoppe, B.; Sabel, B.; Conjeti, S.; Ridder, K.; Sistermanns, M.; Wang, L.; et al. An Artificial Intelligence–Based Chest X-ray Model on Human Nodule Detection Accuracy from a Multicenter Study. JAMA Netw. Open 2021, 4, e2141096. [Google Scholar] [CrossRef] [PubMed]
  27. Mahboub, B.; Tadepalli, M.; Raj, T.; Santhanakrishnan, R.; Hachim, M.; Bastaki, U.; Hamoudi, R.; Haider, E.; Alabousi, A. Identifying malignant nodules on chest X-rays: A validation study of radiologist versus artificial intelligence diagnostic accuracy. Adv. Biomed. Health Sci. 2022, 1, 137. [Google Scholar] [CrossRef]
  28. Monkam, P.; Qi, S.; Ma, H.; Gao, W.; Yao, Y.; Qian, W. Detection and classification of pulmonary nodules using convolutional neural networks: A survey. IEEE Access. 2019, 7, 78075–78091. [Google Scholar] [CrossRef]
  29. Yamashita, R.; Nishio, M.; Do, R.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Into Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  30. Whaley, J.; Pressman, B.; Wilson, J.; Bravo, L.; Sehnert, W.; Foos, D. Investigation of the variability in the assessment of digital chest X-ray image quality. J. Digit. Imaging 2013, 26, 217–226. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Gavelli, G.; Giampalma, E. Sensitivity and specificity of chest X-ray screening for lung cancer. Cancer 2000, 89, 2453–2456. [Google Scholar] [CrossRef]
  32. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  33. Liu, C.; Tao, Y.; Liang, J.; Li, K.; Chen, Y. Object detection based on YOLO network. In Proceedings of the 2018 IEEE 4th Information Technology Furthermore, Mechatronics Engineering Conference (ITOEC), Chongqing, China, 14–16 December 2018; pp. 799–803. [Google Scholar]
  34. Katsamenis, I.; Karolou, E.; Davradou, A.; Protopapadakis, E.; Doulamis, A.; Doulamis, N.; Kalogeras, D. TraCon: A novel dataset for real-time traffic cones detection using deep learning. In Novel & Intelligent Digital Systems, Proceedings of the 2nd International Conference (NiDS 2022), Athens, Greece, 29–30 September 2022; Sprigner: Berlin/Heidelberg, Germany, 2022; pp. 382–391. [Google Scholar]
  35. Janssen-Heijnen, M.; Schipper, R.; Razenberg, P.; Crommelin, M.; Coebergh, J. Prevalence of co-morbidity in lung cancer patients and its relationship with treatment: A population-based study. Lung Cancer 1998, 21, 105–113. [Google Scholar] [CrossRef]
  36. Low, S.; Eng, P.; Keng, G.; Ng, D. Positron emission tomography with CT in the evaluation of non-small cell lung cancer in populations with a high prevalence of tuberculosis. Respirology 2006, 11, 84–89. [Google Scholar] [CrossRef]
  37. Margerie-Mellon, C.; Chassagnon, G. Artificial intelligence: A critical review of applications for lung nodule and lung cancer. Diagn. Interv. Imaging 2022, 104, 11–17. [Google Scholar] [CrossRef] [PubMed]
  38. Bi, W.; Hosny, A.; Schabath, M.; Giger, M.; Birkbak, N.; Mehrtash, A.; Allison, T.; Arnaout, O.; Abbosh, C.; Dunn, I.; et al. Artificial intelligence in cancer imaging: Clinical challenges and applications. CA Cancer J. Clin. 2019, 69, 127–157. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Initial and follow-up CXRs and CT images with a pulmonary lesion. (a) Initial CXR of a 65-year-old male patient with metastatic renal cell carcinoma in the left upper lobe and (b) CT examination of the patient (a). (c) Follow-up CXR of (a). (d) Initial CXR of an 81-year-old male patient with metastatic adenocarcinoma in the right middle lobe and (e) CT examination corresponding to (d). (f) Follow-up CXR of (d). The yellow arrows indicate the localization of suspected pulmonary lesions [16].
Figure 1. Initial and follow-up CXRs and CT images with a pulmonary lesion. (a) Initial CXR of a 65-year-old male patient with metastatic renal cell carcinoma in the left upper lobe and (b) CT examination of the patient (a). (c) Follow-up CXR of (a). (d) Initial CXR of an 81-year-old male patient with metastatic adenocarcinoma in the right middle lobe and (e) CT examination corresponding to (d). (f) Follow-up CXR of (d). The yellow arrows indicate the localization of suspected pulmonary lesions [16].
Diagnostics 13 01043 g001
Figure 2. (a) The proposed DLAD (Carebot AI CXR v2.00, implemented in CloudPACS by OR-CZ) and other commercial solutions: (b) Qure AI qXR, (c) Lunit INSIGHT CXR, and (d) Arterys Chest AI.
Figure 2. (a) The proposed DLAD (Carebot AI CXR v2.00, implemented in CloudPACS by OR-CZ) and other commercial solutions: (b) Qure AI qXR, (c) Lunit INSIGHT CXR, and (d) Arterys Chest AI.
Diagnostics 13 01043 g002
Figure 3. Overview of the YOLOv5 model architecture [34].
Figure 3. Overview of the YOLOv5 model architecture [34].
Diagnostics 13 01043 g003
Figure 4. Examples of the respected ground truth pixel-level annotations and correct DLAD predictions (TP). The proposed DLAD correctly identified 91 out of 100 CXRs (Se of 0.910 (0.854–0.966)) pulmonary lesions (LES+ Abnormal).
Figure 4. Examples of the respected ground truth pixel-level annotations and correct DLAD predictions (TP). The proposed DLAD correctly identified 91 out of 100 CXRs (Se of 0.910 (0.854–0.966)) pulmonary lesions (LES+ Abnormal).
Diagnostics 13 01043 g004
Figure 5. Forest plots showing the mean sensitivity (Se), specificity (Sp), positive (PLR), and negative likelihood ratio (NLR) and corresponding 95% confidence interval estimates for DLAD and individual radiologists. For all assessed radiologists, the DLAD achieved a statistically significantly higher Se than that of radiologists, indicating that it would be useful in identifying patients with pulmonary lesions that were not identified by the radiologists.
Figure 5. Forest plots showing the mean sensitivity (Se), specificity (Sp), positive (PLR), and negative likelihood ratio (NLR) and corresponding 95% confidence interval estimates for DLAD and individual radiologists. For all assessed radiologists, the DLAD achieved a statistically significantly higher Se than that of radiologists, indicating that it would be useful in identifying patients with pulmonary lesions that were not identified by the radiologists.
Diagnostics 13 01043 g005
Table 1. Methodology of determining the ground truth by assessment of three radiologists with different levels of experience.
Table 1. Methodology of determining the ground truth by assessment of three radiologists with different levels of experience.
ClassInclusion Criteria
LES+ AbnormalA consensus of 2/3 is required to confirm the presence of one or more pulmonary lesions.
In addition, CXR may contain other pathological abnormalities.
A consensus of 2/3 is required to confirm these.
LES− AbnormalA consensus of 3/3 is required to confirm the absence of any pulmonary lesions.
In addition, CXR contained other pathological abnormalities.
A consensus of 2/3 is required to confirm these.
NormalA consensus of 3/3 is required to confirm the CXR did not show any pathological abnormalities.
Table 2. Patient characteristics indicating the demographic data and the prevalence of individual findings.
Table 2. Patient characteristics indicating the demographic data and the prevalence of individual findings.
DataLES+ AbnormalLES− AbnormalNormal
Total100100100
Patient’s Sex
   Female (♀)556879
   Male (♂)453221
   F:M Ratio1.22:12.13:13.76:1
Prevalence
   LES only4000
   With other findings601000
Findings
   Atelectasis560
   Consolidation21210
   Cardiomegaly5380
   Fracture1100
   Mediastinal widening100
   Pneumoperitoneum010
   Pneumothorax000
   Pulmonary edema090
   Pleural effusion21340
   Pulmonary lesion10000
   Hilar enlargement220
   Subcutaneous emphysema010
Table 3. Pooled results of DLAD and individual radiologists sensitivity (Se) and specificity (Sp) with corresponding 95% confidence interval estimates. The experience levels of the radiologists were as follows: RAD 1 and RAD 2 were junior radiologists with less than 2 years of experience, RAD 3 was a radiologist with more than 2 years but less than 5 years of experience, and RAD 4 and RAD 5 were board-certified radiologists with more than 5 years of experience.
Table 3. Pooled results of DLAD and individual radiologists sensitivity (Se) and specificity (Sp) with corresponding 95% confidence interval estimates. The experience levels of the radiologists were as follows: RAD 1 and RAD 2 were junior radiologists with less than 2 years of experience, RAD 3 was a radiologist with more than 2 years but less than 5 years of experience, and RAD 4 and RAD 5 were board-certified radiologists with more than 5 years of experience.
BASe (95% CI)Sp (95% CI)Se p-ValueSp p-Value
DLAD0.8430.910 (0.854–0.966)0.775 (0.717–0.833)
RAD 10.6450.290 (0.201–0.379)1.000 (0.984–1.000)<0.001<0.001
RAD 20.7100.450 (0.352–0.548)0.970 (0.946–0.994)<0.001<0.001
RAD 30.8250.670 (0.578–0.762)0.980 (0.961–1.000)<0.001<0.001
RAD 40.8930.810 (0.733–0.887)0.975 (0.953–0.997)0.025<0.001
RAD 50.8480.700 (0.610–0.790)0.995 (0.985–1.000)<0.001<0.001
Table 4. Pooled results of DLAD and individual radiologists likelihood ratio (LR) with corresponding 95% confidence interval estimates. The experience levels of the radiologists were as follows: RAD 1 and RAD 2 were junior radiologists with less than 2 years of experience, RAD 3 was a radiologist with more than 2 years but less than 5 years of experience, and RAD 4 and RAD 5 were board-certified radiologists with more than 5 years of experience.
Table 4. Pooled results of DLAD and individual radiologists likelihood ratio (LR) with corresponding 95% confidence interval estimates. The experience levels of the radiologists were as follows: RAD 1 and RAD 2 were junior radiologists with less than 2 years of experience, RAD 3 was a radiologist with more than 2 years but less than 5 years of experience, and RAD 4 and RAD 5 were board-certified radiologists with more than 5 years of experience.
PLRNLRPLR p-ValueNLR p-Value
DLAD4.044 (3.104–5.269)0.116 (0.062–0.218)
RAD 1N/A0.710 (0.626–0.804)N/A<0.001
RAD 215.000 (6.624–33.966)0.567 (0.474–0.678)0.002<0.001
RAD 333.500 (12.575–89.245)0.337 (0.254–0.445)<0.001<0.001
RAD 432.400 (13.564–77.389)0.194 (0.130–0.292)<0.0010.132
RAD 5140.000 (19.734–993.174)0.301 (0.223–0.406)<0.0010.003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kvak, D.; Chromcová, A.; Hrubý, R.; Janů, E.; Biroš, M.; Pajdaković, M.; Kvaková, K.; Al-antari, M.A.; Polášková, P.; Strukov, S. Leveraging Deep Learning Decision-Support System in Specialized Oncology Center: A Multi-Reader Retrospective Study on Detection of Pulmonary Lesions in Chest X-ray Images. Diagnostics 2023, 13, 1043. https://doi.org/10.3390/diagnostics13061043

AMA Style

Kvak D, Chromcová A, Hrubý R, Janů E, Biroš M, Pajdaković M, Kvaková K, Al-antari MA, Polášková P, Strukov S. Leveraging Deep Learning Decision-Support System in Specialized Oncology Center: A Multi-Reader Retrospective Study on Detection of Pulmonary Lesions in Chest X-ray Images. Diagnostics. 2023; 13(6):1043. https://doi.org/10.3390/diagnostics13061043

Chicago/Turabian Style

Kvak, Daniel, Anna Chromcová, Robert Hrubý, Eva Janů, Marek Biroš, Marija Pajdaković, Karolína Kvaková, Mugahed A. Al-antari, Pavlína Polášková, and Sergei Strukov. 2023. "Leveraging Deep Learning Decision-Support System in Specialized Oncology Center: A Multi-Reader Retrospective Study on Detection of Pulmonary Lesions in Chest X-ray Images" Diagnostics 13, no. 6: 1043. https://doi.org/10.3390/diagnostics13061043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop