Next Article in Journal
Improvement of the Diagnosis of Isoattenuating Pancreatic Carcinomas by Defining Their Characteristics on Contrast Enhanced Computed Tomography and Endosonography with Fine-Needle Aspiration (EUS-FNA)
Next Article in Special Issue
Dilated Semantic Segmentation for Breast Ultrasonic Lesion Detection Using Parallel Feature Fusion
Previous Article in Journal
The Reconstruction of Magnetic Particle Imaging: Current Approaches Based on the System Matrix
Previous Article in Special Issue
A Novel Deep Learning Method for Recognition and Classification of Brain Tumors from MRI Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NanoChest-Net: A Simple Convolutional Network for Radiological Studies Classification

by
Juan Eduardo Luján-García
1,
Yenny Villuendas-Rey
2,*,
Itzamá López-Yáñez
2,*,
Oscar Camacho-Nieto
2,* and
Cornelio Yáñez-Márquez
1,*
1
Centro de Investigación en Computación, Instituto Politécnico Nacional, Mexico City 07700, Mexico
2
Centro de Innovación y Desarrollo Tecnológico en Cómputo, Instituto Politécnico Nacional, Mexico City 07738, Mexico
*
Authors to whom correspondence should be addressed.
Diagnostics 2021, 11(5), 775; https://doi.org/10.3390/diagnostics11050775
Submission received: 27 March 2021 / Revised: 13 April 2021 / Accepted: 22 April 2021 / Published: 26 April 2021
(This article belongs to the Special Issue Deep Learning for Computer-Aided Diagnosis in Biomedical Imaging)

Abstract

:
The new coronavirus disease (COVID-19), pneumonia, tuberculosis, and breast cancer have one thing in common: these diseases can be diagnosed using radiological studies such as X-rays images. With radiological studies and technology, computer-aided diagnosis (CAD) results in a very useful technique to analyze and detect abnormalities using the images generated by X-ray machines. Some deep-learning techniques such as a convolutional neural network (CNN) can help physicians to obtain an effective pre-diagnosis. However, popular CNNs are enormous models and need a huge amount of data to obtain good results. In this paper, we introduce NanoChest-net, which is a small but effective CNN model that can be used to classify among different diseases using images from radiological studies. NanoChest-net proves to be effective in classifying among different diseases such as tuberculosis, pneumonia, and COVID-19. In two of the five datasets used in the experiments, NanoChest-net obtained the best results, while on the remaining datasets our model proved to be as good as baseline models from the state of the art such as the ResNet50, Xception, and DenseNet121. In addition, NanoChest-net is useful to classify radiological studies on the same level as state-of-the-art algorithms with the advantage that it does not require a large number of operations.

1. Introduction

The new coronavirus disease (COVID-19) has achieved historical records. Until 8 March 2021, the World Health Organization (WHO) had registered more than 116 million confirmed cases and over 2.5 million deaths [1]. COVID-19 is an infectious disease caused by SARS-CoV2 virus that affects severely the lungs of people infected, and the virus is easily propagated in the air and by contact. COVID-19 can cause complications and lead to development of pneumonia and other symptoms that can be confused with other diseases [1].
In addition, pneumonia is also an infectious disease that affects the lungs and can be caused by bacteria such as Streptococcus pneumoniae and Haemophilus influenzae, and viruses apart from the one that provokes COVID-19. It has been a major disease and cause of death for children and senior people around the world. According to the WHO, pneumonia causes 15% of all deaths of children under 5 years old [2]. Moreover, pneumonia killed 808,694 children in 2017.
On the other hand, tuberculosis disease, caused by Mycobacterium tuberculosis is also an infectious disease that causes antimicrobial resistance and death of tissue on different parts of the body, affecting principally the lungs. According to the WHO, tuberculosis is the top infectious killer around the world and causes around 1.5 million deaths every year. In 2019 alone, an estimated 10 million people fell ill with tuberculosis [3].
At the same time, according to the World Cancer Research Fund (WCRF) breast cancer is the most common cancer in women, and until 2018 it was the second most common cancer overall in the world, with over 2 million new cases in the same year [4]. Consequently, COVID-19, pneumonia, tuberculosis, and breast cancer have one thing in common: they can be diagnosed using radiological studies such as X-ray images. Lung infections can be detected by taking a chest X-ray of the patient; breast cancer can be detected by taking an X-ray image of the breast of a woman, called mammography [5]. In the case of COVID-19, radiological features include peripheral damage in one or both lungs, and a crazy paving pattern is commonly found in chest X-ray images from infected patients [1]. On the other hand, pneumonia not caused by COVID-19 causes pus and fluid in the lung, identifying radiopaque segments in the X-ray images without a specific pattern [6]. Tuberculosis presents radiolucent segments on the X-ray images due to solid necrosis in the center of the affected area (tubercles) [7]. As breast cancer concerns, mammography images show nodules and microcalcifications (radiopaque segments) near the mammary glands [8].
X-ray images usually are not as accurate as computed tomography (CT) and magnetic resonance imaging (MRI), but developing countries do not always have specialized equipment available to acquire CT and MRI. Therefore, X-ray images become a crucial tool to help physicians to diagnose diseases. With radiological studies and technology, computer-aided diagnosis (CAD) results in a very useful technique to analyze and detect abnormalities using the images generated by X-ray machines [9].
Prior to deep learning (DL) frameworks, medical image classification was based on traditional feature extraction and classification algorithms. For example, Livieris et al. [10] presented a framework that consisted in an ensemble of semi-supervised learning (SSL) algorithms to identify and classify lung abnormalities. In this work, authors used a tuberculosis dataset and different configurations of SSL such as self-training, co-training, and tri-training, obtaining accuracies under 74% for tuberculosis classification. Another popular work was presented by Minaee et al. [11] in which they manually extracted features from MRI to track the damage on patients with brain injuries. They used feature selection and linear regression as classification algorithms. Nowadays, CAD is mostly aided by computer vision (CV)-specialized algorithms from DL such as a convolutional neural network (CNN) [12]. CNNs are the most popular type of DL algorithms and the most used for medical image diagnosis. We can find several works from segmentation of lesions [13,14,15] to classification of different diseases [16,17,18,19].
As a major relevance for this work, we can find works that use radiological studies to classify among different diseases. Rajan et al. [20] presented a few-shot learning approach to classify among 14 chest diseases using X-ray images, and they proposed a solution to train a CNN with few data and solve the problem of acquiring a vast amount of medical imaging data.
As COVID-19 concerns, Sharma et al. [21] presented a CNN called CORONA-19 NET in which they used transfer learning to classify with a MobileNetV2 between normal and sick patients using a small dataset of 20 images. In addition, Zebin and Rezvy [22] used multiple pretrained CNNs as feature extractors to classify among patients. Moreover, Yu et al. [23] presented a framework that used four pretrained CNNs as baseline to classify among patients using CT scans, obtaining accuracies superior to 94%. Luján-García et al. [24] used an Xception CNN to classify among COVID-19 and pneumonia patients using a pretrained model on ImageNet. They showed that the Xception network was the fastest among several baselines. More recently, Yazdani et al. [25] presented a CNN with an attention mechanism to classify COVID-19 patients, obtaining a sensitivity of 90% using CT scans. Finally, Gupta et al. [26] presented a framework called InstaCovNet-19, which consists in using five pretrained baselines and stacking them to classify patients using X-ray images, and this obtained excellent results of almost 100% accuracy compared with other researches.
Pneumonia classification also plays an important role in radiological classification studies. Zhang et al. [27] presented a confidence-aware framework that uses a CNN as a feature extractor, a confidence module, and a prediction module achieving a sensitivity of 71.70%. Rahman et al. [28] presented a comparison between several baseline models to classify images of children infected with pneumonia, achieving up to 99% of sensitivity, and Luján-García et al. [29] used the same dataset but added a preprocessing technique and used a different pretrained baseline.
Recently, Rajpurkar et al. [30] presented a DL assistance tool to classify tuberculosis from patients with human immunodeficiency virus (HIV) using a CNN and a linear classifier to predict six clinical findings. On the other hand, Pasa et al. [31] presented a new small CNN to classify X-ray images from two small datasets, and they achieved good results despite the fact that no pretrained models were used. Moreover, using the same dataset as Pasa et al., Khatibi et al. [32] used an ensemble of CNNs to achieve classification accuracies up to 99.2%.
On the other hand, breast cancer has been improved using DL techniques. Shen et al. [33] used pretrained baselines on a large mammography dataset to classify among malign and benign mass and calcification, obtaining a sensitivity of 96%. Moreover, Agarwal [34] used a pretrained CNN to detect masses in mammography images, which achieved a better result, compared to Shen et al., with a sensitivity of 98% on the same dataset. Finally, Wu et al. [35] presented a custom ResNet-based CNN to classify over 1 million images from multiple views of patients with benign and malign masses, achieving an area under the curve score of 0.895.
Nonetheless, popular CNNs are enormous models and need a large amount of data for the purpose of being trained properly to get good results. Therefore, we aim to preset a small but effective CNN model that can be used to classify among different diseases using images from radiological studies.

2. Materials and Methods

In this section, datasets used for this research are described. In addition, we briefly introduce some of the CNN baseline models used for comparison purposes. Finally, metrics used for evaluating the algorithms are detailed.

2.1. Datasets

2.1.1. Tuberculosis Dataset

The tuberculosis dataset is a collection of two sets of chest X-ray images from two different hospitals presented by the National Institute of Health of the United States [36]. The tuberculosis dataset is divided in two sets: the Montgomery County set and the Shenzhen set.
The Montgomery County set contains 138 frontal chest X-ray images, in which 80 of them are normal cases and 58 are from tuberculosis patients. Similarly, the Shenzhen set contains 662 frontal chest X-ray images, of which 326 are normal cases and 336 are tuberculosis patients.

2.1.2. Pneumonia Children Dataset

The Pneumonia children dataset was published by Kermany et al. [37]. The dataset contains 5856 chest X-ray images of healthy and sick children up to five years old. All images are given as a training set, with 5232 images and an official test set of 624 images. From the 5232 images, 3883 are from patients infected with pneumonia. The remaining 1349 images are from healthy children. On the other hand, the test set is divided as follows: 390 images are from pneumonia-infected children, and 234 images are from healthy children.

2.1.3. COVID-19 Dataset

Presented by Cohen et al. [38], The COVID-19 Image Data Collection was one of the first open available datasets that contained chest X-rays from patients infected with COVID-19. We initially used the dataset from November 2020, which contained 930 images from different diseases such as pneumonia, severe acute respiratory syndrome (SARS), and middle east respiratory syndrome (MERS), among others. At this time, only 478 images were from patients infected with COVID-19. These images were used to generate two different sets of images to perform experiments, explained in the next section.

2.1.4. RSNA Pneumonia Challenge Dataset

The RSNA Pneumonia Challenge (RSNA-PC) is the only competition (from Kaggle.com) to classify and provide bounding boxes for damaged areas of the lung caused by pneumonia. The dataset contains 26,684 unique chest X-ray images of both normal (29%) and not normal/opacities (71%) for the training set, and 3000 images for the test set.

2.1.5. BCDR Dataset

The Breast Cancer Digital Repository (BCDR), by Moura and Guevara [39], offers multiple datasets for both digital and scanned mammography in which principal classes are malign and benign tumors. For this work, we have used only the two datasets of digital mammography.
The BCDR-D01 contains full-field digital mammography and is composed of 79 biopsy-proven lesions of 64 women, rendering 143 segmentations for 80 unique images of patients with benign tumors, and 57 patients with malign tumors.
The BCDR-D02 contains full-field digital mammography and is composed of 230 biopsy-proven lesions of 162 women, rendering 455 segmentations for 359 unique images of patients with benign tumors, and 48 patients with malign tumors.

2.2. CNN Models from the State of the Art

Back in 2015, since the formal introduction of Deep Learning [40], the research community has dedicated a lot of attention and effort on developing DL algorithms for different purposes, such as image recognition, CV, CAD systems, and natural language processing, among others. Due to the capacity of extracting features withing the algorithm itself from different kind of signals (including images), CNNs have achieved magnificent results. Nowadays, a huge number of CNN models exist and are used for distinct purposes. As a result, we can find custom models [16,19,31] and the ones that use key baselines for classification of different diseases [17,22,24,29,41,42,43].
For this research, we have compared the proposed method with the most popular CNNs used for computer vision tasks such as the ResNet50 [44], the Xception network [45], and the DenseNet121 [46].

2.3. Metrics

In a binary classification problem, we can measure the performance according to the examples correctly classified that belong to each class as true positives (tp) and true negatives (tn), and we take into account the mistakes or errors when classifying instances such as the false positives (fp) and false negatives (fn). Normally, tp, tn, fp, and fn are shown in tabular form as a confusion matrix (Figure 1).
From the confusion matrix, we can compute a variety of metrics. Accuracy is commonly used when we have a classification task among two or more different classes. Moreover, we can also compute other metrics such as precision, sensitivity, specificity, F1-Score, and the area under the ROC curve (AUC) [47]. Following, we can find the definition of the metrics used in this work (Equations (1)–(5)).
A c c u r a c y = t p + t n t p + f n + f p + t n
P r e c i s i o n = t p t p + f p
S e n s i t i v i t y = t p t p + f n
S p e c i f i c i t y = t n f p + t n
F 1 - S c o r e = 2   t p 2   t p + f p + t n = 2   p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l .
In general, accuracy not always represents an unbiased performance measurement due to different imbalances within the instances of a dataset. Therefore, precision, sensitivity, specificity, F1-Score, and AUC are always helpful to measure the performance of a model. For this work, an AUC using thresholds was computed.

3. Proposal

In this section, a detailed description of the proposed custom CNN model is given. Moreover, final datasets and their partitions are explained. On the other hand, preprocessing and data augmentation techniques are described. Finally, hyperparameters used to train the models are mentioned.

3.1. DL Model

Inspired by the Separable Convolutions from the Xception network, we have designed the NanoChest-net to classify between images from radiological studies, such as X-ray images. The complete block diagram of our CNN model is shown in Figure 2. A complete specification of each layer of the CNN is described in Table 1.
We have used the depth multiplier of Separable Convolution layers to increment the number of output channels on each layer. In addition, we have used a dilation rate of 2 to increment the size of the spatial perception field on each layer. As a result, Table 1 shows us the total number of layers of our proposal, which is 28. If we count the set of a convolutional layer, the batch normalization layer, and its activation as a complete layer (as commonly used in the literature), then our proposal is composed of a very small number of 14 layers. Moreover, if we only focus on weighted layers, then our proposal is as small as 10 layers in depth. In comparison, baseline models such as VGG-16, which comes next in layer size, contain 19 weighted layers and have an average of 136.4 million parameters [48]. Therefore, our proposal has the advantage of halving the depth according to weighted layers, and it has 40 times fewer parameters with only 3.4 million.
As consequence, the aforementioned reasons are motive to call this small model as NanoChest-net due to the minimal number of layers on the CNN model and its application to radiological studies, primarily of the chest.

3.2. Datasets Splitting and Validation Method

3.2.1. Splitting and Final Datasets

We have maintained the original examples for Shenzhen, Montgomery, Pneumonia children, BCDR-D01, and BCDR-D02. Nonetheless, we have generated two new subsets using the COVID-19 dataset and the RSNA-PC dataset. We took the 478 COVID-19 images from the COVID-19 dataset and 478 images of healthy patients of the RSNA-PC to generate the COVID-NORMAL dataset. Moreover, we took the same 478 images from COVID-19 dataset, but now 478 images of pneumonia-sick patients from the RSNA-PC to generate the COVID-PNEUMONIA dataset. Table 2 shows the final datasets used for this research.

3.2.2. Validation Method

Hold-out validation was performed in order to obtain the training, development (Dev), and test set for each dataset. Hold-out validation consists of randomly dividing the original number of images on the training, Dev, and test set. Figure 3 shows the behavior of the hold-out validation method.
A hold-out 70-10-20 was used over each dataset, except for the Pneumonia children, in which an official test set was established by the authors. Therefore, partitions for each dataset are as follows (Table 3).

3.3. Preprocessing and Data Augmentation

All images were normalized before feeding the CNN models. In addition, all datasets were resized to 500 × 500 pixels to avoid resizing each image from its original size to the input size of each CNN model on each step of the training. Moreover, we fed the models with the original input size. Then, the input size of each model is as follows (Table 4).

Tuberculosis Montgomery County Dataset

For the Montgomery County dataset, first we cropped the central region of the images with the intention of deleting black bars from the original images (Figure 4). We followed the algorithm by Pasa et al. [31]. Then, we applied preprocessing and data augmentation techniques.
On the other hand, data augmentation techniques were applied to each dataset aiming to obtain a better generalization of the models. For tuberculosis datasets, pneumonia dataset, and COVID-19 dataset we applied horizontal flip, magnification in a range of 0.90 to 1.2, random width and height shift with a factor of 0.20, random rotation of 20 degrees, and brightness changes in a factor range of 0.80 to 1.05. In the case of the BCDR datasets, we changed the random rotation to 30 degrees and added vertical flip.

3.4. Hyperparameter Tuning

We conducted the same experiments using the state-of-the-art CNNs and our proposed method. Equivalent hyperparameters were used through all models, except for the input size. We used the original input size for each model, as mentioned in Section 4.3. We trained all models using a logistic layer of two units with Sigmoid activation to get the probability of having each of two classes per dataset. Binary cross-entropy was used as a cost function (computed as in [29]) and Adam [49] as optimization algorithm with parameters β 1 = 0.9 ,   β 2 = 0.999 (recommended values from original paper). In addition, we performed several experiments with different optimizers to see the impact in the training of our proposal as seen in Table 5 (best results are highlighted in bold).
From Table 5, we can see that the stochastic gradient descent (SGD) [50] did not obtain good results in any dataset. On the other hand, Adam obtained the best scores 25 times, and RMSProp (introduced by Hinton in 2012) obtained the best scores 17 times. Adam was selected due to the fact it is a combination of SGD (using momentum) and RMSProp (squared gradients). As a result, we also performed experiments to see the impact of changing the learning rate for training using Adam optimizer as seen on Table 6 (best results are highlighted in bold).
From previous results (Table 6) we can see that using a learning rate of 0.001 provided the best scores 16 times. Nonetheless, using a learning of 0.0005, the best scores were obtained 23 times. A learning rate of 0.0005 was selected because it showed better performance than a larger one. Moreover, a small learning rate on the order of 10–4 benefits all models when training on small datasets.
As a result, a learning rate of 0.0005 was selected to perform all experiments with all datasets. In the same way, the Adam algorithm, which is a combination of SGD and RMSProp, was selected as an optimization algorithm. Moreover, the number of epochs was selected according to the size of each dataset and our technological capabilities. In addition, the batch size was selected considering the size of each dataset and its partitions. Learning rate, epochs, and batch size configurations are shown in Table 7.
We applied weights to BCDR-D01 and BCDR-D02 datasets to combat class imbalance. We used 0.8636 and 1.1875 for benign and malign on BCDR-D01, and 0.568 and 4.1765 for benign and malign on BCDR-D02.

4. Results

In this section, the experimental framework is described. In addition, performance and comparison between models are presented. Furthermore, statistical analysis is presented, considering metrics and time measurements.

4.1. Experimental Framework

Experiments for this research were conducted on a PC with AMD Ryzen 3700x processor; 16 GB of RAM; 512 SSD + 2 TB of storage; GPU Nvidia RTX 2070 Super with 8 GB GDDR5; Python 3.7.9 was used as programming language; TensorFlow 2.1.0 with Keras as high-level DL framework; sci-kit learn 0.23.2 [51] as machine learning (ML) library; and OpenCV 3.4.2 [52] as main image processing library. Moreover, we set a fixed seed for TensorFlow, Python random generator, and NumPy library to get the repeatability of the experiments.
In addition, we want to clarify that all baseline CNN models have been randomly initialized with the intention of making a fair comparison with our proposal. Neither transfer learning nor finetuning were performed in these tests. Apart from state-of-the-art works [13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,32,33,34,35] pretrained in several medical image datasets and ImageNet, our proposal received no extra training and only took the training partition of each presented dataset. Our source code is available on https://github.com/zotrick/NanoChest-net.

4.2. Test Sets Results

We trained all baseline models and the proposed NanoChest-net using the same hyperparameters (apart from the input size) on all the datasets specified in Section 2 and computed the performance metrics. The results over the respective test set for each dataset can be found in Table 8 (best results are highlighted in bold).
As seen in Table 8. The proposed model behaved similar to the baseline ones. A detailed discussion will be given in the next section.

4.3. Training Time Results

Apart from evaluating the metrics, we also measured the time taken for each model at training time. We measured the total time taken by each model, the average time per epoch, the time of each model to process a single example, and the time taken in achieving the best result through all the training epochs. Therefore, results can be found in Table 9 (best results are highlighted in bold).
Despite the fact our proposal did not seem be faster among the CNN models, we will perform further analyses in the next section.

4.4. Size of the Models

With the final structure of each CNN model, and after training them, we measured the total number of parameters of each one. In addition, as the best-performance model for all CNNs was saved, we measured the size of storage of each model. Results can be found in Table 10 (smaller number of parameters and size are highlighted in bold).

4.5. Statistical Analysis

We performed the Friedman test [53] over computed metrics for each model on each dataset (Table 11). The Friedman test tells with a 95% confidence if a significant statistical difference exists between five or more instances ranked in order of significance. If p < 0.05 is found, then significant differences exist. We arranged the results of the Friedman test in Table 11 (best results are highlighted in bold).
From Table 11, we can observe that p-values were greater than 0.05 (confidence level of 95%). Therefore, the null hypothesis for equality on the performance of compared algorithms was not rejected.
As for time per example, statistical analysis showed that the Friedman test obtained a p-value of 0.001213, being the ResNet50 the best of the ranking. As a result, the null hypothesis was rejected. However, after applying the Holm test [54], we obtained the results as shown in Table 12.
From Table 12, we can observe that the Holm test rejected the hypothesis with unadjusted p-values smaller than 0.001213. Therefore, neither DenseNet121 nor NanoChest-net was rejected. On the contrary, only the Xception network was rejected, showing significant differences (inferior performance) compared with the other algorithms. Consequently, we will perform further analyses in the next section.

5. Discussion

In this section, advantages of the proposal are highlighted. An evaluation of the classification results on the different datasets is performed, as well as the time analysis of the training.
From the classification results (Table 8), our proposal obtained good results through all the datasets. On the Montgomery County dataset, our model obtained the best results on accuracy, sensitivity, F1-Score, and AUC, with scores of 0.931, 1.000, 0.929, and 0.928, respectively. For the Shenzhen dataset, our model obtained the second-best scores on accuracy (0.828), sensitivity (0.897), F1-Score (0.841), and AUC (0.928) only behind the Xception network. On the Pneumonia children dataset, our method achieved the best results among all CNN models, with an accuracy of 0.931, precision of 0.904, sensitivity of 0.995, specificity of 0.825, F1-Score of 0.947, and an AUC score of 0.992. Again, for the COVID-NORMAL dataset, our proposal achieved the best scores of all models with an accuracy of 0.933, precision score of 0.912, sensitivity score of 0.959, specificity score of 0.907, F1-Score of 0.935, and an AUC score of 0.970. On the other hand, for the COVID-PNEUMONIA dataset, the NanoChest-net obtained the best results for precision (0.860), specificity (0.878), and AUC (0.919), and the remaining metrics had good results, only behind the Xception network. Nevertheless, for the BCDR-D01 dataset, we fell behind to third place behind DenseNet121 and Xception networks, except for sensitivity (best result). We obtained the following scores: accuracy of 0.621, precision of 0.500, sensitivity of 0.818, specificity of 0.500, F1-Score of 0.621, and AUC of 0.702. Finally, for the BCDR-D02 dataset we obtained the best results for sensitivity (0.556) and F1-Score (0.278). For the remaining metrics, we obtained second place only behind the DenseNet121 with an accuracy of 0.687, precision of 0.185, specificity of 0.703, and an AUC of 0.664. Finally, from the statistical analysis for metrics (Table 9), the Friedman test did not show evidence to reject the hypothesis. Therefore, there were no statistically significant differences among the models. However, the scores from the test placed our proposal at the top of the ranking on each metric, showing a superior behavior compared to the state-of-the-art baseline models. In addition, if we used a confidence level of 90%, then there will exist differences on F1-Score and AUC in favor of NanoChest-net due to its first position in the ranking.
On the other hand, time measurement results (Table 9) showed that in the Montgomery County and BCDR-D01 our proposal obtained the lowest values for total training time, epoch average time, and time per example. For Shenzhen, Pneumonia children, COVID-NORMAL, COVID-PNEUMONIA, and BCDR-D02, our method was always in third place, behind the ResNet50 and DenseNet121 networks. Nonetheless, from the Friedman test and Holm results (Table 11 and Table 12) we can observe that there were no significant differences among ResNet50, DenseNet121, and our method. The Xception network was the only worse model considering the time taken per example.
At the same time, our proposal took an important step further on a crucial aspect. Apart from showing decent classification and time results, our method was significantly smaller in parameter count and storage size. From Table 10, we can observe that our method had less than half the parameters and size compared to DenseNet121, and 6 to 7 times fewer parameters and smaller size when compared to Xception and ResNet50 models. Consequently, our model could be used in computers, embedded devices, and mobile devices with limited storage, memory, and computation capabilities.

6. Conclusions

In this paper, we have introduced a new, full custom, and small convolutional neural network model called NanoChest-net. Our proposed model is used to classify medical images from radiological studies such as X-rays from the chest and mammography from the breasts of women. As a result, our model proves to be effective in classifying among different diseases like tuberculosis, pneumonia, and COVID-19. Moreover, NanoChest-net obtained the best results on both the Pneumonia children and COVID-NORMAL datasets. On the remaining datasets, our model proved to be as good as baseline models from the state of the art such as the ResNet50, Xception, and DenseNet121, finding no statistically relevant differences among models, neither in performance nor training time. On the contrary, we can find an abrupt difference on the number of parameters and storage size of our model, being two to seven times smaller compared with baseline models. In short, the NanoChest-net model is useful to classify radiological studies on the same level as state-of-the-art algorithms and without computing large numbers of operations and occupying more than 40 MB of storage, making our proposal suitable for embedded and mobile devices.
As future work we plan to further study the correlation of radiological features between pneumonia caused by COVID-19 and other viruses or bacteria. In addition, we plan to train the NanoChest-net on ImageNet and perform a comparison with state-of-the-art frameworks for medical image classification.

Author Contributions

Conceptualization, J.E.L.-G., Y.V.-R. and C.Y.-M.; validation, O.C.-N. and I.L.-Y.; formal analysis, J.E.L.-G., Y.V.-R. and C.Y.-M.; investigation, I.L.-Y. and O.C.-N.; writing—original draft preparation, J.E.L.-G.; writing—review and editing, O.C.-N., Y.V.-R. and C.Y.-M.; visualization, J.E.L.-G.; supervision, I.L.-Y., Y.V.-R. and C.Y.-M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Tuberculosis datasets: available from National Institute of Health at https://lhncbc.nlm.nih.gov/LHC-publications/pubs/TuberculosisChestXrayImageDataSets.html (accessed on 25 October 2020); Pneumonia children dataset: available at https://data.mendeley.com/datasets/rscbjbr9sj/3 (accessed on 5 October 2020); COVID-19 dataset: available on GitHub repository at https://github.com/ieee8023/covid-chestxray-dataset (accessed on 1 November 2020). RSNA-PC dataset: available at https://www.kaggle.com/c/rsna-pneumonia-detection-challenge/ (accessed on 5 October 2020); BCDR datasets can be accessed by filling a form at https://bcdr.eu/ (accessed on 27 November 2020).

Acknowledgments

The authors gratefully acknowledge the Instituto Politécnico Nacional (Secretaría Académica, Comisión de Operación y Fomento de Actividades Académicas, Secretaría de Investigación y Posgrado, Centro de Investigación en Computación, and Centro de Innovación y Desarrollo Tecnológico en Cómputo), the Consejo Nacional de Ciencia y Tecnología (CONACyT), and Sistema Nacional de Investigadores for their economic support to develop this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. Coronavirus Disease (COVID-19) Pandemic. 2020. Available online: https://www.who.int/emergencies/diseases/novel-coronavirus-2019 (accessed on 24 April 2021).
  2. World Health Organization. Pneumonia. 2019. Available online: https://www.who.int/news-room/fact-sheets/detail/pneumonia (accessed on 12 March 2021).
  3. World Health Organization. Tuberculosis. 2020. Available online: https://www.who.int/westernpacific/health-topics/tuberculosis (accessed on 12 March 2021).
  4. World Cancer Research Fund. Breast Cancer Statistics. 2018. Available online: https://www.wcrf.org/dietandcancer/cancer-trends/breast-cancer-statistics (accessed on 12 March 2021).
  5. Suetens, P. Fundamentals of Medical Imaging, 2nd ed.; Cambridge University Press: New York, NY, USA, 2009. [Google Scholar]
  6. Sutton, D. Textbook of Radiology and Imaging, 7th ed.; Chirchill Livingstone: London, UK, 2003. [Google Scholar]
  7. Goodman, L.R. Felson’s Principles of Chest Roentgenology: A Programmed Text, 3rd ed.; Saunders: Philadelphia, PA, USA, 2007. [Google Scholar]
  8. Fauci, K.; Longo, H.; Loscalzo, J. Harrison’s Principles of Internal Medicine, 19th ed.; McGraw-Hill Education: New York, NY, USA, 2015. [Google Scholar]
  9. Doi, K. Computer-Aided Diagnosis in Medical Imaging: Historical Review, Current Status and Future Poten-tial. Computerized medical imaging and graphics. Off. J. Comput. Med. Imaging Soc. 2007, 31, 198–211. [Google Scholar] [CrossRef] [Green Version]
  10. Livieris, I.E.; Kanavos, A.; Tampakas, V.; Pintelas, P. An Ensemble SSL Algorithm for Efficient Chest X-Ray Image Classification. J. Imaging 2018, 4, 95. [Google Scholar] [CrossRef] [Green Version]
  11. Minaee, S.; Wang, Y.; Lui, Y.W. Prediction of Longterm Outcome of Neuropsychological Tests of MTBI Patients Using Imaging Features. In Proceedings of the 2013 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), Brooklyn, NY, USA, 7 December 2013. [Google Scholar]
  12. Chan, H.; Hadjiiski, L.M.; Samala, R.K. Computer-aided diagnosis in the era of deep learning. Med. Phys. 2020, 47, e218–e227. [Google Scholar] [CrossRef] [PubMed]
  13. Pathan, S.; Kumar, P.; Pai, R.M.; Bhandary, S.V. Automated segmentation and classifcation of retinal features for glaucoma diagnosis. Biomed. Signal Process. Control 2021, 63, 102244. [Google Scholar] [CrossRef]
  14. Oulefki, A.; Agaian, S.; Trongtirakul, T.; Laouar, A.K. Automatic COVID-19 lung infected region segmentation and measurement using CT-scans images. Pattern Recognit. 2021, 114, 107747. [Google Scholar] [CrossRef] [PubMed]
  15. Qiu, Y.; Liu, Y.; Li, S.; Xu, J. MiniSeg: An Extremely Minimum Network for Efficient COVID-19 Segmentation. arXiv 2021, arXiv:2004.09750. [Google Scholar]
  16. Gautam, A.; Raman, B. Towards effective classification of brain hemorrhagic and ischemic stroke using CNN. Biomed. Signal Process. Control 2021, 63, 102178. [Google Scholar] [CrossRef]
  17. Mbarki, W.; Bouchouicha, M.; Frizzi, S.; Tshibasu, F.; Ben Farhat, L.; Sayadi, M. Lumbar spine discs classification based on deep convolutional neural networks using axial view MRI. Interdiscip. Neurosurg. 2020, 22, 100837. [Google Scholar] [CrossRef]
  18. Martínez-Más, J.; Bueno-Crespo, A.; Martínez-España, R.; Remezal-Solano, M.; Ortiz-González, A.; Ortiz-Reina, S.; Martínez-Cendán, J.P. Classifying Papanicolaou cervical smears through a cell merger approach by deep learning technique. Expert Syst. Appl. 2020, 160, 113707. [Google Scholar] [CrossRef]
  19. Zhou, H.; Wang, K.; Tian, J. Online Transfer Learning for Differential Diagnosis of Benign and Malignant Thyroid Nodules with Ultrasound Images. IEEE Trans. Biomed. Eng. 2020, 67, 2773–2780. [Google Scholar] [CrossRef]
  20. Rajan, D.; Thiagarajan, J.J.; Karargyris, A.; Kashyap, S. Self-Training with Improved Regularization for Few-Shot Chest X-Ray Classification. arXiv 2020, arXiv:2005.02231. [Google Scholar]
  21. Sharma, S.; Ghose, S.; Datta, S.; Malathy, C.; Gayathri, M.; Prabhakaran, M. CORONA-19 NET: Transfer Learning Approach for Automatic Classification of Coronavirus Infections in Chest Radiographs. In Advances in Intelligent Systems and Computing; Springer International Publishing: New York, NY, USA, 2021; Volume 1200 AISC, pp. 526–534. [Google Scholar]
  22. Zebin, T.; Rezvy, S. COVID-19 detection and disease progression visualization: Deep learning on chest X-rays for classification and coarse localization. Appl. Intell. 2021, 51, 1010–1021. [Google Scholar] [CrossRef]
  23. Yu, Z.; Li, X.; Sun, H.; Wang, J.; Zhao, T.; Chen, H.; Ma, Y.; Zhu, S.; Xie, Z. Rapid identification of COVID-19 severity in CT scans through classification of deep features. Biomed. Eng. Online 2020, 19, 1–13. [Google Scholar] [CrossRef] [PubMed]
  24. Luján-García, J.E.; Moreno-Ibarra, M.A.; Villuendas-Rey, Y.; Yáñez-Márquez, C. Fast COVID-19 and Pneumonia Classification Using Chest X-ray Images. Mathematics 2020, 8, 1423. [Google Scholar] [CrossRef]
  25. Yazdani, S.; Minaee, S.; Kafieh, R.; Saeedizadeh, N.; Sonka, M. COVID CT-Net: Predicting Covid-19 from Chest CT Images Using Attentional Convolutional Network. arXiv 2020, arXiv:2009.05096. [Google Scholar]
  26. Gupta, A.; Anjum; Gupta, S.; Katarya, R. InstaCovNet-19: A deep learning classification model for the detection of COVID-19 patients using Chest X-ray. Appl. Soft Comput. 2020, 99, 106859. [Google Scholar] [CrossRef]
  27. Zhang, J.; Xie, Y.; Liao, Z.; Pang, G.; Verjans, J.; Li, W.; Sun, Z.; He, J.; Li, Y.; Shen, C.; et al. Viral Pneumonia Screening on Chest X-Ray Images Using Confidence-Aware Anomaly Detection. arXiv 2020, arXiv:2003.12338. [Google Scholar]
  28. Rahman, T.; Chowdhury, M.E.H.; Khandakar, A.; Islam, K.R.; Mahbub, Z.B.; Kadir, M.A.; Kashem, S. Transfer Learning with Deep Convolutional Neural Network (CNN) for Pneumonia Detection Using Chest X-ray. Appl. Sci. 2020, 10, 3233. [Google Scholar] [CrossRef]
  29. Luján-García, J.E.; Yáñez-Márquez, C.; Villuendas-Rey, Y.; Camacho-Nieto, O. A Transfer Learning Method for Pneumonia Classification and Visualization. Appl. Sci. 2020, 10, 2908. [Google Scholar] [CrossRef] [Green Version]
  30. Rajpurkar, P.; O’Connell, C.; Schechter, A.; Asnani, N.; Li, J.; Kiani, A.; Ball, R.L.; Mendelson, M.; Maartens, G.; Van Hoving, D.J.; et al. CheXaid: Deep learning assistance for physician diagnosis of tuberculosis using chest x-rays in patients with HIV. npj Digit. Med. 2020, 3, 1–8. [Google Scholar] [CrossRef]
  31. Pasa, F.; Golkov, V.; Pfeiffer, F.; Cremers, D. Efficient Deep Network Architectures for Fast Chest X-Ray Tuberculosis Screening and Visualization. Sci. Rep. 2019, 9, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Khatibi, T.; Shahsavari, A.; Farahani, A. Proposing a novel multi-instance learning model for tuberculosis recognition from chest X-ray images based on CNNs, complex networks and stacked ensemble. Phys. Eng. Sci. Med. 2021, 44, 291–311. [Google Scholar] [CrossRef]
  33. Shen, L.; Margolies, L.R.; Rothstein, J.H.; Fluder, E.; McBride, R.; Sieh, W. Deep Learning to Improve Breast Cancer Detection on Screening Mammography. Sci. Rep. 2019, 9, 1–12. [Google Scholar] [CrossRef] [PubMed]
  34. Agarwal, R.; Diaz, O.; Lladó, X.; Yap, M.H.; Martí, R. Automatic mass detection in mammograms using deep convolutional neural networks. J. Med. Imaging 2019, 6, 31409. [Google Scholar] [CrossRef]
  35. Wu, N.; Phang, J.; Park, J.; Shen, Y.; Huang, Z.; Zorin, M.; Jastrzebski, S.; Fevry, T.; Katsnelson, J.; Kim, E.; et al. Deep Neural Networks Improve Radiologists’ Performance in Breast Cancer Screening. IEEE Trans. Med. Imaging 2020, 39, 1184–1194. [Google Scholar] [CrossRef] [Green Version]
  36. Jaeger, S.; Candemir, S.; Antani, S.; Wáng, Y.-X.J.; Lu, P.-X.; Thoma, G. Two public chest X-ray datasets for computer-aided screening of pulmonary diseases. Quant. Imaging Med. Surg. 2014, 4, 475–477. [Google Scholar]
  37. Kermany, D.; Zhang, K.; Goldbaum, M. Labeled Optical Coherence Tomography (OCT) and Chest X-Ray Images for Classification. Mendeley Data 2018. [Google Scholar] [CrossRef]
  38. Cohen, J.P.; Morrison, P.; Dao, L. COVID-19 Image Data Collection. arXiv 2020, arXiv:2003.11597. [Google Scholar]
  39. Moura, D.C.; Guevara López, M.A. An Evaluation of Image Descriptors Combined with Clinical Data for Breast Cancer Diagnosis. Int. J. Comput. Assist. Radiol. Surg. 2013, 8, 561–574. [Google Scholar] [CrossRef]
  40. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  41. Kamil, M.Y. A deep learning framework to detect Covid-19 disease via chest X-ray and CT scan images. Int. J. Electr. Comput. Eng. IJECE 2021, 11, 844–850. [Google Scholar] [CrossRef]
  42. Chouhan, V.; Singh, S.K.; Khamparia, A.; Gupta, D.; Tiwari, P.; Moreira, C.; Damaševičius, R.; De Albuquerque, V.H.C. A Novel Transfer Learning Based Approach for Pneumonia Detection in Chest X-ray Images. Appl. Sci. 2020, 10, 559. [Google Scholar] [CrossRef] [Green Version]
  43. Liang, G.; Zheng, L. A transfer learning method with deep residual network for pediatric pneumonia diagnosis. Comput. Methods Progr. Biomed. 2020, 187, 104964. [Google Scholar] [CrossRef] [PubMed]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  45. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 125–1258. [Google Scholar]
  46. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  47. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
  48. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, Banff, AL, Canada, 14–16 April 2014. [Google Scholar]
  49. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference Learn, Representations, San Diego, CA, USA, 5–8 May 2015. [Google Scholar]
  50. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the Importance of Initialization and Momentum in Deep Learning. In Proceedings of the 30th International Conference on International Conference on Machine Learning, Atlanta, GA, USA, 16 June 2013; Volume 28, pp. III-1139–III-1147. [Google Scholar]
  51. Pedregosa, F.; Michel, V.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Vanderplas, J.; Cournapeau, D.; Pedregosa, F.; Varoquaux, G.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  52. Bradski, G. The Open CV Library. Dr. Dobb’s J. Softw. Tools 2000, 120, 122–125. [Google Scholar]
  53. Friedman, M. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance. J. Am. Stat. Assoc. 1937, 32, 675. [Google Scholar] [CrossRef]
  54. Holm, S. A Simple Sequentially Rejective Multiple Test Procedure. Scand. J. Stat. 1979, 6, 65–70. [Google Scholar]
Figure 1. Confusion matrix for binary classification.
Figure 1. Confusion matrix for binary classification.
Diagnostics 11 00775 g001
Figure 2. Block diagram of NanoChest-net.
Figure 2. Block diagram of NanoChest-net.
Diagnostics 11 00775 g002
Figure 3. Hold-out validation method.
Figure 3. Hold-out validation method.
Diagnostics 11 00775 g003
Figure 4. Center region cropping applied to the Montgomery County dataset. (a) Shows the original image with black bands; (b) shows the preprocessed image.
Figure 4. Center region cropping applied to the Montgomery County dataset. (a) Shows the original image with black bands; (b) shows the preprocessed image.
Diagnostics 11 00775 g004
Table 1. Layer specification of NanoChest-net.
Table 1. Layer specification of NanoChest-net.
Layer TypeSpecifications
InputSize = (250, 250, 3)
ConvolutionNumber of filters = 64
kernel size = (3, 3)
dilatation rate = 2
padding = valid
ReluNonlinearity relu
ConvolutionNumber of filters = 64
kernel size = (3, 3)
dilatation rate = 2
padding = valid
ReluNonlinearity relu
Max PoolingPool size = (3, 3)
Separable
Convolution
Number of filters = 128
kernel size = (3, 3)
dilatation rate = 2
depth multiplier = 3
padding = valid
Batch
Normalization
Normalization
ReluNonlinearity relu
Separable
Convolution
Number of filters = 256
kernel size = (3, 3)
dilatation rate = 2
depth multiplier = 3
padding = valid
Batch
Normalization
Normalization
ReluNonlinearity relu
Max PoolingPool size = (3, 3)
Separable
Convolution
Number of filters = 256
kernel size = (3, 3)
dilatation rate = 2
depth multiplier = 3
padding = valid
Batch
Normalization
Normalization
ReluNonlinearity relu
Separable
Convolution
Number of filters = 512
kernel size = (3, 3)
dilatation rate = 2
depth multiplier = 3
padding = valid
Batch
Normalization
Normalization
ReluNonlinearity relu
Separable
Convolution
Number of filters = 1024
kernel size = (3, 3)
padding = same
Batch
Normalization
Normalization
ReluNonlinearity relu
Separable
Convolution
Number of filters = 2048
kernel size = (3, 3)
padding = same
Batch
Normalization
Normalization
ReluNonlinearity relu
Global Average
Pooling
Global Pooling
DropoutKeeping rate = 0.25
Logistic-OutputUnits = 2
activation = Softmax
Table 2. Number of images of each class from each dataset.
Table 2. Number of images of each class from each dataset.
DatasetClassesImages per ClassOfficial Test Set
Montgomery County{NORMAL, TUBERCULOSIS}80, 58-
Shenzhen{NORMAL, TUBERCULOSIS}326, 336-
Pneumonia children{NORMAL, PNEUMONIA}1349, 3883234, 390 [37]
COVID-NORMAL{COVID, NORMAL}478, 478-
COVID-PNEUMONIA{COVID, PNEUMONIA}478, 478-
BCDR-D01{BENIGN, MALIGN}80, 57-
BCDR-D02{BENIGN, MALIGN}359, 48-
Table 3. Partitions for each dataset.
Table 3. Partitions for each dataset.
DatasetPartitionClass 1Class 2
Montgomery CountyTraining set5640
Dev set85
Test set1613
ShenzhenTraining set228235
Dev set3233
Test set6668
Pneumonia childrenTraining set12143494
Dev set135389
Test set234390
COVID-NORMALTraining set334334
Dev set4747
Test set9797
COVID-PNEUMONIATraining set334334
Dev set4747
Test set9797
BCDR-D01Training set5640
Dev set86
Test set1611
BCDR-D02Training set25134
Dev set365
Test set729
Table 4. Input size for each CNN model.
Table 4. Input size for each CNN model.
ModelInput Size
ResNet50224 × 224 × 3
Xception299 × 299 × 3
DenseNet121224 × 224 × 3
NanoChest-net250 × 250 × 3
Table 5. Results using different optimizers for NanoChest-net.
Table 5. Results using different optimizers for NanoChest-net.
DatasetOptimizerAccuracyPrecisionSensitivitySpecificityF1AUC
Montgomery CountySGD0.5520.0000.0001.0000.0000.587
RMSProp0.8620.7651.0000.7500.8670.981
Adam0.9310.8671.0000.8750.9290.928
ShenzhenSGD0.7390.7390.7500.7270.7450.861
RMSProp0.8810.9060.8530.9090.8790.932
Adam0.8280.7920.8970.7580.8410.928
Pneumonia childrenSGD0.8940.8600.9920.7310.9210.984
RMSProp0.9200.8861.0000.7860.9400.994
Adam0.9310.9040.9950.8250.9470.992
COVID-NORMALSGD0.7320.6960.8250.6390.7550.844
RMSProp0.8710.8600.8870.8560.8730.930
Adam0.9330.9120.9590.9070.9350.970
COVID-PNEUMONIASGD0.6940.6790.7350.6530.7060.787
RMSProp0.7860.7800.7960.7760.7880.881
Adam0.8160.8600.7550.8780.8040.919
BCDR-D01SGD0.4830.2500.1820.6670.2110.379
RMSProp0.7240.6360.6360.7780.6360.768
Adam0.6210.5000.8180.5000.6210.702
BCDR-D02SGD0.6390.1610.5560.6490.2500.679
RMSProp0.6140.1890.7780.5950.3040.707
Adam0.6870.1850.5560.7030.2780.664
Table 6. Results using Adam optimizer and different learning rates for NanoChest-net.
Table 6. Results using Adam optimizer and different learning rates for NanoChest-net.
DatasetLearning RateAccuracyPrecisionSensitivitySpecificityF1AUC
Montgomery County0.0010.7930.6841.0000.6250.8131.000
0.00050.9310.8671.0000.8750.9290.928
Shenzhen0.0010.8580.8660.8530.8640.8590.937
0.00050.8280.7920.8970.7580.8410.928
Pneumonia children0.0010.9310.9060.9920.8290.9470.992
0.00050.9310.9040.9950.8250.9470.992
COVID-NORMAL0.0010.8610.8300.9070.8140.8670.927
0.00050.9330.9120.9590.9070.9350.970
COVID-PNEUMONIA0.0010.8470.8860.7960.8980.8390.869
0.00050.8160.8600.7550.8780.8040.919
BCDR-D010.0010.6900.5830.6360.7220.6090.657
0.00050.6210.5000.8180.5000.6210.702
BCDR-D020.0010.4580.1090.5560.4460.1820.545
0.00050.6870.1850.5560.7030.2780.664
Table 7. Hyperparameters for each dataset.
Table 7. Hyperparameters for each dataset.
DatasetLearning RateEpochsBatch Size
Montgomery County0.00052004
Shenzhen0.00052008
Pneumonia children0.000510016
COVID-NORMAL0.000520016
COVID-PNEUMONIA0.000520016
BCDR-D010.00052004
BCDR-D020.00052008
Table 8. Metrics of CNN models over all datasets.
Table 8. Metrics of CNN models over all datasets.
DatasetModelAccuracyPrecisionSensitivitySpecificityF1AUC
Montgomery CountyResNet500.8621.0000.6921.0000.8180.885
Xception0.6900.6110.8460.5630.7100.851
DenseNet1210.7930.8180.6920.8750.7500.755
NanoChest-net0.9310.8671.0000.8750.9290.928
ShenzhenResNet500.8130.8640.7500.8790.8030.871
Xception0.8510.8000.9410.7580.8650.937
DenseNet1210.7760.7880.7650.7880.7760.883
NanoChest-net0.8280.7920.8970.7580.8410.928
Pneumonia childrenResNet500.9210.8920.9950.7990.9410.990
Xception0.9170.8860.9950.7860.9370.992
DenseNet1210.9210.8940.9920.8030.9400.989
NanoChest-net0.9310.9040.9950.8250.9470.992
COVID-NORMALResNet500.8450.8020.9180.7730.8560.953
Xception0.8870.8710.9070.8660.8890.960
DenseNet1210.8660.8900.8350.8970.8620.924
NanoChest-net0.9330.9120.9590.9070.9350.970
COVID-PNEUMONIAResNet500.7960.7960.7960.7960.7960.843
Xception0.8370.8240.8570.8160.8400.872
DenseNet1210.7760.7550.8160.7350.7840.857
NanoChest-net0.8160.8600.7550.8780.8040.919
BCDR-D01ResNet500.5860.4740.8180.4440.6000.662
Xception0.6550.5710.3640.8330.4440.732
DenseNet1210.7590.6670.7270.7780.6960.854
NanoChest-net0.6210.5000.8180.5000.6210.702
BCDR-D02ResNet500.5900.1430.5560.5950.2270.659
Xception0.6270.1560.5560.6350.2440.565
DenseNet1210.7350.1900.4440.7700.2670.673
NanoChest-net0.6870.1850.5560.7030.2780.664
Table 9. Time measurements of CNN models over all datasets.
Table 9. Time measurements of CNN models over all datasets.
DatasetModelTotal Training Time (s)Epoch Avg Time (s)Time per Example (s)Convergence Time (s)
Montgomery CountyResNet50251.85981.25930.0131166.2275
Xception490.96862.45480.0256198.8423
DenseNet121268.74261.34370.0140143.7773
NanoChest-net227.58041.13790.0119216.2014
ShenzhenResNet50955.67774.77840.0105793.2125
Xception2112.323910.56160.02321193.4630
DenseNet121997.16724.98580.0109623.2295
NanoChest-net1071.14025.35570.0117599.8385
Pneumonia childrenResNet504649.312346.49310.00994416.8467
Xception9404.984194.04980.02007241.8378
DenseNet1214898.910248.98910.01044115.0845
NanoChest-net5474.782454.74780.01163941.8433
COVID-NORMALResNet501317.23706.58620.01001172.3409
Xception2691.657113.45830.0205753.6640
DenseNet1211384.64046.92320.0106851.5538
NanoChest-net1518.70237.59350.01161260.5229
COVID-PNEUMONIAResNet501387.44206.93720.01061200.1374
Xception2796.158513.98080.0213503.3085
DenseNet1211423.22667.11610.01081095.8845
NanoChest-net1581.97847.90990.0121450.8638
BCDR-D01ResNet50245.41581.22710.0133158.2932
Xception472.35122.36180.0257340.0929
DenseNet121271.22061.35610.0147269.8645
NanoChest-net225.83491.12920.0123195.3472
BCDR-D02ResNet50574.56282.87280.0103255.6805
Xception1239.43536.19720.02211171.2663
DenseNet121594.56772.97280.0106335.9307
NanoChest-net630.95673.15480.0113498.4558
Table 10. Size comparison of the CNN models.
Table 10. Size comparison of the CNN models.
ModelTotal ParametersSize (MB)
ResNet5023,591,810270
Xception20,865,578239
DenseNet1217,039,55481.8
NanoChest-net3,393,98638.9
Table 11. Friedman test over each metric for all models.
Table 11. Friedman test over each metric for all models.
ModelAccuracyPrecisionSensitivitySpecificityF1-ScoreAUC
Friedman Test
0.1836720.4188180.1707540.4402270.0858010.087436
Ranking
NanoChest-net1.711.85711.928621.42861.6429
Xception2.422.85712.14292.92862.71432.2143
DenseNet1212.642.42863.35712.21432.85712.8571
ResNet503.212.85712.57142.857133.2857
Table 12. Adjusted p-values for time per example obtained through the post hoc method (Friedman).
Table 12. Adjusted p-values for time per example obtained through the post hoc method (Friedman).
iAlgorithmUnadjusted p
1Xception0.000084
2NanoChest-net0.09769
3DenseNet1210.147299
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Luján-García, J.E.; Villuendas-Rey, Y.; López-Yáñez, I.; Camacho-Nieto, O.; Yáñez-Márquez, C. NanoChest-Net: A Simple Convolutional Network for Radiological Studies Classification. Diagnostics 2021, 11, 775. https://doi.org/10.3390/diagnostics11050775

AMA Style

Luján-García JE, Villuendas-Rey Y, López-Yáñez I, Camacho-Nieto O, Yáñez-Márquez C. NanoChest-Net: A Simple Convolutional Network for Radiological Studies Classification. Diagnostics. 2021; 11(5):775. https://doi.org/10.3390/diagnostics11050775

Chicago/Turabian Style

Luján-García, Juan Eduardo, Yenny Villuendas-Rey, Itzamá López-Yáñez, Oscar Camacho-Nieto, and Cornelio Yáñez-Márquez. 2021. "NanoChest-Net: A Simple Convolutional Network for Radiological Studies Classification" Diagnostics 11, no. 5: 775. https://doi.org/10.3390/diagnostics11050775

APA Style

Luján-García, J. E., Villuendas-Rey, Y., López-Yáñez, I., Camacho-Nieto, O., & Yáñez-Márquez, C. (2021). NanoChest-Net: A Simple Convolutional Network for Radiological Studies Classification. Diagnostics, 11(5), 775. https://doi.org/10.3390/diagnostics11050775

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop