Next Article in Journal
Biomechanical Rupture Risk Assessment in Management of Patients with Abdominal Aortic Aneurysm in COVID-19 Pandemic
Next Article in Special Issue
Tumor Lung Visualization and Localization through Virtual Reality and Thermal Feedback Interface
Previous Article in Journal
The Need for Artificial Intelligence Based Risk Factor Analysis for Age-Related Macular Degeneration: A Review
Previous Article in Special Issue
Well-Aerated Lung and Mean Lung Density Quantified by CT at Discharge to Predict Pulmonary Diffusion Function 5 Months after COVID-19
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Deep Learning Method for Detection of COVID-19 Infection Using Chest X-ray Images

1
Amity School of Engineering and Technology, Amity University Uttar Pradesh, Noida 201301, India
2
Department of Computer Science and Engineering, Malaviya National Institute of Technology, Jaipur 302017, India
3
Department of Electrical Engineering, Indian Institute of Technology Indore, Indore 453552, India
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(1), 131; https://doi.org/10.3390/diagnostics13010131
Submission received: 22 November 2022 / Revised: 6 December 2022 / Accepted: 19 December 2022 / Published: 30 December 2022
(This article belongs to the Special Issue Leading Diagnosis on Chest Imaging)

Abstract

:
The research community has recently shown significant interest in designing automated systems to detect coronavirus disease 2019 (COVID-19) using deep learning approaches and chest radiography images. However, state-of-the-art deep learning techniques, especially convolutional neural networks (CNNs), demand more learnable parameters and memory. Therefore, they may not be suitable for real-time diagnosis. Thus, the design of a lightweight CNN model for fast and accurate COVID-19 detection is an urgent need. In this paper, a lightweight CNN model called LW-CORONet is proposed that comprises a sequence of convolution, rectified linear unit (ReLU), and pooling layers followed by two fully connected layers. The proposed model facilitates extracting meaningful features from the chest X-ray (CXR) images with only five learnable layers. The proposed model is evaluated using two larger CXR datasets (Dataset-1: 2250 images and Dataset-2: 15,999 images) and the classification accuracy obtained are 98.67% and 99.00% on Dataset-1 and 95.67% and 96.25% on Dataset-2 for multi-class and binary classification cases, respectively. The results are compared with four contemporary pre-trained CNN models as well as state-of-the-art models. The effect of several hyperparameters: different optimization techniques, batch size, and learning rate have also been investigated. The proposed model demands fewer parameters and requires less memory space. Hence, it is effective for COVID-19 detection and can be utilized as a supplementary tool to assist radiologists in their diagnosis.

1. Introduction

The COVID-19 pandemic has caused severe healthcare crises across the globe in a very short period. The pandemic broke out in early December 2019 in Wuhan, China, and was declared a global pandemic on 11th March 2020 by the World Health Organization (WHO) [1]. Researchers across the world have reported COVID-19 to be a highly infectious disease that severely affects the respiratory system and has common symptoms like dry cough, myalgia, fever, headache, chest pain, and sore throat [2]. The current medical diagnostic processes lack proper medicine and drugs as well as hospital resources for the treatment of COVID-19 infection [3,4]. Reverse transcription-polymerase chain reaction (RT-PCR), a manual, time-consuming, and costly tool, is the most frequently used diagnostic method for detection which causes the risk to medical staff [5,6]. It is still an ongoing pandemic and has led to various variants, thereby, resulting in high mortality rates in many countries. Hence, there is a strong need for a safe and effective methodology that can detect COVID-19 infection at an early stage. The significance of two imaging modalities such as chest X-ray (CXR) and computed tomography (CT) has been studied for diagnosing COVID-19 [7,8,9]. However, manual visual inspection of both CXR and CT images is time-taking and tedious which may sometimes result in an inaccurate diagnosis [7,10]. Recently, tremendous efforts have been made in developing artificial intelligence (AI) based automated models for accurate diagnosis of COVID-19 to lessen the workload of radiologists [11].
Deep learning (DL) algorithms, especially, convolutional neural network (CNN), have offered efficient solutions for pneumonia detection in CXR images [12,13,14,15,16]. Soon after, many DL-based approaches have been reported to diagnose COVID-19 using CXR images. Ozturk et al. [17] proposed the DarkCovidNet model for COVID-19 detection in CXR images which obtained a classification accuracy of 98.08% and 87.02% for two-class and three-class scenarios, respectively. Hemdan et al. [18] proposed COVIDX-Net for binary classification task which was validated using only 50 CXR images. Narin et al. [19] proposed a model using the ResNet-50 and achieved a binary classification accuracy of 98% over 100 images. Ucar and Korkmaz [20] presented a multi-class classification system based on SqueezeNet and Bayesian optimizer that yielded an accuracy of 98.3%. Rahimzadeh and Attar [21] designed a model that concatenates Xception and ResNet50 networks and yields 91.4% accuracy for multi-class cases. Wang et al. [22] developed a tailored deep CNN model using CXR images that achieved an accuracy of 93.3%. Toğaçar et al. [23] achieved a classification accuracy of 97.06% using fuzzy color, stacking approach, and two DL approaches such as MobileNetV2 and SqueezeNet. Toraman et al. [24] proposed a convolutional CapsNet model to detect COVID-19 using CXR images and achieved an accuracy of 84.22% for multi-class classification cases. Han et al. [25] developed a deep 3D multiple instance learning methodology using chest CT images and obtained an accuracy of 94.3% on a three-class classification task. Zhang et al. [26] proposed a 7-layer-based CNN with stochastic pooling to detect COVID-19 from CT images. Wang et al. [27] developed a novel CCSHNet system for COVID-19 classification where two best pre-trained CNN models were selected to learn features and a discriminant correlation analysis method was used to fuse those features. Chaudhary and Pachori [28] used pretrained CNN models over the sub-band images generated from ourier-Bessel series expansion-based decomposition (FBSE) for COVID-19 detection. Joshi et al. [29] designed a multi-scale CNN for effective COVID-19 diagnosis from CT images. Recently, Bhattacharyya et al. [30] employed the VGG-19 model with the binary robust invariant scalable key points (BRISK) to detect COVID-19 cases from X-ray images. Jyoti et al. [31] proposed an automated COVID-19 detection method using memristive crossbar array-based tunable Q-wavelet transform (TQWT) and ResNet-50.
The literature shows that many existing DL-based methods have been validated using a few annotated CXR/CT images for COVID-19 diagnosis. The most frequently used DL models require a lot of training parameters and memory. The design of a lightweight DL method thus assumes significant urgency. Moreover, the performance of DL models relies on several hyperparameters: learning rate, batch size, type of optimizer, and the number of epochs, etc. However, there are only a limited number of studies on these parameters. Hence, there exists ample scope to conduct an investigation that includes an in-depth analysis of different hyperparameters to obtain the best possible results for COVID-19 detection. In a recent study [32], the COVID-19 classification performance was evaluated using eight pre-trained CNN models over a small dataset. All those models required a huge number of parameters. The performance of these models was evaluated for only binary classification scenarios. Nonetheless, despite being very challenging, a multi-class classification is in high demand.
To address the aforementioned issues, a lightweight CNN is proposed for the automated detection of COVID-19 infection in the current study. The major contributions of this study are summarized as follows:
  • A light-weight CNN model is proposed to diagnose COVID-19 infection in CXR images which require low computational cost and memory, thereby making it more suitable for real-time diagnosis.
  • The impact of several hyperparameters like different optimization techniques, number of epochs, batch size, and the learning rate is analyzed.
  • The performances of both binary (Normal and COVID-19) and multi-class (Normal, COVID-19, and pneumonia) classification cases have been studied.
  • A performance evaluation is conducted employing two larger datasets. Finally, the proposed model is compared with a few contemporary pre-trained CNN architectures including VGG-19 [33], ResNet-101 [34], DenseNet-121 [35], and Xception [36] in terms of the number of parameters and memory required. The performance of the model is also compared with a few state-of-the-art methods.
The remaining of this article is structured as follows: In Section 2, the datasets used and the suggested method for automated COVID-19 diagnosis are detailed. Section 3 presents the experimental setup and results. Finally, Section 4 draws the concluding remarks.

2. Materials and Methodology

This section exhaustively elucidates the dataset used in the current investigation and the proposed approach for the classification of various lung diseases. In addition, it delineates a few contemporary CNN architectures which are used for comparison purposes.

2.1. Datasets Used

The first dataset (Dataset-1) considered in our study comprises CXR images of three classes: Normal, COVID-19, and pneumonia that were collected from Figshare repository https://figshare.com/articles/COVID-19_Chest_X-RayImage_Repository/12580328, accessed on 21 November 2022 [37]. The quantity of available radiography images of COVID-19 patients is still limited. In addition, data imbalance has remained a major concern. Therefore, a total of 2250 CXR samples were collected for the three classes with 750 samples in each class. Dataset-1 mainly considered the COVID-19 cases from several open sources which are updated periodically.
To further verify the efficacy of the proposed model, we considered another dataset COVIDx-V7A (termed as Dataset-2 in this study) which comprises 15,999 CXR images from 15,122 patients across 51 countries and is the largest open-access dataset available to date [22]. These images were collected from five different data repositories and are available at https://github.com/lindawangg/COVID-Net, accessed on 21 November 2022. Table 1 shows the detailed specification of the considered datasets. The samples of a few frontal-view CXR images of different classes from Dataset-1 are presented in Figure 1.

2.2. Proposed Methodology

The proposed automated framework to classify COVID-19 infection cases from normal and pneumonia using CXR images are depicted in Figure 2. The framework consists of two stages: (i) prepossessing which includes image enhancement, data augmentation, and image normalization, and (ii) classification using the proposed light-weight CNN model (LW-CORONet). The detailed elucidation of each stage is given below.

2.2.1. Preprocessing

It is one of the most critical stages in the proposed approach. This section describes each technique deployed in this stage.
Image Enhancement: Image enhancement is frequently used in the biomedical image processing domain to improve image quality [38,39]. Therefore, in this study, we performed the image enhancement by employing bilinear interpolation (BI) followed by the contrast-limited adaptive histogram equalization (CLAHE) technique. Initially, BI was applied to replace the missing pixel by computing a weighted average of the nearest boundary pixels [40]. Then, CLAHE was employed for contrast enhancement that prevents over-enhancement of noise present in the images [41,42], thereby, facilitating better diagnosis. Figure 3 depicts the outcomes of the preprocessing stage (BI and CLAHE).
Data Augmentation: The CNN models perform well when trained using a large dataset [43,44]. However, the majority of biomedical imaging datasets lack sufficient data. Therefore, data augmentation has been primarily used to address the above concerns which enhances variability in images and acts as a regularizer [45,46]. In the current study, data augmentation is applied over the training dataset using four transformations: horizontal flipping, rotation by an angle of 15 degrees (clockwise), scaling with 10%, and Gaussian noise with zero mean and variance of 0.25.
Image Normalization: To preserve numerical stability, normalization has been widely adopted in the CNN models. It also facilitates faster learning and stability in the gradient descent technique [47]. Therefore, we employed normalization that normalizes each image pixel by multiplying it by 1/255 so that it falls in a range, i.e., 0 to 1.

2.2.2. Proposed Model

The DL models have revolutionized the field of AI because of their self-learning capabilities and superior classification performance in both binary and multi-class classification problems. Nevertheless, studies have been conducted recently in the field of biomedical image analysis either by implementing CNN models from scratch or by using transfer learning to achieve the best performance [48]. Both these methods have been recently studied for the classification of different lung infection diseases. However, most of these models deal with high computational costs and memory. Hence, the design of a lightweight CNN model is highly essential for the diagnosis of COVID-19 infection from CXR images. The proposed model includes three blocks: CBR blocks which contain convolution (CONV), batch normalization (BN), and rectified linear unit (ReLU) layer, each followed by a pooling layer and two fully connected (FC) layers at the end. A brief description of each layer is given below.

CONV Layer

This layer generates various feature maps by convolving the input volume with a set of filters [45]. It requires hyperparameters like the number of filters, size of the filter, stride, and padding.

BN Layer

BN layer helps to speed up the training process and normalizes the activation of the previous layer over a mini-batch [49]. It also acts as a regularizer to prevent overfitting issues. This layer is usually placed between CONV and ReLU layers.

ReLU Layer

This layer is essentially used to establish non-linearity in the model [45]. Due to its computational efficiency, ReLU has been extensively in DL models. It converts the negative values of inputs into zero and keeps positive values unchanged. The ReLU function can be defined mathematically as follows:
ϕ ( v ) = m a x ( 0 , v )

Pooling Layer

This layer down-samples the spatial size of the feature map and is generally placed in between consecutive CONV layers. It lessens the learnable parameters and computational cost, thereby handling the overfitting issues. Three popular pooling types are (i) max pooling, (ii) average pooling, and (iii) sum pooling; out of them, max pooling has been used in most of the recent studies [50]. Therefore, we used max pooling in this study.

FC Layer

The features extracted from the CONV layers are first flattened and then forwarded to the FC layers. They are placed at the end and are considered to be the final layers. These layers have similar characteristics to that of traditional neural networks. To handle the overfitting issue, a dropout layer is utilized between the FC layers. The last FC layer follows a softmax layer that yields a probability score with respect to every class which is used for classification.

2.2.3. Proposed LW-CORONet Model

The proposed LW-CORONet model is depicted in Figure 4. It has three CBR blocks, each block includes a sequence of CONV, BN, and ReLU layers and follows a maxpool layer and two consecutive FC layers at the end. The 50% dropout is applied on each of the FC layers to avoid overfitting. Eventually, a softmax layer is placed. The prime objectives of designing the lightweight custom CNN model include the reduction of computational cost and the number of learning parameters, which helps in increasing its learning speed as compared to the state-of-art CNN methods.
The first CONV layer inputs a CXR image of size 224 × 224 and convolves it employing 64 filters of size 5 × 5 with stride 2 which produces an output volume of size 110 × 110 × 64 . A max pooling operation with filter size 3 × 3 and a stride 3 is harnessed following the first CBR block, resulting in an output volume of size 36 × 36 × 64 . Then, this output is fed to the second CBR block where 128 kernels of size 3 × 3 are used that yield an output volume of size 34 × 34 × 128 . Further, a max pooling operation is employed to obtain an output of size 11 × 11 × 128 . Similarly, in the third block, 256 filters of size 3 × 3 are applied followed by a max pooling operation to generate a volume of size 3 × 3 × 256 . Then, this output volume is flattened and fed to a dense layer of 128 neurons. This layer is preceded by BN and dropout layers and followed by ReLU and BN layers. Finally, a soft-max layer is used at the end that produces the classification results. The learnable parameters involved in each layer and the detailed architecture is given in Table 2.

2.2.4. Pre-Trained CNN Models with Transfer Learning

A brief overview of four pre-trained CNN architectures considered in this work with the concept of transfer learning (TL) is presented in this section. With TL, the pre-trained models that were trained on large-scale datasets like ImageNet are further trained to learn a similar task. Therefore, the learning process is computationally faster than learning a model from the scratch. It also performs well in the absence of huge training data. A large number of efforts have been made using TL for COVID-19 diagnosis due to the unavailability of a large number of samples. Notably, we fine-tune only the final layer while keeping the pre-trained weights the same. In this study, we evaluated four architectures: VGG-19 [33], ResNet-101 [34], DenseNet-121 [35], and Xception [36] and compared the impact of these models with our proposed model. Table 3 shows the architectural overview of each pre-trained model along with the proposed scheme.

3. Experimental Setup and Results

In this section, we present the experimental setup and results to verify the effectiveness of the proposed model. A set of experiments were performed using two datasets: Dataset-1 of 2250 frontal-view CXR images (pneumonia: 750, COVID-19: 750, and normal: 750) and Dataset-2 of 15,999 CXR images (pneumonia: 5575, COVID-19: 2358, and Normal: 8066). These images were first rescaled into a size 224 × 224 . All CNN models were developed using the PyTorch toolbox and all experiments were conducted on the Google Colab GPU platform with NVIDIA Tesla T4 GPU of 16 GB RAM. The performance of both the suggested approach and pre-trained architectures was evaluated using 10-fold cross-validation (CV) for Dataset-1, wherein in each trial, one fold was utilized for testing, and the rest folds for training. From the training set, 20% of samples were randomly chosen for validation. For Dataset-2, results were evaluated based on the train-test division strategy reported by Wang et al. [22], wherein the test set contained 200 CXR images from each class and the remaining samples were retained for training the model, of which 20% are used for validation.
A different set of performance metrics such as sensitivity, specificity, precision, accuracy, and F1-score were used to assess each model. Furthermore, the heap-map results were computed using gradient-weighted class activation mapping (Grad-CAM) [51] to visually interpret the effectiveness of the model by highlighting the relevant regions. We evaluated the performance of the model in two different scenarios: the first scenario deals with three-class classification (pneumonia, COVID-19, and normal) and the second scenario deals with binary classification (Normal and COVID-19). The hyperparameter setting used in our study is presented in Table 4 which has been set empirically. Furthermore, a comparison analysis with pre-trained CNN architectures, namely VGG-19 [33], ResNet-101 [34], DenseNet-121 [35], and Xception [36] was done.

3.1. Results of the Proposed Model

The classification results obtained by the proposed model on both datasets are presented in this section.

3.1.1. Results on Dataset-1

The fold-wise results in terms of accuracy, precision, sensitivity, and F1-score for Dataset-1 are shown in Table 5. The proposed model was also tested over binary classification tasks and the results are shown in Table 6. Notably, the results provided in each fold were average results computed over all three classes. The average accuracy of 98.67% and 99.00% was obtained for multi-class and binary-class classification tasks, respectively. The confusion matrices in each run of 10-fold CV for multi-class and binary class scenarios are exhibited in Figure 5 and Figure 6, respectively. Figure 7 shows the training and validation loss curves. It can be observed that the proposed model converges within 100 epochs. The plot is shown for a single run of a 10-fold CV. The Mathew correlation coefficient (MCC) and kappa score were computed as 0.9730 and 0.9860, respectively which indicate better prediction results for each class. The pictorial presentation of MCC and kappa score along with multi-class classification accuracy is depicted in Figure 8. These values were recorded at different epochs and computed for a single fold. Figure 9 shows the receiver operating characteristic (ROC) curves obtained by the suggested and pre-trained models for binary classification scenarios.
Finally, heat maps were provided using Grad-CAM to verify the visual interpretability of the proposed model. Figure 10 illustrates the heat map results of a few sample normal and COVID-19 CXR images. The proposed model could locate the suspicious regions that indicated better interpretability of the classification results. Hence, it can be helpful to assist radiologists in their diagnosis. Figure 10h depicts the heat map of a normal sample where suspicious regions are not indicated.
The computational cost of the proposed model was evaluated in terms of time (in seconds). Our model takes around 8.5 × 10 3 s in the training stage to converge, which is comparatively faster than training the state-of-the-art CNN architectures. While testing an image takes approximately 0.035 s using the proposed model.

3.1.2. Results on Dataset-2

The performance of the proposed model is evaluated on Dataset 2. Table 7 shows the classification results for both the classification scenarios (multi-class and binary classification) for Dataset-2. The average accuracy of 95.67% and 96.25% was achieved for multi-class and binary classification, respectively. Figure 11 shows the confusion matrices for both scenarios.

3.2. Experiment on Different Hyperparameters

In this section, the impact of different hyperparameters, namely the number of epochs, batch size, learning rate, different optimizers, etc., is evaluated to identify the best classification performance. It is worth noting here that all experiments were conducted using Dataset-1.

3.2.1. Impact of Optimization Techniques

In this experiment, we explore several optimization techniques such as SGD [52], Adam [53], RMSProp [54], and AdaGrad [55] to ascertain the best classification performance. Table 8 shows the detailed classification results using different optimizers. It can be seen that observed that the classification performance of the LW-CORONet with Adam optimizer is promising as compared to others. Therefore, all remaining experiments in this study were carried out using the Adam optimizer.

3.2.2. Impact of Learning Rate

In this experiment, we investigated the impact of different learning rates, and the learning rate with minimum validation loss was chosen. Figure 12 illustrates a plot of learning rates versus validation loss. The point (represented in red dot) in the Figure 12 specifies the optimal learning rate from which the loss drops significantly, thereby, facilitating better classification performance.

3.2.3. Impact of Different Batch Sizes

We evaluated the effect of different batch sizes in this experiment. The performance of LW-CORONet when trained with batch sizes of 32, 16, and 8 is tabulated in Table 9. These results showed that the proposed model with a batch size of 32 results in stable and higher testing performance.

3.3. Misclassification Results Analysis

Figure 13 shows few misclassification results yielded by LW-CORONet on Dataset-1. These errors possibly occurred because of the similar visual features among the CXR images of the three classes.

3.4. Comparative Analysis of Pre-Trained CNN Architectures and the Proposed Model

We performed a comparative analysis among the pre-trained classification architectures, VGG-19 [33], ResNet-101 [34], DenseNet [35], and Xception [36] and the proposed model. The detailed classification performance on Dataset-1 is tabulated in Table 10. It is worth mentioning here that these models were validated on the same images as done for the proposed model and the experimental setup was also the same. The proposed model required fewer parameters and memory but obtained a comparable or better performance compared to effective architectures like ResNet-101. All these pre-trained models were originally designed over ImageNet [56] dataset but, were fined tuned later using the CXR images to perform multi-class classification. We fine-tuned only the last layer of these models and initialized other layers with the pre-trained weights.

3.5. Comparison with Existing Approaches

The proposed model was compared against the existing DL-based schemes for automated COVID-19 detection in CXR images. Table 11 summarizes the obtained performance. It can be observed that the suggested model obtained promising results as compared with other approaches on Dataset-1 and achieved comparable or better performance even on the larger dataset (Dataset-2). The proposed model was validated using a relatively larger number of CXR samples compared to several recent studies [17,18,19,23,28,30,31,32]). Although studies in [20,21] employed datasets of comparable sizes, the COVID-19 class samples are only 76 and 180, respectively which imposes the class imbalance problem. Further, several studies were focused on solving either binary or multi-class classification tasks. But, the current study investigated both tasks and achieved an accuracy of 99.00% and 98.67% for binary and multi-class classification respectively on Dataset-1. Similarly, for Dataset-2, an accuracy of 96.25% and 95.67% was achieved for binary and multi-class classification, respectively.

3.6. Discussion

An automated DL-based model, LW-CORONet, is proposed for the effective detection of COVID-19 infection using CXR images. Of late, in the setting of COVID-19 detection, many DL-based studies have been performed; however, most of these studies were limited to smaller datasets and require huge memory space and higher computational costs. Hence, LW-CORONet was designed aiming to handle these issues. LW-CORONet was capable of learning discriminant features directly from CXR images while demanding fewer learning parameters and memory.
To verify the effectiveness of the proposed scheme, several extensive experiments were carried out using two larger datasets of 2250 and 15999 CXR samples from three categories such as normal, pneumonia, and COVID-19. This study also explored the impact of various hyperparameters: batch size, learning rate, and optimizers. A comparative analysis was done with four contemporary pre-trained classification networks and state-of-art approaches. The proposed scheme yielded a higher classification accuracy of 98.67% and 99.00% for multi-class and binary cases respectively on Dataset-1, whereas it is 95.67% and 96.25% on a larger dataset (Dataset-2). The proposed model is thus effective, lightweight, and hence can be utilized by radiologists for the early diagnosis of COVID-19 infection and pneumonia. The main advantages of the proposed model are as follows:
  • The proposed LW-CORONet has only five learnable layers which promote the learning of high-level features automatically from the CXR samples.
  • The proposed model is well-suited for both binary and multi-class classification scenarios and does not involve a hand-held feature extraction process.
  • The proposed architecture demands very few parameters compared to other CNN models and thus, necessitates low computational cost.
  • The proposed architecture is lightweight and uses less memory space.
The major drawback of this investigation is that the proposed model was trained with limited COVID-19 data due to the unavailability of a large-scale COVID-19 dataset.

4. Conclusions

This paper proposed, LW-CORONet, a novel model, for the early and accurate detection of COVID-19 infection from CXR images. The LW-CORONet was composed of five learnable layers with low computational power and memory space to extract detailed features. For the validation of the proposed scheme, an extensive set of experiments was carried out using two publicly available CXR datasets with larger COVID-19 samples. The effect of notable hyperparameters was verified on the proposed model to detect the best detection performance. Comparisons with pre-trained CNN models as well as current existing approaches revealed the superiority of the proposed approach in both multi-class and binary classification scenarios. Furthermore, the LW-CORONet model is better in terms of memory and computational cost. Overall, the suggested model is effective, lightweight, and hence can be suitable for clinicians for real-time COVID-19 diagnosis. However, in the future, the performance of LW-CORONet is suggested to be verified over a big and diverse dataset with more COVID-19 samples. Also, the lungs affected regions for both pneumonia and COVID-19 cases can be identified simultaneously with the classification of CXR images.

Author Contributions

Conceptualization, D.R.N.; methodology, D.R.N. and S.R.N.; software, D.R.N., U.S. and V.A.; validation, D.R.N., S.R.N. and R.B.P.; formal analysis, U.S. and V.A.; investigation, S.R.N., D.R.N. and R.B.P.; resources, S.R.N.; data curation, U.S. and V.A.; writing—original draft preparation, S.R.N., U.S. and V.A.; writing—review and editing, D.R.N. and R.B.P.; visualization, D.R.N., U.S. and V.A.; supervision, D.R.N. and R.B.P.; project administration, S.R.N. and D.R.N.; funding acquisition, D.R.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

In this study, we used two publicly available COVID-19 chest X-ray datasets: Dataset-1 (https://figshare.com/articles/COVID-19_Chest_X-RayImage_Repository/12580328) and Dataset-2 (https://github.com/lindawangg/COVID-Net).

Acknowledgments

We are thankful to the authors of [22] for making the database available publicly.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. WHO. Coronavirus Disease 2019 (COVID-19) Situation Report–127. 2020. Available online: https://www.who.int/docs/default-source/coronaviruse/situation-reports/20200526-covid-19-sitrep-127.pdf?sfvrsn=7b6655ab_8 (accessed on 21 November 2022).
  2. Huang, C.; Wang, Y.; Li, X.; Ren, L.; Zhao, J.; Hu, Y.; Zhang, L.; Fan, G.; Xu, J.; Gu, X.; et al. Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China. Lancet 2020, 395, 497–506. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Giacomo, G.; Antonio, P.; Maurizio, C. Critical Care Utilization for the COVID-19 Outbreak in Lombardy, Italy. JAMA 2020, 323, 1545. [Google Scholar]
  4. Jiang, X.; Coffee, M.; Bari, A.; Wang, J.; Jiang, X.; Huang, J.; Shi, J.; Dai, J.; Cai, J.; Zhang, T.; et al. Towards an artificial intelligence framework for data-driven prediction of coronavirus clinical severity. CMC Comput. Mater. Contin. 2020, 63, 537–551. [Google Scholar] [CrossRef]
  5. Fang, Y.; Zhang, H.; Xie, J.; Lin, M.; Ying, L.; Pang, P.; Ji, W. Sensitivity of chest CT for COVID-19: Comparison to RT-PCR. Radiology 2020, 296, 200432. [Google Scholar] [CrossRef] [PubMed]
  6. Dong, D.; Tang, Z.; Wang, S.; Hui, H.; Gong, L.; Lu, Y.; Xue, Z.; Liao, H.; Chen, F.; Yang, F.; et al. The role of imaging in the detection and management of COVID-19: A review. IEEE Rev. Biomed. Eng. 2021, 14, 16–29. [Google Scholar] [CrossRef]
  7. Kanne, J.P.; Little, B.P.; Chung, J.H.; Elicker, B.M.; Ketai, L.H. Essentials for radiologists on COVID-19: An update radiology scientific expert panel. Radiology 2020, 296, E113–E114. [Google Scholar] [CrossRef] [Green Version]
  8. Chung, M.; Bernheim, A.; Mei, X.; Zhang, N.; Huang, M.; Zeng, X.; Cui, J.; Xu, W.; Yang, Y.; Fayad, Z.A.; et al. CT imaging features of 2019 novel coronavirus (2019-nCoV). Radiology 2020, 295, 202–207. [Google Scholar] [CrossRef] [Green Version]
  9. Joshi, A.M.; Nayak, D.R. MFL-Net: An Efficient Lightweight Multi-Scale Feature Learning CNN for COVID-19 Diagnosis from CT Images. IEEE J. Biomed. Health Inform. 2022, 26, 5355–5363. [Google Scholar] [CrossRef]
  10. Xie, X.; Zhong, Z.; Zhao, W.; Zheng, C.; Wang, F.; Liu, J. Chest CT for typical 2019-nCoV pneumonia: Relationship to negative RT-PCR testing. Radiology 2020, 296, 200343. [Google Scholar] [CrossRef] [Green Version]
  11. Shi, F.; Wang, J.; Shi, J.; Wu, Z.; Wang, Q.; Tang, Z.; He, K.; Shi, Y.; Shen, D. Review of artificial intelligence techniques in imaging data acquisition, segmentation and diagnosis for COVID-19. IEEE Rev. Biomed. Eng. 2021, 14, 4–15. [Google Scholar] [CrossRef] [Green Version]
  12. Gu, X.; Pan, L.; Liang, H.; Yang, R. Classification of bacterial and viral childhood pneumonia using deep learning in chest radiography. In Proceedings of the 3rd International Conference on Multimedia and Image Processing, Guiyang, China, 16–18 March 2018; pp. 88–93. [Google Scholar]
  13. Chouhan, V.; Singh, S.K.; Khamparia, A.; Gupta, D.; Tiwari, P.; Moreira, C.; Damaševičius, R.; De Albuquerque, V.H.C. A novel transfer learning based approach for pneumonia detection in chest X-ray images. Appl. Sci. 2020, 10, 559. [Google Scholar] [CrossRef]
  14. Lakhani, P.; Sundaram, B. Deep learning at chest radiography: Automated classification of pulmonary tuberculosis by using convolutional neural networks. Radiology 2017, 284, 574–582. [Google Scholar] [CrossRef] [PubMed]
  15. Rajpurkar, P.; Irvin, J.; Ball, R.L.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.P.; et al. Deep learning for chest radiograph diagnosis: A retrospective comparison of the CheXNeXt algorithm to practicing radiologists. PLoS Med. 2018, 15, e1002686. [Google Scholar] [CrossRef] [PubMed]
  16. Chaudhary, P.K.; Pachori, R.B. Automatic diagnosis of COVID-19 and pneumonia using FBD method. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Republic of Korea, 16–19 December 2020; pp. 2257–2263. [Google Scholar]
  17. Ozturk, T.; Talo, M.; Yildirim, E.A.; Baloglu, U.B.; Yildirim, O.; Acharya, U.R. Automated detection of COVID-19 cases using deep neural networks with X-ray images. Comput. Biol. Med. 2020, 121, 103792. [Google Scholar] [CrossRef] [PubMed]
  18. Hemdan, E.E.D.; Shouman, M.A.; Karar, M.E. COVIDX-Net: A framework of deep learning classifiers to diagnose covid-19 in X-ray images. arXiv 2020, arXiv:2003.11055. [Google Scholar]
  19. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. arXiv 2020, arXiv:2003.10849. [Google Scholar] [CrossRef]
  20. Ucar, F.; Korkmaz, D. COVIDiagnosis-Net: Deep Bayes-SqueezeNet based Diagnostic of the Coronavirus Disease 2019 (COVID-19) from X-Ray Images. Med. Hypotheses 2020, 140, 109761. [Google Scholar] [CrossRef]
  21. Rahimzadeh, M.; Attar, A. A modified deep convolutional neural network for detecting COVID-19 and pneumonia from chest X-ray images based on the concatenation of Xception and ResNet50V2. Inform. Med. Unlocked 2020, 19, 100360. [Google Scholar] [CrossRef]
  22. Wang, L.; Lin, Z.Q.; Wong, A. COVID-Net: A tailored deep convolutional neural network design for detection of COVID-19 cases from chest X-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef]
  23. Toğaçar, M.; Ergen, B.; Cömert, Z. COVID-19 detection using deep learning models to exploit Social Mimic Optimization and structured chest X-ray images using fuzzy color and stacking approaches. Comput. Biol. Med. 2020, 121, 103805. [Google Scholar] [CrossRef]
  24. Toraman, S.; Alakuş, T.B.; Türkoğlu, İ. Convolutional capsnet: A novel artificial neural network approach to detect COVID-19 disease from X-ray images using capsule networks. Chaos Solitons Fractals 2020, 140, 110122. [Google Scholar] [CrossRef] [PubMed]
  25. Han, Z.; Wei, B.; Hong, Y.; Li, T.; Cong, J.; Zhu, X.; Wei, H.; Zhang, W. Accurate Screening of COVID-19 using Attention Based Deep 3D Multiple Instance Learning. IEEE Trans. Med. Imaging 2020, 39, 2584–2594. [Google Scholar] [CrossRef] [PubMed]
  26. Zhang, Y.D.; Satapathy, S.C.; Zhu, L.Y.; Górriz, J.M.; Wang, S.H. A seven-layer convolutional neural network for chest CT based COVID-19 diagnosis using stochastic pooling. IEEE Sens. J. 2020, 22, 17573–17582. [Google Scholar] [CrossRef] [PubMed]
  27. Wang, S.H.; Nayak, D.R.; Guttery, D.S.; Zhang, X.; Zhang, Y.D. COVID-19 classification by CCSHNet with deep fusion using transfer learning and discriminant correlation analysis. Inf. Fusion 2021, 68, 131–148. [Google Scholar] [CrossRef] [PubMed]
  28. Chaudhary, P.K.; Pachori, R.B. FBSED based automatic diagnosis of COVID-19 using X-ray and CT images. Comput. Biol. Med. 2021, 134, 104454. [Google Scholar] [CrossRef] [PubMed]
  29. Joshi, A.M.; Nayak, D.R.; Das, D.; Zhang, Y.D. LiMS-Net: A Lightweight Multi-Scale CNN for COVID-19 Detection from Chest CT Scans. ACM Trans. Manag. Inf. Syst. (TMIS) 2022. [Google Scholar] [CrossRef]
  30. Bhattacharyya, A.; Bhaik, D.; Kumar, S.; Thakur, P.; Sharma, R.; Pachori, R.B. A deep learning based approach for automatic detection of COVID-19 cases using chest X-ray images. Biomed. Signal Process. Control 2022, 71, 103182. [Google Scholar] [CrossRef]
  31. Jyoti, K.; Sushma, S.; Kumar, P.; Yadav, S.; Pachori, R.B.; Mukherjee, S. Automatic diagnosis of COVID-19 with MCA-inspired TQWT-based classification of chest X-ray images. Comput. Biol. Med. 2022, 152, 106331. [Google Scholar] [CrossRef]
  32. Nayak, S.R.; Nayak, D.R.; Sinha, U.; Arora, V.; Pachori, R.B. Application of deep learning techniques for detection of COVID-19 cases using chest X-ray images: A comprehensive study. Biomed. Signal Process. Control 2021, 64, 102365. [Google Scholar] [CrossRef]
  33. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  34. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  35. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  36. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  37. Haghanifar, A.; Majdabadi, M.M.; Choi, Y.; Deivalakshmi, S.; Ko, S. COVID-CXNet: Detecting COVID-19 in frontal chest X-ray images using deep learning. arXiv 2020, arXiv:2006.13807. [Google Scholar] [CrossRef]
  38. Senthilkumaran, N.; Thimmiaraja, J. Histogram equalization for image enhancement using MRI brain images. In Proceedings of the 2014 World Congress on Computing and Communication Technologies, Trichirappalli, India, 27 February–1 March 2014; pp. 80–83. [Google Scholar]
  39. Soomro, T.A.; Afifi, A.J.; Shah, A.A.; Soomro, S.; Baloch, G.A.; Zheng, L.; Yin, M.; Gao, J. Impact of Image Enhancement Technique on CNN Model for Retinal Blood Vessels Segmentation. IEEE Access 2019, 7, 158183–158197. [Google Scholar] [CrossRef]
  40. Agrafiotis, D. Video Error Concealment. In Academic Press Library in Signal Processing; Elsevier: Amsterdam, The Netherlands, 2014; Volume 5, pp. 295–321. [Google Scholar]
  41. Pisano, E.D.; Zong, S.; Hemminger, B.M.; DeLuca, M.; Johnston, R.E.; Muller, K.; Braeuning, M.P.; Pizer, S.M. Contrast limited adaptive histogram equalization image processing to improve the detection of simulated spiculations in dense mammograms. J. Digit. Imaging 1998, 11, 193. [Google Scholar] [CrossRef] [PubMed]
  42. Nayak, D.R.; Dash, R.; Majhi, B. Discrete ripplet-II transform and modified PSO based improved evolutionary extreme learning machine for pathological brain detection. Neurocomputing 2018, 282, 232–247. [Google Scholar] [CrossRef]
  43. Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  45. Nayak, D.R.; Dash, R.; Majhi, B. Automated Diagnosis of Multi-class Brain Abnormalities using MRI Images: A Deep Convolutional Neural Network based Method. Pattern Recognit. Lett. 2020, 138, 385–391. [Google Scholar] [CrossRef]
  46. Zhang, Y.D.; Nayak, D.R.; Zhang, X.; Wang, S.H. Diagnosis of secondary pulmonary tuberculosis by an eight-layer improved convolutional neural network with stochastic pooling and hyperparameter optimization. J. Ambient. Intell. Humaniz. Comput. 2020, 1–18. [Google Scholar] [CrossRef]
  47. Swati, Z.N.K.; Zhao, Q.; Kabir, M.; Ali, F.; Ali, Z.; Ahmed, S.; Lu, J. Brain tumor classification for MR images using transfer learning and fine-tuning. Comput. Med. Imaging Graph. 2019, 75, 34–46. [Google Scholar] [CrossRef]
  48. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Van Der Laak, J.A.; Van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  49. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  50. Boureau, Y.L.; Ponce, J.; LeCun, Y. A theoretical analysis of feature pooling in visual recognition. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 June 2010; pp. 111–118. [Google Scholar]
  51. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  52. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the importance of initialization and momentum in deep learning. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 1139–1147. [Google Scholar]
  53. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  54. Hinton, G.; Srivastava, N.; Swersky, K. Neural Networks for Machine Learning. Lecture 6a Overview of Mini-Batch Gradient Descent Course. 2012. Available online: https://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf (accessed on 21 November 2022).
  55. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  56. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
Figure 1. Frontal-view CXR samples of three categories: (a) Normal, (b) Pneumonia, and (c) COVID-19.
Figure 1. Frontal-view CXR samples of three categories: (a) Normal, (b) Pneumonia, and (c) COVID-19.
Diagnostics 13 00131 g001
Figure 2. Block diagram of the proposed approach for automated detection of COVID-19 infection.
Figure 2. Block diagram of the proposed approach for automated detection of COVID-19 infection.
Diagnostics 13 00131 g002
Figure 3. Results of image enhancement (BI and CLAHE): (ac) indicate original images, (df) indicate enhanced images.
Figure 3. Results of image enhancement (BI and CLAHE): (ac) indicate original images, (df) indicate enhanced images.
Diagnostics 13 00131 g003
Figure 4. Illustration of proposed LW-CORONet model.
Figure 4. Illustration of proposed LW-CORONet model.
Diagnostics 13 00131 g004
Figure 5. (aj) Confusion matrix obtained by LW-CORONet fold-wise (fold 1 to fold 10) for multi-class classification case.
Figure 5. (aj) Confusion matrix obtained by LW-CORONet fold-wise (fold 1 to fold 10) for multi-class classification case.
Diagnostics 13 00131 g005
Figure 6. (aj) Confusion matrix obtained by LW-CORONet fold-wise (fold 1 to fold 10) for binary classification case.
Figure 6. (aj) Confusion matrix obtained by LW-CORONet fold-wise (fold 1 to fold 10) for binary classification case.
Diagnostics 13 00131 g006
Figure 7. Illustration of the loss convergence plot for the first run of 10-fold CV.
Figure 7. Illustration of the loss convergence plot for the first run of 10-fold CV.
Diagnostics 13 00131 g007
Figure 8. Graphical plot of Mathews correlation, kappa score, and accuracy for LW-CORONet model.
Figure 8. Graphical plot of Mathews correlation, kappa score, and accuracy for LW-CORONet model.
Diagnostics 13 00131 g008
Figure 9. ROC curves of LW-CORONet and the pre-trained CNNs.
Figure 9. ROC curves of LW-CORONet and the pre-trained CNNs.
Diagnostics 13 00131 g009
Figure 10. Grad-CAM visualization results of the proposed model on a few sample CXR images. (ad) The top row indicates the original images. (eh) The bottom row indicates the heat maps obtained by LW-CORONet.
Figure 10. Grad-CAM visualization results of the proposed model on a few sample CXR images. (ad) The top row indicates the original images. (eh) The bottom row indicates the heat maps obtained by LW-CORONet.
Diagnostics 13 00131 g010
Figure 11. (a,b) Confusion matrix obtained for both cases on Dataset-2.
Figure 11. (a,b) Confusion matrix obtained for both cases on Dataset-2.
Diagnostics 13 00131 g011
Figure 12. Plot between the learning rate and validation loss.
Figure 12. Plot between the learning rate and validation loss.
Diagnostics 13 00131 g012
Figure 13. Sample misclassification outcomes of LW-CORONet.
Figure 13. Sample misclassification outcomes of LW-CORONet.
Diagnostics 13 00131 g013
Table 1. Description of the datasets.
Table 1. Description of the datasets.
DatasetNumber of CXR ImagesTotal
NormalPneumoniaCOVID-19
Dataset-17507507502250
Dataset-280665575235815,999
Table 2. Detail configuration of the proposed LW-CORONet.
Table 2. Detail configuration of the proposed LW-CORONet.
LayerActivation SizeParametersStrideKernel SizeFilter
Input(3, 224, 224)0---
CONV(64, 110, 110)48642 5 × 5 64
BN(64, 110, 110)128---
ReLU(64, 110, 110)0---
MaxPool(64, 36, 36)03 3 × 3 -
CONV(128, 34, 34)73,8561 3 × 3 128
BN(128, 34, 34)256---
ReLU(128, 34, 34)0---
MaxPool(128, 11, 11)03 3 × 3 -
CONV(256, 9, 9)295,1681 3 × 3 256
BN(256, 9, 9)512---
ReLU(256, 9, 9)0---
MaxPool(256, 3, 3)03 3 × 3 -
BN23044068---
Dropout (0.5)23040---
FC128295,040---
ReLU1280---
BN128256---
Dropout (0.5)1280---
Softmax3387---
Total ≊ 0.68 million
Table 3. Architectural details of pre-trained CNN architecture.
Table 3. Architectural details of pre-trained CNN architecture.
ArchitectureLayersInput Layer SizeOutput Layer SizeParameters
ResNet-101101(224,224,3)(3,1) 44.1 × 10 6
VGG-1919(224,224,3)(3,1) 143 × 10 6
DenseNet-121121(224,224,3)(3,1) 7.5 × 10 6
Xception71(299,299,3)(3,1) 22.3 × 10 6
Proposed LW-CORONet5(224,224,3)(3,1) 0.68 × 10 6
Table 4. Illustration of hyperparameter setting used in this study.
Table 4. Illustration of hyperparameter setting used in this study.
HyperparameterValue
Batch size32
Learning rate0.0005
Epochs100
OptimizerAdam
Loss functionCategorical cross-entropy
Table 5. Fold-wise multi-class classification performance (in %) on Dataset-1.
Table 5. Fold-wise multi-class classification performance (in %) on Dataset-1.
FoldAccuracyPrecisionSensitivitySpecificityF1-ScoreCOVID-19 Class Accuracy
197.9296.8996.8998.4496.8898.66
299.4199.1299.1199.5699.1199.56
397.9797.1597.0398.4996.9999.11
498.8298.6597.3399.3397.8899.56
597.6396.4596.4498.2296.4497.33
697.3596.0296.0298.0196.0197.78
799.1198.6998.6799.3398.6799.11
899.4199.1299.1199.5699.1199.56
999.4199.1399.1199.5699.1199.11
1099.7099.5699.5699.7899.56100.00
Average98.6798.0897.9399.0397.9898.98
Table 6. Fold-wise binary classification performance (in %) on Dataset-1.
Table 6. Fold-wise binary classification performance (in %) on Dataset-1.
FoldAccuracyPrecisionSensitivitySpecificityF1-ScoreCOVID-19 Class Accuracy
199.33100.0098.67100.0099.3398.67
298.6797.40100.0097.3398.68100.00
399.3398.68100.0098.6799.33100.00
498.6798.6798.6798.6798.6798.67
5100.00100.00100.00100.00100.00100.00
698.0098.6597.3398.6797.9997.33
799.3398.68100.0098.6799.34100.00
898.6797.40100.0097.3398.68100.00
999.33100.0098.67100.0099.3398.67
1098.6798.6798.6798.6798.6798.67
Average99.0098.8299.2098.8099.0099.20
Table 7. Classification performance (in %) of LW-CORONet on Dataset-2.
Table 7. Classification performance (in %) of LW-CORONet on Dataset-2.
TaskAccuracyPrecisionSensitivitySpecificityF1-ScoreCOVID-19 Class Accuracy
Multi-class95.6793.5193.5096.7593.5095.00
Binary96.2595.5797.0095.5096.2897.00
Table 8. Classification results (in %) obtained by various optimizers.
Table 8. Classification results (in %) obtained by various optimizers.
OptimizerAccuracyPrecisionSensitivitySpecificityF1-Score
SGD98.1897.5597.4298.6197.39
RMSProp97.9497.2097.0598.4397.03
AdaGrad98.5598.0197.9198.8997.90
Adam98.6798.0897.9399.0397.98
Table 9. Accuracy of LW-CORONet with different batch sizes.
Table 9. Accuracy of LW-CORONet with different batch sizes.
ModelBatch Size
32168
Proposed LW-CORONet98.6798.0397.26
Table 10. Classification performance comparison (in %) among pre-trained CNNs.
Table 10. Classification performance comparison (in %) among pre-trained CNNs.
ModelAccuracyPrecisionSensitivitySpecificityF1-ScoreParameterMemory
ResNet-10198.1997.2997.2998.6497.2944.1 M168 MB
VGG-1997.8196.7696.7198.3696.71143 M547 MB
DenseNet-12197.4596.1696.2998.1096.207.5 M31 MB
Xception96.9795.4395.5397.7495.4622.3 M84 MB
LW-CORONet98.6798.0897.9399.0397.980.68 M6 MB
Table 11. Comparison of proposed model with existing COVID-19 detection approaches.
Table 11. Comparison of proposed model with existing COVID-19 detection approaches.
ReferenceMethodNumber of CXR SamplesAccuracy (%)
Binary ClassMulti-Class
Hemdan et al. [18]COVIDX-Net5090.00
( C 1 : 25 and C 2 : 25)
Narin et al. [19]ResNet-5010098.00
( C 1 : 50 and C 2 : 50)
Ozturk et al. [17]DarkCovidNet112598.0887.02
( C 1 : 125, C 2 : 500 and C 3 : 500)
Ucar and Korkmaz [20]Bayes-SqueezeNet594998.30
( C 1 : 76, C 2 : 1583 and C 3 : 4290)
Rahimzadeh and Attar [21]Xception and ResNet50V21508591.40
C 1 : 180, C 2 : 6054 and C 3 : 8851
Toğaçar et al. [23]SqueezeNet and MobileNetV245898.25
SVM( C 1 : 295, C 2 : 65 and C 3 : 98)
Nayak et al. [32]ResNet-3440698.33
( C 1 :203 and C 2 : 203)
Toramana et al. [24]CapsNet315097.2484.22
C 1 : 1050, C 2 : 1050 and C 3 : 1050
Chaudhary and Pachori [28]FBSED and Inception-ResNet-v21446-93.06
C 1 : 482, C 2 : 482 and C 3 : 482
Bhattacharyya et al. [30]VGG-19 and BRISK1030-96.60
C 1 : 342, C 2 : 341 and C 3 : 347
Jyoti et al. [31]ResNet-50 and MCA-TQWT527594.64-
C 1 : 2409 and C 2 : 2866
Proposed modelLW-CORONetDataset-1: 225099.0098.67
( C 1 : 750, C 2 : 750 and C 3 : 750)
Dataset-2: 15,99996.2595.67
( C 1 : 2358, C 2 : 8066 and C 3 : 5575)
C1: COVID-19, C2: Normal, C3: Pneumonia.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nayak, S.R.; Nayak, D.R.; Sinha, U.; Arora, V.; Pachori, R.B. An Efficient Deep Learning Method for Detection of COVID-19 Infection Using Chest X-ray Images. Diagnostics 2023, 13, 131. https://doi.org/10.3390/diagnostics13010131

AMA Style

Nayak SR, Nayak DR, Sinha U, Arora V, Pachori RB. An Efficient Deep Learning Method for Detection of COVID-19 Infection Using Chest X-ray Images. Diagnostics. 2023; 13(1):131. https://doi.org/10.3390/diagnostics13010131

Chicago/Turabian Style

Nayak, Soumya Ranjan, Deepak Ranjan Nayak, Utkarsh Sinha, Vaibhav Arora, and Ram Bilas Pachori. 2023. "An Efficient Deep Learning Method for Detection of COVID-19 Infection Using Chest X-ray Images" Diagnostics 13, no. 1: 131. https://doi.org/10.3390/diagnostics13010131

APA Style

Nayak, S. R., Nayak, D. R., Sinha, U., Arora, V., & Pachori, R. B. (2023). An Efficient Deep Learning Method for Detection of COVID-19 Infection Using Chest X-ray Images. Diagnostics, 13(1), 131. https://doi.org/10.3390/diagnostics13010131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop