Next Article in Journal
An Improved COVID-19 Forecasting by Infectious Disease Modelling Using Machine Learning
Next Article in Special Issue
In Silico Screening of Potential Phytocompounds from Several Herbs against SARS-CoV-2 Indian Delta Variant B.1.617.2 to Inhibit the Spike Glycoprotein Trimer
Previous Article in Journal
Understanding Horizon 2020 Data: A Knowledge Graph-Based Approach
Previous Article in Special Issue
Screening Support System Based on Patient Survey Data—Case Study on Classification of Initial, Locally Collected COVID-19 Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

COVID-19 Patient Detection Based on Fusion of Transfer Learning and Fuzzy Ensemble Models Using CXR Images

by
Chandrakanta Mahanty
1,
Raghvendra Kumar
1,
Panagiotis G. Asteris
2 and
Amir H. Gandomi
3,*
1
Department of Computer Science and Engineering, GIET University, Rayagada 765022, Odisha, India
2
Computational Mechanics Laboratory, Department of Civil Engineering, School of Pedagogical and Technological Education, 15122 Athens, Greece
3
Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(23), 11423; https://doi.org/10.3390/app112311423
Submission received: 27 October 2021 / Revised: 24 November 2021 / Accepted: 29 November 2021 / Published: 2 December 2021
(This article belongs to the Special Issue Artificial Intelligence Computing and Applications for COVID-19)

Abstract

:
The COVID-19 pandemic has claimed the lives of millions of people and put a significant strain on healthcare facilities. To combat this disease, it is necessary to monitor affected patients in a timely and cost-effective manner. In this work, CXR images were used to identify COVID-19 patients. We compiled a CXR dataset with equal number of 2313 COVID positive, pneumonia and normal CXR images and utilized various transfer learning models as base classifiers, including VGG16, GoogleNet, and Xception. The proposed methodology combines fuzzy ensemble techniques, such as Majority Voting, Sugeno Integral, and Choquet Fuzzy, and adaptively combines the decision scores of the transfer learning models to identify coronavirus infection from CXR images. The proposed fuzzy ensemble methods outperformed each individual transfer learning technique and several state-of-the-art ensemble techniques in terms of accuracy and prediction. Specifically, VGG16 + Choquet Fuzzy, GoogleNet + Choquet Fuzzy, and Xception + Choquet Fuzzy achieved accuracies of 97.04%, 98.48%, and 99.57%, respectively. The results of this work are intended to help medical practitioners achieve an earlier detection of coronavirus compared to other detection strategies, which can further save millions of lives and advantageously influence society.

1. Introduction

Coronavirus is an extremely contagious virus that mostly affects the lungs and is characterized by symptoms of cough, fever, and exhaustion. The disease has the potential to spread to the lower respiratory system, where it can potentially cause inflammation and swelling, resulting in pneumonia [1].
COVID-19 spreads by respiratory droplet production following coughing and patient contact. As a result, preventative methods, such as wearing masks, social distancing, and isolating infected people, have been implemented to deal with viral transmission thus far [2]. COVID-19 has devastated most healthcare systems around the world, causing economic damage and claiming the lives of numerous individuals. Although vaccines are being produced and administered around the globe, it will be a very long time before each human can be vaccinated. As new strains of infection arise, such worldwide vaccination will be further delayed and will lead to subsequent lockdowns. As a result, there is an even greater need for the accurate and early identification of COVID-19 to prevent viral spread.
Currently, RT-PCR remains the most utilized detection technique by health professionals, which requires samples to be collected from the upper and lower parts of the sputum and nasal cavity from people suspected of being infected with COVID-19. The RT-PCR test faces several limitations, such as the manual collection of samples from infected patients, reliance on patient consent, and the abilities of doctors, which could also impact the proper identification of coronavirus. Furthermore, the RT-PCR procedure is time-consuming and high cost. Because of these problems, multiple tests are necessary for proper assessment [3]. Chest radiography (CXR) is amongst the most widely-used techniques for diagnosing pneumonia globally [4]. CXR diagnosis offers the advantages of a simple detection procedure and low cost with low ionizing radiation intensity when compared to CT. CXR also exposes the patient to less radiation [5,6]. However, proper diagnosis using CXR images requires professional expertise and skill, thus demanding high-accuracy diagnoses via transfer learning models. Medical imaging professionals are encouraged to use these models to achieve fast and accurate results [7]. Due to the COVID-19 pandemic and the proceeding high demand for health services, the health systems risk collapse owing to the lack of health professionals and hospital beds. Other concerns are that coronavirus is highly infectious, and nurses, caretakers and physicians are the most vulnerable. Therefore, early identification of pneumonia is critical in terms of limiting the dissemination of the pandemic by isolating affected individuals [8].
CT and CXR images have been utilized as the imaging techniques for COVID-19 diagnosis in the majority of studies. In this regard, the presented study includes numerous pieces of research on various transfer learning techniques used to detect coronavirus via CXR imaging.
Nayak et al. [9] suggested a deep learning-based automated approach, named the ResNet-34 model, for the early detection of COVID-19 infection using CXR images. They compared the performance of eight pre-trained CNN models for discriminating COVID-19 instances from normal CXR images and achieved an accuracy of 98.33% for the proposed model. Dilshad et al. [10] introduced and applied a CNN-based MobileNet model on the Indian CXR dataset, achieving an accuracy of 96.33%. Similarly, a CNN-based binary classification model (Haque and Abdelgawad et al. [11]), modified VGG-19 model (Bhatt et al. [12]), and a deep CNN architecture named CVDNet (Ouchicha et al. [13]) have been proposed to detect coronavirus in patients using CXR images, resulting in accuracies of98.3%, 97.37%, and 97.20%, respectively. Other recommended models include two novel models, namely a dilated and depth-wise separable CNN by Li et al. [14], a modified AlexNet model by Kaur et al. [15], a deep CNN architecture named CoroNet based on Xception architecture by Khan et al. [16], an SVM-based hybrid differential evolution PSO model by Dixit et al. [17], and a novel MAnet based two-stage classification technique for coronavirus detection utilizing CXR images by Xu et al. [18]. Moreover, a computer-aided diagnostic combined approach based on graph CNN and pre-trained CNN model (Kumar et al. [19]), a deep Covix-net model (Vinod et al. [20]), new deep hybrid and deep boosted hybrid learning models (Khan et al. [21]), and a gradient weighted class activation mapping technique (Panwar et al. [22]) for coronavirus detection exhibited accuracies of 97.60%, 97%, 98.53%, and 95%, respectively. Likewise, a DenseNet-201 architecture reported by Alhudhaif et al. [23], a PSO-based eXtreme Gradient Boosting model recommended by Dias Júnior et al. [24], an automatic AI-based system using majority voting ensemble techniques suggested by Chandra et al. [25], and a bi-level prediction model by Das et al. [26] exhibited respective accuracies of 94.96%,98.71%, 91.329%, and96.74%to diagnose COVID-19. Other novel techniques to detect coronavirus include a deep LSTM model (Demir et al. [27]), Inception-v3 model based on deep CNN associated with Multi-Layered Perceptron model called CovScanNet (Sait et al. [28]), and a hybrid deep CNN technique with discrete wavelet transform features (Mostafiz et al. [29]. For COVID-19 detection in patients, Mahmud et al. [30] recommended a CovXNet model, and Ergen and Zafer et al. [31] suggested a MobileNetV2 deep CNN model with fuzzy color technique, which achieved an accuracy of 97.4% and 98.25%, respectively. Recently, Dey et al. [32] proposed Choquet fuzzy classifier ensemble technique and Kundu et al. [33] proposed CNN models based on fuzzy rank and Gompertz function to detect coronavirus from CXR and CT images, respectively. Rahaman et al. [34] utilized 15 distinct deep CNN models and Manokaran et al. [35] suggested a DenseNet201 model based on three-class dataset for identification of COVID-19 patients with an accuracy of 89.3% and 94%, respectively. When reviewing the related work, it was discovered that few researchers concentrated on fuzzy ensemble approaches paired with transfer learning models. An ensemble approach is a machine learning technique that integrates many base models to generate a single optimal prediction model. The outcomes of numerous models are integrated in order to enhance overall performance. Specifically, ensemble methods integrate numerous machine learning approaches into a single predictive model to enhance overall performance and predictions while reducing bias and variance. Furthermore, combining deep transfer learning models with fuzzy ensemble approaches will assist in increasing the accuracy and robustness of a detection system. The presented study considered the following.
1. CXR images were used to diagnose coronavirus, and a transfer learning and fuzzy ensemble fusion model was proposed for COVID-19 detection.
2. At first, we trained the VGG16, GoogleNet, and Xception base classifiers using deep transfer learning techniques. Using fully connected layers and the softmax layer, we extracted features and classified CXR images using pre-trained transfer learning models.
3. Then, using Majority Voting, Sugeno Integral, and Choquet fuzzy integral, the data prediction scores from multiple models were integrated to create the final forecasted labels that are more accurate than the separate model predictions.
4. For the application of Choquet fuzzy integral with other state-of-the-art models, such as Majority Voting and Sugeno Integral, we considered the validation accuracy of each classifier to compute the fuzzy membership values of each classifier. The advantage of this type of fusion is that it creates the final prediction of each sample using adaptive weights depending on each sample’s confidence scores.
5. The performance of each pre-trained model, including VGG16, GoogleNet, and Xception combined individually with Majority Voting, Sugeno Integral, and Choquet fuzzy ensemble techniques was measured in terms of precision, recall, F1-Score, sensitivityand specificity.
The goal of this article was to use CXR images to predict coronavirus, aiding healthcare providers and radiologists in the early detection and isolation of positive coronavirus patients to halt the virus’ rapid spread in vulnerable countries with limited hospital capacity and a low patient-to-doctor ratio and to prevent increasing fatality rates.
The remainder of the article is organized as follows. The preliminary models are provided in Section 2. Section 3 describes the proposed algorithm. Section 4 provides a brief explanation of the proposed model and comparison analysis. Section 5 presents the conclusion and future works.

2. Preliminaries

VGG16: VGG16 was proposed by Karen Simonyan et al. [36] in 2014, which was termed based on its 16 layers. VGG-16 contains different layers, such as three dense layers, two FL, five max-pooling layers, and 13 convolutional layers.
GoogleNet: With a reported performance of 93.3%, GoogleNet [37] succeeds at the leading edge of the ImageNet detection and classification problem. The greater utilization of computer resources in the model is its advantage, in which a construction component is treated as an “inception module” that reflects the model’s increased width and depth. This was one of the earliest CNN designs that diverge from the traditional method of just adding convolution and pooling layers in a sequential manner. Figure 1 illustrates the GoogleNet architecture, which is composed of nine initialization modules. GoogleNet is a 22 deep CNN that is a version of Google researchers’ Inception Network that is built on the Inception architecture. In this case, we utilized the third version of the GoogleNet model, Inception V3. This version is made up of a hierarchy of sophisticated inception modules that integrate channel spatial convolution, pooling, and reprojection operations at different scales in each module. The model shrinks the parameter space by breaking spatial convolutions with bigger filter sizes n × n into a series of two convolutional operations with n × 1 and 1 × n filter sizes. The resultant network model is more complicated and deeper than AlexNet, although it has fewer features and lower computing complexity than AlexNet. GoogleNet does not employ fully-connected layers. Instead, the final convolutional map is subjected to channel-wise average activation values, and global average pooling values of the 2048 channels are utilized as the input image’s feature vector.
Xception: The Xception [38] architecture includes a depth-wise separable convolution layer, convolutional layer, residual connections, and inception module. The design of Xception, which stands for “extreme inception”, has 36 convolution layers that form the network’s extractor base. Xception is an extension of Inception architecture by using deeply separable convolutions for standard Inception modules and, thus, has improved speed and accuracy compared to Inception, ImageNet, Inception V3, ResNet-101, ResNet-50, and VGGNet models. The Xception model contains three major sections: entry, middle, and exit flow. The data first enter the input flow, and then go through the eight-fold middle flow, and finally the exit flow. Batch normalization is accompanied by all the Convolution layers. Xception replaced the conventional starting block by expanding it and changing the various (1 × 1, 5 × 5, 3 × 3) spatial dimensions with a (3 × 3) single dimension preceded by a (1 × 1) transition to maintain computational burden. Equation(1) and Equation(2) represent the mathematical modelling of Xception:
f l + 1 k ( p , q ) = x , y f l k ( x ,   y ) .   e l k ( u ,   v )
F l + 2 k = g c ( F l + 2 k ,   k l + 1 )
Choquet fuzzy: In this study, aggregation refers to the process of merging each classifier’s performance scores into a single global score. For the said purpose, an aggregation operator was used. Normally, we consider weighted average as the aggregation operator; however, in this situation, we employed the Choquet fuzzy integral [39]. This approach has already been performed in a variety of pattern identification applications due to its benefit of capturing the unpredictability that exists in the decision values obtained as extra data via classifier fusion, which is lacking in traditional ensemble approaches. As a result, fuzzy measures are the extension of aggregation operations on a set of confidence values. Weighted values are assigned to distinct classifier combinations in these fuzzy measures. To use the Choquet fuzzy approach, we must first establish the fuzzy measure values that define the strength of each classifier. We include three different types of interaction in this section: f   { A   U   B } = f { A } + f { B } = sum of individual strength of A and B   is equal to { A   U   B } , f   { A   U   B } f { A } + f { B } and f   { A   U   B } f { A } + f { B } ; Where A and B are two subsets of classifiers and mutually exclusive. f ( S ) indicates the strength of a set of classifiers.
After obtaining a complete set of fuzzy measure scores and relevant scores for each classifier, we can utilize the Choquet fuzzy method to compute the classifiers’ aggregated global score. Let g   ( )   generate the fuzzy measure score of a collection of classifiers   D = ( d 1 , d 2 , d 3 .. ,   d n ), and   P = ( p 1 , p 2 , p 3 .. ,   p n ) are the individual classifier performance scores in D. Assume L i is a subset of D, such that L i = ( d 1 , d 2 , d 3 .. ,   d n ) of D, where i is in between 1 and n. This indicates that, if i = 1 then L 1 = ( d 1 ), and if i = 3, then L 3 = ( d 1 , d 2 , d 3 ). Wealsoassumethat   (   p 1 p 2 p 3 ..   p n ), then the aggregated score may be computed using the Choquet fuzzy integral, as shown in Equation (3):
C h o q u e t g ( p 1 , p 2 , p 3 .. ,   p n ) = p n * g ( L n ) + ( p n 1 p n ) * g ( L n 1 ) + + ( p 1 p 2 ) * g ( L 1 )
We compute the value of λ for determination of the fuzzy membership values using Equation (4):
1 + λ = i = 1 n ( g ( { d i } ) λ + 1 )
By obtaining the roots of the characteristic equation, we can find the value of λ. By repeating Equation (5), the fuzzy membership value of each classifier combination can be determined:
g ( { d l ,   d o } ) = g ( { d l } ) + g ( { d o } ) + λ g ( { d l } ) g ( { d o } )
where 1 l, o n .
We have p classifiers that are used to categorize q classes when dealing with classifier combinations. Let x i j denote the confidence score of the j th class of the i th classifier. Choquet integral is now utilized for each j to produce a fuzzy confidence score. In this scenario, g ( { b i } ) represents each classifier’s fuzzy membership scores, which are empirically determined for j = 1 ,   2 ,   3 q , b i = x i j , where i = 1 ,   2 ,   3 p . In this study, however, we set the fuzzy measure g ( { b i } ) = w i / w 1 + w 2 + w 3 . + w p , where w i indicates the validation accuracy of the i th classifier. Even after the values of fuzzy measures have been set, the Choquet fuzzy integral can modify the weightage of each classifier by considering the decision scores supplied by the rest of the classifiers. This makes the system dynamic, as opposed to other assembly techniques that employ the majority voting and sugeno integral system.
We must first obtain fuzzy measure values for each classifier before we can use the Choquet fuzzy ensemble. The fuzzy measure values indicate the strength of each classifier and all possible classifiers. Algorithm 1 depicts the key phases of the proposed methodology.
Majority Voting: MV is a basic but effective ensemble learning approach that has been effectively utilized in a variety of industries. The fundamental concept behind MV is to use the majority voting rule to aggregate the output results from many data sources [40]. The majority voting rule is divided into three types: (i) unanimous voting, in which all classifiers must agree on a prediction; (ii) simple majority voting, in which the prediction must be supported by at least one more than half of the classifiers; and (iii) plurality or majority voting, in which the highest number of votes is considered for the ensemble decision whether the sum of those exceeds 50%. In the case of independent classifier outputs, the majority voting rule combiner always improves prediction performance. The MV technique works by accumulating the labels associated with each classifier in the input image. The final label is then assigned to the class with the most labels (votes) among the classifiers. If two classes receive equal votes, we choose to assign the final label to the class with the lowest index.
Sugeno Integral: Takagi-Sugeno [41] is a fuzzy inference technique that creates a systematic strategy for generating fuzzy rules from a given input-output dataset. The inputs are blurry, but the outcome is clear. Takagi-Sugeno computes the crisp output using a weighted average. This method is computationally more efficient and may be used in conjunction with optimization and adaptive approaches.

3. Methodology

Herein, we present a model that employs the deep transfer learning approach to classify CXR images as COVID-positive, negative or pneumonia. This approach first employs pre-trained transfer learning models from VGG16, GoogleNet, and Xception as feature extractors, then uses classifiers to classify the features. The classifier outputs are subsequently merged using Majority Voting, Sugeno Integral, and Choquet fuzzy integral, whereby each classifier’s fuzzy measures are paired with an empirically established ratio. We used two conventional ensemble techniques (Majority Voting and Sugeno Integral) to make a comparison of the Choquet fuzzy ensemble approach for COVID-19 patient identification based on the CXR dataset. We evaluated these models for consistent performance comparison. Figure 2 depicts the proposed ensemble model.
Algorithm 1 Detection of COVID-19 +ve and –ve Patients.
Step 1: As input, use CXR images (COVID-positive, negative or pneumonia images)
Step 2: Resize the size of the training pictures to 224 × 224 × 3 pixels
Step 3: Load pre-trained models like VGG16, GoogleNet and Xception. Let the feature maps for VGG16, GoogleNet and Xception are denoted as FM1, FM2 and FM3 respectively
Step 4: In training phase of VGG16, we use GlobalAveragePooling2D followed by FCL with dropout 0.5, FCL with ReLU 98 and finally FCL with softmax as output layer
Step 5: In training phase of GoogleNet and Xception, we use FCL with dropout 0.5 followed by FCL with ReLU 98 and finally FCL with softmax as output layer
Step 6: Create the classifiers CF1, CF2 and CF3 by layering two FC layers on top of feature maps and a softmax layer.
Step 7: Divide all of the images in the dataset into training and testing sets in proportions of 80% and 20%, respectively.
Step 8: Using the training dataset, train the classifiers CF1, CF2 and CF3
Step 9: Using the validation set, produce validation accuracies for the w 1 , w 2 , w 3 feature maps
Step 10: Determine the fuzzy measure values for the k t h classifier is w i / w 1 + w 2 + w 3
Step 11: Resize the size of the testing pictures to 224 × 224 × 3 pixels.
Step 12: Based on the testing samples, use the learned classifiers to create prediction scores P.
Step 13: Using fuzzy measures F, apply Choquet Fuzzy to the prediction scores P to get the final prediction scores.
Step 14: Use final prediction scores to determine if a particular sample is COVID-19 positive, negative or pneumonia.
2D Global Average Pooling is employed on the spatial dimensions until each spatial dimension is equal to one. To avoid overfitting, FCL with a dropout of 0.5 is utilized. FCL combined with ReLU 98 is then utilized to enhance neural networks by accelerating training. The FCL function with the softmax function is used to normalize the outputs between 0 and 1.

4. Discussion

4.1. Experimental Setup

The entire experiment was carried out on Google Colaboratory using a Python3 engine, a GPU as a backend, and a system RAM of 12 GB. TensorFlow 1.x was employed. For image processing, we loaded the picture as a color image with three RGB channels and transformed it into a NumPy array. On the CXR dataset, we trained all three transfer learning models with an identical set of hyperparameters. During the entry procedure, the images were scaled to 224 × 224 pixels with various dimensions. For training purposes, we setthe learning rate to 2 × 10−4 with 60 epochs, which are small enough to prevent the fitting problem of the transfer learning models. An Adam optimizer was used as a compilation, and FCL with ReLU 98 was employed as the activation function for feature extraction.

4.2. Dataset Description

We obtained the COVID-19 CXR image dataset from Kaggle [42], which includes equal number of COVID positive, pneumonia and normal CXR images. Each class contains 2313 images. The active COVID-19, pneumonia and normal patient CXR images were used to train the Tensor Flow and Keras deep learning models to detect whether or not a person has COVID-19. The data were divided into 80% for the training set and 20% for the test sets, with the identical sets utilized for all models.

4.3. Results and Outcomes

Figure 3 and Figure 4 show the accuracy and loss graphs of the transfer learning models (VGG16, GoogleNet, and Xception). Specifically, Figure 3 indicates that Xception learns faster than GoogleNet and VGG16 and excelled in the competition in terms of accuracy. Early halting and model checkpoints were used to determine the best model and stop training if there was no substantial gain in accuracy. The ImageNet weights for the models were set. In Figure 4, the loss for the above three models is depicted, whereby the loss rate of VGG16 was initially very high but gradually declined, while the loss rate of Xception exhibited a smooth descent. For all three models, the loss was <0.1. We compared the performance of the Choquet fuzzy integral-based ensemble strategy for coronavirus detection to two traditional ensemble approaches (Majority Voting and Sugeno Integral) using the current dataset.

4.4. Performance Analysis

The performance of each model was evaluated in terms of recall, precision, F1-Score, sensitivity, and specificity, and the results are represented in Table 1. Confusion matrices of the VGG16, GoogleNet, and Xception-associated fuzzy ensemble techniques are illustrated in Figure 5, Figure 6 and Figure 7, respectively. Performance comparisons of various transfer learning models with the proposed model using CXR images are presented in Table 2. The following equations were used to calculate the performance metrics:
1 .   P r e c i s i o n = T r u e   P o s i t i v e   ( T P ) T r u e   P o s i t i v e + F a l s e   P o s i t i v e   ( F P )
2 .   S e n s i t i v i t y   o r   R e c a l l = T r u e   P o s i t i v e T r u e   P o s i t i v e + F a l s e   N e g a t i v e   ( F N )
3 .   F 1 S c o r e = 2   x ( P r e c i s i o n   x   R e c a l l ) ( P r e c i s i o n + R e c a l l )
4 .   S p e c i f i c i t y = T r u e   N e g a t i v e   ( T N ) T r u e   N e g a t i v e + F a l s e   P o s i t i v e
5 .   A c c u r a c y = T P + T N T P + T N + F P + F N

5. Conclusions and Future Work

With the growing threat of new coronavirus strains throughout the world, early diagnosis of COVID-19 is critical especially due to the global scarcity of healthcare facilities. In this work, we propose a hybrid fuzzy ensemble-based method for detecting COVID-19-positive patients using CXR images. Ensemble learning can improve the overall performance by combining the prominent properties of its constituent models, resulting in better predictions than the individual contributing models. Moreover, ensemble models can enhance performance by lowering the variance of prediction errors by adding some bias to compete with base learners. In the proposed work, we used various transfer learning models, which were separately trained to make independent predictions. The models were then integrated to forecast a class value, utilizing a novel approach combining Majority Voting, Sugeno Integral, and Choquet Fuzzy. A publicly accessible database containing 6939 CXR pictures was employed to assess the effectiveness of the proposed method. The data were divided into training and test sets in proportions of 80% and 20%, respectively. The proposed method, Xception + Choquet Fuzzy ensemble methods, obtained a better accuracy of 99.57% compared to other transfer learning models. The suggested methodology failed to categorize 0.43% of test data, which may be due to the relatively poor performance of the individual transfer learning models. As a result, creating or selecting better individual transfer learning models may help to amend this issue. In addition, only three basic transfer learning models were applied in the aforementioned ensemble method. Therefore, adding more classifier techniques might be an additional option to improve the current method. To demonstrate its robustness across various domains, the proposed model might be applied to various classification-based image analyses. Such advancements will aid radiologists to develop a graphical user interface (GUI) for recognizing COVID-19 to keep up with the disease’s various stages of progression and imagistic patterns.

Author Contributions

Conceptualization, R.K. and C.M.; methodology, P.G.A. and A.H.G. software, R.K.; validation, C.M. and A.H.G.; formal analysis, P.G.A.; investigation, A.H.G.; resources, C.M.; data curation, R.K.; writing—original draft preparation, A.H.G.; writing—review and editing, A.H.G.; visualization, P.G.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, D.; Lee, J.-Y.; Yang, J.-S.; Kim, J.W.; Kim, V.N.; Chang, H. The Architecture of SARS-CoV-2 Transcriptome. Cell 2020, 181, 914–921.e10. [Google Scholar] [CrossRef]
  2. Silva, P.C.; Batista, P.V.; Lima, H.S.; Alves, M.A.; Guimarães, F.G.; Silva, R.C. COVID-ABS: An agent-based model of COVID-19 epidemic to simulate health and economic effects of social distancing interventions. Chaos Solitons Fractals 2020, 139, 110088. [Google Scholar] [CrossRef]
  3. Li, Y.; Yao, L.; Li, J.; Chen, L.; Song, Y.; Cai, Z.; Yang, C. Stability issues of RT-PCR testing of SARS-CoV-2 for hospitalized patients clinically diagnosed with COVID-19. J. Med.Virol. 2020, 92, 903–908. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Jaiswal, A.K.; Tiwari, P.; Kumar, S.; Gupta, D.; Khanna, A.; Rodrigues, J.J. Identifying pneumonia in chest X-rays: A deep learning approach. Measurement 2019, 145, 511–518. [Google Scholar] [CrossRef]
  5. Gaál, G.; Maga, B.; Lukács, A. Attention u-net based adversarial architectures for chest x-ray lung segmentation. arXiv 2020, arXiv:2003.10304. [Google Scholar]
  6. Smith-Bindman, R.; Lipson, J.; Marcus, R.; Kim, K.P.; Mahesh, M.; Gould, R.; De González, A.B.; Miglioretti, D.L. Radiation dose associated with common computed tomography examinations and the associated lifetime attributable risk of cancer. Arch. Intern. Med. 2009, 169, 2078–2086. [Google Scholar] [CrossRef]
  7. Minaee, S.; Kafieh, R.; Sonka, M.; Yazdani, S.; Soufi, G.J. Deep-COVID: Predicting COVID-19 from chest X-ray images using deep transfer learning. Med. Image Anal. 2020, 65, 101794. [Google Scholar] [CrossRef]
  8. Sen-Crowe, B.; Sutherland, M.; McKenney, M.; Elkbuli, A. A closer look into global hospital beds capacity and resource shortages during the COVID-19 pandemic. J. Surg. Res. 2021, 260, 56–63. [Google Scholar] [CrossRef]
  9. Nayak, S.R.; Nayak, D.R.; Sinha, U.; Arora, V.; Pachori, R.B. Application of deep learning techniques for detection of COVID-19 cases using chest X-ray images: A comprehensive study. Biomed. Signal Process. Control. 2021, 64, 102365. [Google Scholar] [CrossRef]
  10. Dilshad, S.; Singh, N.; Atif, M.; Hanif, A.; Yaqub, N.; Farooq, W.; Ahmad, H.; Chu, Y.-M.; Masood, M.T. Automated image classification of chest X-rays of COVID-19 using deep transfer learning. Results Phys. 2021, 28, 104529. [Google Scholar] [CrossRef]
  11. Haque, K.F.; Abdelgawad, A. A deep learning approach to detect COVID-19 patients from chest X-ray images. AI 2020, 1, 418–435. [Google Scholar] [CrossRef]
  12. Bhatt, A.; Ganatra, A.; Kotecha, K. COVID-19 pulmonary consolidations detection in chest X-ray using progressive resizing and transfer learning techniques. Heliyon 2021, 7, e07211. [Google Scholar] [CrossRef] [PubMed]
  13. Ouchicha, C.; Ammor, O.; Meknassi, M. CVDNet: A novel deep learning architecture for detection of coronavirus (COVID-19) from chest x-ray images. Chaos Solitons Fractals 2020, 140, 110245. [Google Scholar] [CrossRef]
  14. Li, X.; Zhai, M.; Sun, J. DDCNNC: Dilated and depthwise separable convolutional neural Network for diagnosis COVID-19 via chest X-ray images. Int. J. Cogn. Comput. Eng. 2021, 2, 71–82. [Google Scholar] [CrossRef]
  15. Kaur, M.; Kumar, V.; Yadav, V.; Singh, D.; Kumar, N.; Das, N.N. Metaheuristic-based Deep COVID-19 screening model from chest X-ray images. J. Health Eng. 2021, 2021, 1–9. [Google Scholar] [CrossRef]
  16. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Comput. Methods Programs Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  17. Dixit, A.; Mani, A.; Bansal, R. CoV2-Detect-Net: Design of COVID-19 prediction model based on hybrid DE-PSO with SVM using chest X-ray images. Inf. Sci. 2021, 571, 676–692. [Google Scholar] [CrossRef] [PubMed]
  18. Xu, Y.; Lam, H.-K.; Jia, G. MANet: A two-stage deep learning method for classification of COVID-19 from Chest X-ray images. Neurocomputing 2021, 443, 96–105. [Google Scholar] [CrossRef]
  19. Kumar, A.; Tripathi, A.R.; Satapathy, S.C.; Zhang, Y.D. SARS-Net: COVID-19 Detection from Chest X-Rays by Combining Graph Convolutional Network and Convolutional Neural Network. Pattern Recognit. 2020, 122, 108255. [Google Scholar] [CrossRef] [PubMed]
  20. Vinod, D.N.; Jeyavadhanam, B.R.; Zungeru, A.M.; Prabaharan, S. Fully automated unified prognosis of Covid-19 chest X-ray/CT scan images using Deep Covix-Net model. Comput. Biol. Med. 2021, 136, 104729. [Google Scholar] [CrossRef]
  21. Khan, S.H.; Sohail, A.; Khan, A.; Hassan, M.; Lee, Y.S.; Alam, J.; Basit, A.; Zubair, S. COVID-19 detection in chest X-ray images using deep boosted hybrid learning. Comput. Biol. Med. 2021, 137, 104816. [Google Scholar] [CrossRef]
  22. Panwar, H.; Gupta, P.; Siddiqui, M.K.; Morales-Menendez, R.; Bhardwaj, P.; Singh, V. A deep learning and grad-CAM based color visualization approach for fast detection of COVID-19 cases using chest X-ray and CT-Scan images. Chaos Solitons Fractals 2020, 140, 110190. [Google Scholar] [CrossRef]
  23. Alhudhaif, A.; Polat, K.; Karaman, O. Determination of COVID-19 pneumonia based on generalized convolutional neural network model from chest X-ray images. Expert Syst. Appl. 2021, 180, 115141. [Google Scholar] [CrossRef]
  24. Júnior, D.A.D.; da Cruz, L.B.; Diniz, J.O.B.; da Silva, G.L.F.; Junior, G.B.; Silva, A.C.; de Paiva, A.C.; Nunes, R.A.; Gattass, M. Automatic method for classifying COVID-19 patients based on chest X-ray images, using deep features and PSO-optimized XGBoost. Expert Syst. Appl. 2021, 183, 115452. [Google Scholar] [CrossRef]
  25. Chandra, T.B.; Verma, K.; Singh, B.K.; Jain, D.; Netam, S.S. Coronavirus disease (COVID-19) detection in Chest X-Ray images using majority voting based classifier ensemble. Expert Syst. Appl. 2021, 165, 113909. [Google Scholar] [CrossRef] [PubMed]
  26. Das, S.; Roy, S.D.; Malakar, S.; Velásquez, J.D.; Sarkar, R. Bi-Level prediction model for screening COVID-19 patients using chest X-ray images. Big Data Res. 2021, 25, 100233. [Google Scholar] [CrossRef]
  27. Demir, F. DeepCoroNet: A deep LSTM approach for automated detection of COVID-19 cases from chest X-ray images. Appl. Soft Comput. 2021, 103, 107160. [Google Scholar] [CrossRef] [PubMed]
  28. Sait, U.; Lal K.V., G.; Shivakumar, S.; Kumar, T.; Bhaumik, R.; Prajapati, S.; Bhalla, K.; Chakrapani, A. A deep-learning based multimodal system for Covid-19 diagnosis using breathing sounds and chest X-ray images. Appl. Soft Comput. 2021, 109, 107522. [Google Scholar] [CrossRef]
  29. Mostafiz, R.; Uddin, M.S.; Alam, N.-A.; Reza, M.; Rahman, M.M. Covid-19 detection in chest X-ray through random forest classifier using a hybridization of deep CNN and DWT optimized features. J. King Saud Univ. Comput. Inf. Sci. 2020. [Google Scholar] [CrossRef]
  30. Mahmud, T.; Rahman, A.; Fattah, S.A. CovXNet: A multi-dilation convolutional neural network for automatic COVID-19 and other pneumonia detection from chest X-ray images with transferable multi-receptive feature optimization. Comput. Biol. Med. 2020, 122, 103869. [Google Scholar] [CrossRef] [PubMed]
  31. Toğaçar, M.; Ergen, B.; Cömert, Z. COVID-19 detection using deep learning models to exploit Social Mimic Optimization and structured chest X-ray images using fuzzy color and stacking approaches. Comput. Biol. Med. 2020, 121, 103805. [Google Scholar] [CrossRef]
  32. Dey, S.; Bhattacharya, R.; Malakar, S.; Mirjalili, S.; Sarkar, R. Choquet Fuzzy Integral-based Classifier Ensemble Technique for COVID-19 Detection. Comput. Biol. Med. 2021, 135, 104585. [Google Scholar] [CrossRef]
  33. Kundu, R.; Basak, H.; Singh, P.K.; Ahmadian, A.; Ferrara, M.; Sarkar, R. (). Fuzzy rank-based fusion of CNN models using Gompertz function for screening COVID-19 CT-scans. Sci. Rep. 2021, 11, 14133. [Google Scholar] [CrossRef] [PubMed]
  34. Rahaman, M.M.; Li, C.; Yao, Y.; Kulwa, F.; Rahman, M.A.; Wang, Q.; Qi, S.; Kong, F.; Zhu, X.; Zhao, X. Identification of COVID-19 samples from chest X-Ray images using deep learning: A comparison of transfer learning approaches. J. X-ray Sci. Technol. 2020, 28, 821–839. [Google Scholar] [CrossRef]
  35. Manokaran, J.; Zabihollahy, F.; Hamilton-Wright, A.; Ukwatta, E. Detection of COVID-19 from chest x-ray images using transfer learning. J. Med. Imaging 2021, 8, 017503. [Google Scholar] [CrossRef] [PubMed]
  36. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  37. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  38. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1800–1807. [Google Scholar] [CrossRef] [Green Version]
  39. Murofushi, T.; Sugeno, M. An interpretation of fuzzy measures and the Choquet integral as an integral with respect to a fuzzy measure. Fuzzy Sets Syst. 1989, 29, 201–227. [Google Scholar] [CrossRef]
  40. Kittler, J.; Hatef, M.; Duin, R.P.W.; Matas, J. On combining classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 226–239. [Google Scholar] [CrossRef] [Green Version]
  41. Sugeno, M. An introductory survey of fuzzy control. Inf. Sci. 1985, 36, 59–83. [Google Scholar] [CrossRef]
  42. COVID19_Pneumonia_Normal_Chest_Xray_PA_Dataset. 2021. Available online: https://www.kaggle.com/amanullahasraf/covid19-pneumonia-normal-chest-xray-pa-dataset (accessed on 24 August 2021).
  43. Sethy, P.K.; Behera, S.K.; Ratha, P.K.; Biswas, P. Detection of coronavirus disease (COVID-19) based on deep features. Int. J. Math. Eng. Manag. Sci. 2020, 5, 643–651. [Google Scholar] [CrossRef]
  44. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. arXiv 2014, arXiv:2003.10849. [Google Scholar] [CrossRef]
  45. Civit-Masot, J.; Luna-Perejón, F.; Morales, M.D.; Civit, A. Deep learning system for COVID-19 diagnosis aid using X-ray pulmonary images. Appl. Sci. 2020, 10, 4640. [Google Scholar] [CrossRef]
  46. Elaziz, M.A.; Hosny, K.M.; Salah, A.; Darwish, M.M.; Lu, S.; Sahlol, A.T. New machine learning method for image-based diagnosis of COVID-19. PLoS ONE 2020, 15, e0235187. [Google Scholar] [CrossRef]
  47. Ucar, F.; Korkmaz, D. COVIDiagnosis-Net: Deep Bayes-SqueezeNet based diagnosis of the coronavirus disease 2019 (COVID-19) from X-ray images. Med. Hypotheses 2020, 140, 109761. [Google Scholar] [CrossRef] [PubMed]
  48. Martinez, F.; Martínez, F.; Jacinto, E. Performance evaluation of the NASNet convolutional network in the automatic identification of COVID-19. Int. J. Adv. Sci. Eng. Inf. Technol. 2020, 10, 662–667. [Google Scholar] [CrossRef]
  49. Vaid, S.; Kalantar, R.; Bhandari, M. Deep learning COVID-19 detection bias: Accuracy through artificial intelligence. Int. Orthop. 2020, 44, 1539–1542. [Google Scholar] [CrossRef] [PubMed]
  50. Karakanis, S.; Leontidis, G. Lightweight deep learning models for detecting COVID-19 from chest X-ray images. Comput. Biol. Med. 2021, 130, 104181. [Google Scholar] [CrossRef] [PubMed]
  51. Elpeltagy, M.; Sallam, H. Automatic prediction of COVID− 19 from chest images using modified ResNet50. Multimed. Tools Appl. 2021, 4, 1–13. [Google Scholar] [CrossRef]
  52. Iraji, M.S.; Feizi-Derakhshi, M.-R.; Tanha, J. COVID-19 detection using deep convolutional neural networks and binary differential algorithm-based feature selection from X-ray images. Complexity 2021, 2021, 1–10. [Google Scholar] [CrossRef]
  53. Apostolopoulos, I.D.; Mpesiana, T.A. Covid-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [Green Version]
Figure 1. GoogleNet architecture.
Figure 1. GoogleNet architecture.
Applsci 11 11423 g001
Figure 2. Working principle of VGG16, GoogleNet, and Xception pre-trained models for detection of positive, negative and pneumonia COVID-19 patients.
Figure 2. Working principle of VGG16, GoogleNet, and Xception pre-trained models for detection of positive, negative and pneumonia COVID-19 patients.
Applsci 11 11423 g002
Figure 3. Accuracy graph for the training phase of VGG16, Xception and GoogleNet.
Figure 3. Accuracy graph for the training phase of VGG16, Xception and GoogleNet.
Applsci 11 11423 g003
Figure 4. Loss graph for the training phase of VGG16, Xception, and GoogleNet.
Figure 4. Loss graph for the training phase of VGG16, Xception, and GoogleNet.
Applsci 11 11423 g004
Figure 5. Confusion matrices of VGG16 and associated fuzzy ensemble techniques. VGG16, VGG16 + Majority Voting, VGG16 + Sugeno Integral, and VGG16 + Choquet Fuzzy achieved an accuracy of 96.68%, 96.75%, 96.90%, and 97.04%, respectively.
Figure 5. Confusion matrices of VGG16 and associated fuzzy ensemble techniques. VGG16, VGG16 + Majority Voting, VGG16 + Sugeno Integral, and VGG16 + Choquet Fuzzy achieved an accuracy of 96.68%, 96.75%, 96.90%, and 97.04%, respectively.
Applsci 11 11423 g005
Figure 6. Confusion matrices of GoogleNet and associated fuzzy ensemble techniques. GoogleNet, GoogleNet + Majority Voting, GoogleNet + Sugeno Integral, and GoogleNet + Choquet Fuzzy achieved an accuracy of 97.91%, 98.05%, 98.27%, and 98.48%, respectively.
Figure 6. Confusion matrices of GoogleNet and associated fuzzy ensemble techniques. GoogleNet, GoogleNet + Majority Voting, GoogleNet + Sugeno Integral, and GoogleNet + Choquet Fuzzy achieved an accuracy of 97.91%, 98.05%, 98.27%, and 98.48%, respectively.
Applsci 11 11423 g006
Figure 7. Confusion matrices of Xception and associated fuzzy ensemble techniques. Xception, Xception + Majority Voting, Xception + Sugeno Integral, and Xception + Choquet Fuzzy achieved an accuracy of 98.20%, 98.56%, 99.28%, and 99.57%, respectively.
Figure 7. Confusion matrices of Xception and associated fuzzy ensemble techniques. Xception, Xception + Majority Voting, Xception + Sugeno Integral, and Xception + Choquet Fuzzy achieved an accuracy of 98.20%, 98.56%, 99.28%, and 99.57%, respectively.
Applsci 11 11423 g007
Table 1. Performance metrics for various transfer learning models associated with fuzzy ensemble techniques.
Table 1. Performance metrics for various transfer learning models associated with fuzzy ensemble techniques.
ModelsClassPrecision (%)Sensitivity (%)Specificity (%)F1-Score (%)Accuracy (%)
VGG16COVID +ve0.9700.9780.9850.97496.68
COVID -ve0.9630.9570.9820.960
Pneumonia 0.9670.9650.9840.966
VGG16 + Majority VotingCOVID +ve0.9700.9810.9850.97596.75
COVID -ve0.9630.9570.9820.960
Pneumonia 0.9700.9650.9850.967
VGG16 + Sugeno IntegralCOVID +ve0.9720.9740.9860.97396.90
COVID -ve0.9690.9590.9850.964
Pneumonia 0.9660.9740.9830.970
VGG16 + Choquet FuzzyCOVID +ve0.9700.9780.9850.97497.04
COVID -ve0.9710.9570.9860.964
Pneumonia 0.9700.9760.9850.973
GoogleNetCOVID +ve0.9780.9810.9890.97997.91
COVID -ve0.9780.9740.9890.976
Pneumonia 0.9810.9830.9900.982
GoogleNet + Majority VotingCOVID +ve0.9780.9850.9890.98298.05
COVID -ve0.9800.9760.9900.978
Pneumonia 0.9830.9810.9910.982
GoogleNet + Sugeno IntegralCOVID +ve0.9810.9830.9900.98298.27
COVID -ve0.9870.9780.9940.983
Pneumonia 0.9810.9870.9900.984
GoogleNet + Choquet FuzzyCOVID +ve0.9830.9870.9910.98598.48
COVID -ve0.9890.9780.9950.984
Pneumonia 0.9830.9890.9910.986
XceptionCOVID +ve0.9830.9830.9910.98398.20
COVID -ve0.9870.9780.9940.983
Pneumonia 0.9760.9850.9880.981
Xception + Majority VotingCOVID +ve0.9850.9850.9920.98598.56
COVID -ve0.9850.9850.9920.985
Pneumonia 0.9870.9870.9940.987
Xception + Sugeno IntegralCOVID +ve0.9910.9960.9960.99499.28
COVID -ve0.9930.9890.9970.991
Pneumonia 0.9940.9940.9970.994
Proposed ModelCOVID +ve0.9960.9980.9980.99799.57
COVID -ve0.9980.9940.9990.996
Pneumonia 0.9940.9960.9970.995
Table 2. Performance comparisons of various transfer learning models with the suggested model using CXR images.
Table 2. Performance comparisons of various transfer learning models with the suggested model using CXR images.
AuthorsBest MethodologyAccuracy (%)
Sethy et al. [43]ResNet50+ SVM95.38%
Narin et al. [44]Deep CNN; ResNet-5098%
Civit et al. [45]K-Nearest Neighbour98.09%
Elaziz et al. [46]Convolutional Neural Network93.2%
Mahmud et al. [30]Neural Search Architecture Network-Type CNN97%
Ucar et al. [47]VGG1996.3%
Martínez et al. [48]SqueezeNet97%
Vaid et al. [49]VGG1695%
Khan et al. [21]Deep Boosted Hybrid Learning98.53%
Karakanis et al. [50]Binary-Classification with CGANs98.7%
Elpeltagy et al. [51]CNN with pre−trained ResNet5097.1%
Júnior et al. [24]VGG19 and XGBoost + PSO98.71%
Rahaman et al. [34]15 distinct deep CNN models89.3%
Manokaran et al. [35]DenseNet201 model94%
Iraji et al. [52]Binary differential-CNN99.43%
Apostolopoulos et al. [53] MobileNetV2 94.72%
Proposed ModelXception + Choquet Fuzzy99.57%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mahanty, C.; Kumar, R.; Asteris, P.G.; Gandomi, A.H. COVID-19 Patient Detection Based on Fusion of Transfer Learning and Fuzzy Ensemble Models Using CXR Images. Appl. Sci. 2021, 11, 11423. https://doi.org/10.3390/app112311423

AMA Style

Mahanty C, Kumar R, Asteris PG, Gandomi AH. COVID-19 Patient Detection Based on Fusion of Transfer Learning and Fuzzy Ensemble Models Using CXR Images. Applied Sciences. 2021; 11(23):11423. https://doi.org/10.3390/app112311423

Chicago/Turabian Style

Mahanty, Chandrakanta, Raghvendra Kumar, Panagiotis G. Asteris, and Amir H. Gandomi. 2021. "COVID-19 Patient Detection Based on Fusion of Transfer Learning and Fuzzy Ensemble Models Using CXR Images" Applied Sciences 11, no. 23: 11423. https://doi.org/10.3390/app112311423

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop