Next Article in Journal
Rapid and Reproducible MALDI-TOF-Based Method for the Detection of Vancomycin-Resistant Enterococcus faecium Using Classifying Algorithms
Previous Article in Journal
Rapid Evolution of an Aortic Endocarditis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Channel Based Image Processing Scheme for Pneumonia Identification

by
Grace Ugochi Nneji
1,
Jingye Cai
1,*,
Jianhua Deng
1,
Happy Nkanta Monday
2,
Edidiong Christopher James
1 and
Chiagoziem Chima Ukwuoma
1
1
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(2), 325; https://doi.org/10.3390/diagnostics12020325
Submission received: 13 December 2021 / Revised: 14 January 2022 / Accepted: 17 January 2022 / Published: 27 January 2022
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
Pneumonia is a prevalent severe respiratory infection that affects the distal and alveoli airways. Across the globe, it is a serious public health issue that has caused high mortality rate of children below five years old and the aged citizens who must have had previous chronic-related ailment. Pneumonia can be caused by a wide range of microorganisms, including virus, fungus, bacteria, which varies greatly across the globe. The spread of the ailment has gained computer-aided diagnosis (CAD) attention. This paper presents a multi-channel-based image processing scheme to automatically extract features and identify pneumonia from chest X-ray images. The proposed approach intends to address the problem of low quality and identify pneumonia in CXR images. Three channels of CXR images, namely, the Local Binary Pattern (LBP), Contrast Enhanced Canny Edge Detection (CECED), and Contrast Limited Adaptive Histogram Equalization (CLAHE) CXR images are processed by deep neural networks. CXR-related features of LBP images are extracted using shallow CNN, features of the CLAHE CXR images are extracted by pre-trained inception-V3, whereas the features of CECED CXR images are extracted using pre-trained MobileNet-V3. The final feature weights of the three channels are concatenated and softmax classification is utilized to determine the final identification result. The proposed network can accurately classify pneumonia according to the experimental result. The proposed method tested on publicly available dataset reports accuracy of 98.3%, sensitivity of 98.9%, and specificity of 99.2%. Compared with the single models and the state-of-the-art models, our proposed network achieves comparable performance.

1. Introduction

Pneumonia is an infectious lung illness in humans that affects one or both lungs and is caused by fungus, bacteria, and viruses, among other microorganisms. Pneumonia occurs as a result of pathogen-caused inflammation [1], which causes the alveoli in the lungs to fill up with pus or fluid, reducing oxygen (O 2 ) and carbon-dioxide (CO 2 ) exchange between the lungs and blood, making it difficult for the infected person to breathe. Other causes of pneumonia are food aspiration and chemical exposure. Furthermore, patients with cancer, HIV/AIDS, hepatic disease, diabetes, cardiovascular diseases, chronic respiratory diseases, and other comorbidities, are vulnerable to pneumonia [1,2,3]. As the result of the inability to identify this lung illness at an early stage, children below the age of five years and aged people are readily transmitted with this illness.
There are several methods to diagnose pneumonia, which include blood test, pulse oximetry, bronchoscopy, sputum test, pleural fluid culture, chest X-ray, magnetic resonance imaging, and CT scans. The blood tests are carried out to confirm an illness and to attempt to figure out the organism caused by it. However, this method is not always feasible to make an accurate identification. Pulse oximetry test determines how much oxygen is in the blood. Pneumonia may make it difficult for the lungs to get enough oxygen into the circulatory system. Another method of pneumonia diagnose is the sputum test which demands for a patient’s deep cough, then the sample of fluid from the lungs (sputum) is obtained and tested to assist determine the source of the illness. A culture method called pleural fluid culture requires a collection of a fluid sample from the pleural region by inserting a needle between the patient’s ribs and testing it to help diagnose the kind of infection. Chest X-ray is used to diagnose pneumonia and evaluate the degree and location of the illness. However, it cannot tell which bacterium caused the pneumonia. Finally, for a clearer view of the lungs, CT scans are highly recommended by the medical practitioners. For older citizens above 65 age of years with serious symptoms and health condition they are advised to undergo imaging examination such as CT scans [4,5]. In most cases, the causing organism determines how pneumonia is treated. Anti-fungal treatments are used to treat fungus pneumonia whereas antiviral medications are used to treat viral pneumonia including influenza, SARS, and MERS, and antibiotics are used to treat bacterial pneumonia [6]. Despite the presence of pneumonia, the diagnosis is always dependent on the doctor’s knowledge and experience. Howbeit, as the number of infected patients’ increases, it becomes a difficult task for radiologists to access the screening process within a constrained time; therefore, there is a need for an AI diagnostic system for the classification of pneumonia.
Deep learning (DL) algorithms have ascertained greater performance in the classification and detection of pneumonia-related ailment and provided high accuracy rate when compared to other methods. Emphatically, DL can identify hidden features of images which could never be detected by medical professionals. With regards to DL, the convolutional neural network (CNN) is the most often utilized DL technique in the medical system due to its capacity in feature extraction and training in differentiating between multiple classes. Transfer learning (TL) technique has made it simpler to retrain deep neural networks fast and accurately on different datasets [7,8,9]. With the introduction of wavelet in CNN, Happy et al. [10] proposed a CNN model integrated with wavelet multi-resolution analysis to effectively classify COVID-19 pneumonia from radiographs. Interestingly, Happy et al. [11] presented a case study to evaluate the capability of multi-resolution analysis in diagnosing COVID-19 pneumonia.
Several applications of computer vision are so complicated that they cannot be achieved using only one algorithm, which has necessitated the creation of models that combined two or more of the methods evaluated. Since models are chosen based on the issue’s requirements and features extraction, weighted fusion technique incorporates more than one model to address this issue. This technique was created to help single models overcome their flaws and solidify their strengths by aggregating the features extracted from single model in a weighted manner. This method reduces generalization error and minimizes prediction variance [12]. Thus, the goal of this study is to analyze the performance of weight fusing deep neural network models for multi-class of pneumonia. The following are the main contributions of our paper:
  • We preprocess our CXR images into three forms, namely: LBP, CLAHE, and CECED.
  • Each preprocessed image features are extracted individually, that is the LBP CXR features are extracted using the shallow CNN, the CECED CXR features are extracted using the MobileNet-V3 and the CLAHE CXR feature images are extracted using the Inception v3.
  • The feature vectors of these preprocessed CXR images are weighted fused for a robust prediction result.
  • We evaluate the performance of each of the single models and our proposed model in this study.
The remaining part of this paper is broken down as follows; the related works on pneumonia disease are discussed in Section 2. Section 3 explains how we came up with our proposed framework. The experimental results and evaluation of our model are presented in Section 4. Our discussion is presented in Section 5. Finally, Section 6 presents the conclusion.

2. Related Works

In recent years, publications of the some of the state-of-the-art models used the deep learning (DL) approach for the automatic classification and detection of pneumonia from X-ray images. This section reviews and examines the most up-to-date approaches for detecting and classifying pneumonia with DL.
Various investigations and research studies based on artificial intelligence and deep learning have been conducted in the area of disease diagnosis using medical imagery such as chest X-rays (CXR), ultrasound scans, computed tomography (CT) scans, MRI scans, and so on. Deep learning is perhaps the most widely used and reliable medical imaging approach available. It is a quick and accurate way for diagnosing variety of ailments. There are models that have been explicitly trained to classify different categories of specific ailment based on the disease type. These models have proven to achieve satisfactory results in the medical science utilizing image analysis for the detection of cardiac abnormalities, tumors, cancer, and a variety of different applications.
Deep learning has been used to distinguish between scan images of COVID-19-infected and non-infected patients as proposed by Shah et al. [13] using a self-developed framework called CTnet-10, hence, achieving an accuracy of 82.1%. However, to further enhance the accuracy, different pretrained models were introduced to the CT scan dataset and VGG-19 attain a better performance of 94.5%.
Deep neural network has the ability to improve the output of instruments utilized in pharmaceutical processes for analytical technology. Maruthamuthu et al. [14] developed a technique for identifying the contamination of micro-organisms using DL strategies based on Raman spectroscopy. The categorization of numerous sample consisting of micro-organisms and bacteria mixed with Chinese Hamster Ovary cells achieved an accuracy of 95% to 100% using a convolutional neural network (CNN).
Looking at a situation of very large margin of unlabeled dataset, Maruthamuthu et al. [15], demonstrate the superiority of a data-driven strategy over data analytics and machine learning technique. The data analytics strategy, however, has the potential to improve the development of efficient proteins and affinity ligands, affecting the downstream refinement and processing of mAb products, according to the authors. The authors propose a deep neural network approach, which is also called a black box for the construction of soft sensors, which uses accumulated past data from process operation that is not present in a mechanistic orphenomenological approach. The data are incorporated to get apolynomial expression to generate product as a function of the inputs such as cell mass, substrate concentration, and initial product (streptokinase). Given enough dataset collected from various sources under different condition, the accurate data fit with automated error correction generated factor is utilized to compute the major outputs such as the number of workable cells and extent of product as a result of the marginal calibrated input parameters such as inoculum and substrate. Even though the molecular relationship of these variables to cell metabolism is uncertain, yet the deep learning model accounts for the impacts of other conditions such as excitation, individual critical media properties, mixing rate, and variations in metabolic concentrations.
An application of the Inception-v3 model is used by Kermany et al. [16] to classify different types of pneumonia infections in pediatric patients. This method extracts features and utilizes the area under curve (AUC) as a metric to categorize bacterial and viral pneumonia. Santosh et al. [17] suggested a method for detecting pulmonary problems by comparing the symmetrical shape of the CXR lung region. Roux et al. [18] sought to determine the prevalence, intensity, and adverse outcomes for pneumonia in infants during their first year of life in a South African cohort study. They set up pneumonia monitoring systems and recorded outpatient pneumonia and pneumonia that required hospitalization. In addition, combined Poisson regression was used to calculate pneumonia incidence rate ratios.
Table 1 gives a summary of the related works with the following observations:
  • Several DL techniques are thoroughly implemented for the classification task.
  • More so, researchers in [27,28,31,32] use more than five DL techniques to evaluate the performance.
  • The popular medical mode of imaging for the classification and detection of pneumonia-related ailment is the chest X-ray.
  • The publishers focus on either binary classification or multi-classification, although just a few considered the multi-classification task.
  • Some evaluation metrics such as accuracy, sensitivity, and specificity were utilized to evaluate the efficacy of DL models approaches whereas precision, recall, and/or F1-score metrics were used in [25,26] to evaluate the model performance.
However, there was no account where low quality was considered as a challenge for better classification performance of pneumonia. Therefore, we put this into consideration by preprocessing our images into LBP, CECED, and CLAHE for multi-channel weighted fusion technique using neural network for the identification of pneumonia. Additionally, we carried out a study to compare the performance of each single model and the weighted fusion model.

3. Materials and Methods

The methodology employed in this study is detailed in this section. Data collection, image preprocessing, and the transfer learning networks for feature extraction and classification are the three phases of the proposed method. The procedure of the proposed method in this paper is presented in the subsequent subsections.

3.1. Datasets

This study utilized the pneumonia dataset from two public domain. The first dataset contains 3029 scans of bacterial pneumonia, 8851 scans of healthy patients and 2983 scans of viral pneumonia gotten from the Kaggle database of RSNA [33]. The second dataset is taken from the author in [34] which consist of 3616 CXR images of patients diagnosed with COVID-19. In all, we only considered 1000 CXR images each for the four classes (COVID-19, virus, bacteria, and normal), as explicitly illustrated in Table 2.

3.2. Data Preprocessing

It is well known that data quality can be affected by factors such as noise, resolution and artifacts, hence, using such data directly in an algorithm may result to inaccurate outcomes. Data preprocessing phase could help in eliminating or reducing the noise and increase the data quality, thereby improving model performance. In this study, we have use the Local Binary Pattern (LBP), Contrast-Enhanced Canny Edge Detection (CECED), and Contrast Limited Adaptive Histogram Equalization (CLAHE) for our data preprocessing.

3.2.1. Local Binary Pattern (LBP) Images

A descriptor called LBP is often used to collect textural information about a specific object. Coding LBP pixel is determined by comparing its value to that of neighboring pixels [13]. At the left portion of Figure 1 depicts all the values of the pixel in a local region, whereas the right part depicts the binary pixel of the LBP coding while concentrating on the left pixel. The pixel value of LBP can be computed after it has been efficiently encoded using LBP coding. Equation (1) gives the mathematical expression of LBP.
L B P = n = 1 N S ( g n g c ) 2 n
where S ( g n g c ) denotes the indicator function for the threshold binary value obtained from the subtraction of the neighbor pixel value from center pixel value. N is the total number of pixel value surrounding the center pixel value for the 3 × 3 window size since LBP considers 9 pixels at a time. g c represents the center pixel value and g n denotes the neighbor pixels values. After the thresholding, the pixel’s value greater or equal to zero would be “1” and the pixel’s value less than zero would be “0”. Figure 1 illustrates the calculation of the pneumonia LBP CXR image and Figure 2 presents the preprocessing of LBP.

3.2.2. CLAHE Images

CLAHE was used in this study to improve the image’s contrast and features by making abnormalities more noticeable. Among the histogram equalization family, Contrast Limited Adaptive Histogram Equalization (CLAHE) is more natural in appearance and useful in the reduction of noise amplification, and we have investigated CLAHE and applied it to our dataset, as shown in Figure 3. A full explanation of the CLAHE approach is given below to demonstrate its effectiveness:
  • The generation of the image transformation using the bin value of the histogram is the first stage of the CLAHE technique.
  • Following that, using clip boundary, the contrast is confined to a binary count from 0 to 1. Before the image segment is processed, the clip boundary is added to the image.
  • To prevent mapping background areas to gray scale, a specific bin value from the histogram region is used to create the entire image region. Clip boundary is used with the help of histogram clip to obtain better mapping.
  • Finally, the finished CLAHE image is created by computing the image’s regions, then extracting, mapping, and interpolating all of the image pixels to get the most out of the image.

3.2.3. CECED Images

As shown in Figure 4, edges are made up of crucial and relevant particular information and features. By employing an edge detector on an image, the amount of data that has to be processed can be reduced, and the information that is deemed less significant can be filtered. The CEED-Canny strategy combines the local morphological contrast enhancement and the Canny edge detection technology used in [24]. The steps are as follows:
  • Collection of the original pixel’s value, as well as the local minimum and maximum;
  • The image’s morphological contrast is increased;
  • To reduce noise, Gaussian smoothing is applied;
  • The intensity gradient of the image is determined;
  • A non-maximum suppression method is utilized;
  • The hysteresis thresholding technique is adopted.

3.3. Feature Extraction

This article uses three CNN architectures: a shallow CNN, pretrained MobileNet-V3, and Inception-V3 as feature extractors with their layers trained and/or frozen. To prevent the model from overfitting, dropout, and regularizer are applied.

3.3.1. Features Extraction from LBP Images

For an automatic extraction of pneumonia-related features from LBP CXR images, we develop a shallow CNN model. Shallow CNN model simply means a lightweight convolutional neural network constructed from scratch to process LBP CXR images. In our study, we constructed a three-layer convolutional neural network which we called shallow CNN, as illustrated in Figure 5. It has an input layer, three convolution (“C”) and subsequent sub-sampling (“S”) layers, and a feature vector (“fv”) layer, as shown in Figure 5. The first convolution layer (“C1”) employs 32 filters and 7 × 7 convolution kernel that focus on the detailed information of possible indicators of pneumonia. This layer is followed by a sub-sampling layer (“S1”), which reduces the image to half its original size using optional max-pooling (with kernel size 2 × 2 ).
To map the previous layer, a new convolution layer (“C2”) conducts 64 convolutions with 3 × 3 kernel, followed by another sub-sampling layer (“S2”) with 2 × 2 kernel. Lastly, 128 convolution layer (“C3”) with 3 × 3 kernel is applied and followed by a max-pooling (with kernel size 2 × 2 ). Finally, we arrived at a feature vector of 512-neuron after the output is passed to the fully connected layer (“fv”). Adding “ReLU” activation after sub-sampling layers “S1” and “S2” guarantees the capacity to handle nonlinear data. Over-fitting can be avoided by utilizing the “dropout” operation [31] between the “S” layers (“S1” and “S2”) and the “fv” layer (parameter was set to 0.5).

3.3.2. Features Extraction from CECED CXR Images

We utilized the MobileNet-V3 network in this extraction phase because of its efficient performance in image classification and fast convergence. Table 3 depicts the parameters of the network. The model has an input size of 224 × 224 input image. It uses depth-wise separable convolutions, which applies two 1D convolutions with two kernels, rather than a single 2D convolution. This allows for a smaller and more efficient model with less memory and training parameters. The model is divided into two blocks: the residual block with a stride of 1 and the shrinking block with a stride of 2. Each block is made up of three layers: a 1 × 1 convolution with ReLU6, a depth-wise convolution, and another 1 × 1 convolution with different non-linearity. A few changes were made to our network to simplify it by replacing the dense layer with a dimension of 1 × 512 , as shown in Figure 6.

3.3.3. Features Extraction from CLAHE CXR Images

The high performance of Inception-V3 [25] is due to a variety of network connection approaches, including batch normalization, using MLP convolutional layers to replace linear convolutional layers, and factorizing convolutions with larger kernel sizes. These methods significantly reduce the number of network parameters as well as the computational cost of the network, allowing it to be built considerably deeper and with more nonlinear expressive capacity than typical CNN models.
The Inception-V3 model proposed in this paper was previously trained over the huge dataset ImageNet, and then the knowledge gained from transfer learning was transferred to CXR dataset to perform pneumonia identification. All layers before the fully connected (FC) and softmax layers were frozen. The softmax layer is eliminated and a new dense layer is trained for extracting the deep-level features of the CXR images by continuously modifying the network’s parameters, resulting in a feature vector size of 1 × 512 as seen in Figure 7. Similar to the authors in [25], average pooling of 8 × 8 is employed instead of the traditional fully connected layer to flatten the feature vector as seen in Figure 7.

3.4. Weighted Fusion of the Different Output Channels

According to the proposed model shown in Figure 8, shallow CNN is used to extract the feature vector (fv1) from the LBP CXR images, the CLAHE CXR images’ feature vector (fv2) is retrieved using a pretrained Inception-V3 approach and the pretrained MobileNet-V3 is utilized to extract feature vector (fv3) from CECED CXR images. After flattening, each feature vector is connected to two dense layers. The first dense layers for fv1, fv2, and fv3 are fc1_1, fc2_1, and fc3_1, respectively, while the second dense layers are fc1_2, fc2_2, and fc3_2, respectively, as shown in Figure 8. Using fc1_2 and fc2_2 and fc3_2, the network captures the distances between various lung properties and reveals them. Furthermore, fc1_2, fc2_2, and fc3_2 are fused in a weighted way into f1 to generate a fused vector. Softmax is used to classify pneumonia CXR images based on the fused feature vector. As a cost function, the categorical cross entropy is utilized, given in Equation (2)
L l o s s = k = 1 n t k l o g p k
where n represents the number of class labels, p k is the instances for the k-th classes and t k is the corresponding label The chance of event k occurring is t k meaning that the total sum of t k is 1, which implies that only one event is possible. The negative sign minimizes the loss when the distributions come closer to one other.
In the last layer of our proposed weighted model with categorical cross entropy loss function, the softmax function is utilized as a classifier to emphasize the highest values and suppress those that are far below the maximum value as well as normalizing the outputs to the total sum of 1, as illustrated in Equation (3)
p k = e t k j = 1 n e j t
where e t k is the exponential term of the input vector, n is the number of class label, p k is the output vector of the softmax and j = 1 n e t j is the normalization term which divides the exponential term to produce a real value greater than 0.

4. Results

This section gives thorough explanations of (1) the experimental setup and configuration (2) the performance metrics utilized for the evaluation of the proposed framework; (3) the performance of individual networks, (4) the performance of weighted fusion technique, and (5) comparisons with other research models.

4.1. Experimental Setup and Configuration

Our proposed model is evaluated using the Keras framework with Python programming language using NVIDIA GTX 1080 GPU. A data split ratio of 80% to 20% for training and testing, respectively. All the data are resized for each single model. For the shallow CNN, our LBP input dataset are resized to 112 × 112 , whereas for the pretrained MobileNet-V3 network, the CECED CXR input dataset are resized to 224 × 224 and lastly for the pre-trained inception-V3 network, the CLAHE CXR input dataset are resized as 299 × 299 . Using Adam optimization, we utilized batch size of 32, learning rate of 0.0001, L2-regularizer, weight decay, dropout to reduce over-fitting in the model and epoch size of 40. We trained separately each of the the single model (shallow CNN, pre-trained MobileNet-V3 and pre-trained Inception-V3) and the concatenation of these single models using the fusion technique with the same dataset. More so, the output classes of 1000 in ImageNet were replaced with 4 classes in the final dense layer, representing viral, bacteria, normal, and COVID-19.

4.2. Performance Metrics

In evaluating the performance of the single and concatenated models, this paper used the evaluation metrics of accuracy, sensitivity, specificity, precision, and F1-score. The numerical expression for each metric is presented in Equations (4)–(8).
Accuracy = TP + TN TP + TN + FP + FN
Sensitivity = TP TP + FN
Specificity = TN TN + FP
Precision = TP TP + FP
F 1 - score = 2 Precision Recall Precision + Recall
TP, FP, FN, and TN represent true positive, false positive, false negative, and true negative, respectively.

4.3. Evaluation of the Single Model

We begin by presenting the accuracy, loss, ROC-AUC, and precision–recall curves produced by the various deep learning frameworks used in this article (Inception-V3, MobileNet-V3, and Shallow CNN). Then, using the metrics established in Equations (4)–(8), we evaluated the results of all designs to find the best method for classifying CXR images as bacteria, COVID-19, normal, and viral pneumonia as shown in Table 4. The ROC-AUC, precision–recall, accuracy, and loss curves, as well as the several models utilized in this study and their interpretation, are shown in the next section.

4.3.1. Shallow CNN

It is worth noting that the shallow CNN achieved considerable performance. The results from Table 4 show that the shallow CNN obtained 90.9% accuracy, 92.3% sensitivity, and 93.1% specificity. In comparison to Inception-V3 and MobileNet-V3, shallow CNN achieved the least score across all the evaluation metrics. Furthermore, the shallow CNN obtained 91.2% precision and 92.7% F1-score.

4.3.2. Pretrained MobileNet-V3

Considering the evaluation performance metrics for the single model of MobileNet-V3, this model obtained accuracy of 93.7%, sensitivity of 95.4%, specificity of 95.7%, precision of 94.3%, and F1-score of 95.5%. These results further show that the CECED preprocessing technique using the MobileNet-V3 performs better than the LBP shallow CNN technique.

4.3.3. Pretrained Inception-V3

From the experimental results presented in Table 4, Inception-V3 shows satisfactory performance achieving accuracy of 95.6%, 94.9% sensitivity, and 96.2% specificity, precision of 95.8%, and F1-score of 95.3%. Compared to shallow CNN, this inception-V3 seems to perform slightly higher than shallow CNN across all the evaluation metrics, however, Inception-v3 achieves better performance on few evaluation metrics such as accuracy, specificity, and precision over the MobileNet-V3.

4.4. Evaluation of the Weighted Fusion

On the pneumonia dataset, we executed various weighted fusion approaches to see the better performance for the identification of pneumonia. Our first investigation is the weighted fusion of LBP-channel shallow CNN and CECED-channel of MobileNet-V3 (LCSC + CCM). Secondly, we checkmate the fusion of the LBP-channel of shallow CNN and the CLAHE-channel of Inception-V3 (LCSC + CCI). The next analysis is the fusion of CECED-channel of MobileNet-V3, and CLAHE-channel of Inception-V3 (CCM + CCI)) and the last weighted-manner approach is the fusion of LBP-channel of shallow CNN, CLAHE-channel of Inception-V3 and CECED-channel of MobileNet-V3 (LCSC + CCI + CCM) for pneumonia identification. The proposed framework of the weighted fusion, as shown in Figure 8, clearly shows that the complimentary fusion of LBP, CLAHE, and CECED image features is capable of managing low-quality images in pneumonia diagnosis, attaining improved recognition accuracy on the CXR dataset. Figure 9 shows classification performance of all the weighted fusion channels including the single channels across all the evaluation metrics. It is obvious that our proposed model outweighs the other weighted fusion channels, as well as the single channels by a considerable margin.Although the computational time is slightly longer than the other weighted channels and the single models, yet it achieves the best classification performance, as depicted in Table 4.

4.5. Result of the Proposed Model

Figure 10 presents the test accuracy curves for the proposed three-channel weighted fusion including the single- and dual-channel models. For the first 10 epochs, the accuracy rapidly increases to about 90% for the proposed model and then maintained gradual progression. Figure 11 depicts the loss curves which show the gradual loss reduction of the different models. Figure 12 presents the ROC-AUC curves for the single- and dual-channels, as well as the proposed weighted fusion deep learning schemes. The LCSC represents the LBP-based channel of shallow CNN for the identification of pneumonia with 92.1% AUC. The LCSC+CCM represents the weighted fusion of LBP-based channel of shallow CNN and the CECED-based channel of MobileNet-V3 with 93.8%. The CCM represents the CECED-based channel of MobileNet-V3 with 95.0% AUC. The LCSC+CCI represents the weighted fusion of LBP-based channel of shallow CNN and the CLAHE-based channel of Inception-V3 with 94.5% AUC. The CCI represents the CLAHE-based channel of Inception-V3 with 97.2% AUC. The CCI+CCM represents the weighted fusion of CLAHE-based channel of Inception-V3 and the CECED-based channel of MobileNet-V3 with 98.3%. The LCSC+CCI+CCM represents the weighted fusion of LBP-based channel of shallow CNN, the CLAHE-based channel of Inception-V3, and the CECED-based channel of MobileNet-V3 with 99.0% AUC.
The proposed multi-channel weighted fusion model (LCSC + CCI + CCM) outperforms the single-based and dual-based channels on the pneumonia dataset, obtaining 99.0% AUC. The AUC of CECED-based channel of MobileNet-V3 is clearly higher than that of CLAHE-based channel of Inception-V3 and Shallow CNN of LBP-based channel, implying that CECED CXR images contribute more to pneumonia identification than CLAHE and LBP CXR images. The proposed weighted fusion model was further analyzed in terms of precision–recall curve, as shown in Figure 13. As the curve approaches the upper right hand corner of the graph, it is obvious that the proposed multi-channel weighted fusion model outweighs the other schemes, indicating that the model has high precision and high recall.

5. Discussion

In comparison to the single-based and dual-based channels, the proposed model’s performance in diagnosing pneumonia in CXR images from dataset obtained from various sources and combined has been demonstrated, and the identification result is given in Table 4. The proposed model can effectively differentiate distinct pneumonia from healthy CXR images, as evidenced by the above mentioned results. It is worth noting that the weighted fusion of LBP, CLAHE, and CECED CXR images result in a greater generalization ability for our proposed model.
We use the following terms to describe the channels as Shallow CNN for channel that employs LBP CXR images, inception-V3 for channel that uses CLAHE CXR images, and MobileNet-V3 for a channel that utilizes CECED CXR images. Table 4 shows the results of the identification for the different types of pneumonia. On all metrics, our proposed model surpasses the single-based and dual-based channels, with 98.3% accuracy, 98.9% sensitivity, and 99.2% specificity, 98.8% precision and 99% f1-score.
We conducted a comparison between our proposed strategy and some recently published pneumonia classification methods. According to Table 5, the proposed model obtained the maximum accuracy and sensitivity score of 98.3% and 98.9%, respectively, while Correa et al. [22] achieve the highest specificity score of 100% although their accuracy value was not reported. In terms of accuracy, the proposed model achieves the highest score of 98.3%, demonstrating its superiority in the identification of pneumonia. The complementing integration of several channels of CXR images gives our proposed technique a competitive advantage. It is worth noting that different deep learning models will perform differently depending on different circumstances. More so, we selected few recent state-of-the-art models and conducted a fair comparison using the same dataset as depicted in Table 6. Figure 14 presents the performance evaluation of the selected few state-of-the-art models across different matrices using the same dataset while Figure 15 depicts the accuracy performance of the selected few state-of-the-art models using the same dataset. We conducted an ablation study utilizing several transfer learning models pretrained on the ImageNet dataset in order to find the best performing pretrained model for our proposed weighted deep learning fusion architecture.
The results of the experiments employing the proposed scheme with several pretrained frameworks utilizing the same pneumonia dataset are presented in Table 7. According to the results of the experiments, Shallow CNN outperforms MobileNet-V3 and Inception-V3 in extracting features from LBP CXR images, reaching 90.9% accuracy, 92.3% sensitivity, and 93.1% specificity. In comparison to Shallow CNN, the MobileNet-V3 network performs better in extracting features from CECED CXR images sachieving 93.7% accuracy, 95.4% sensitivity, and 95.7% specificity. When compared to Shallow CNN and MobileNet-V3, Inception-V3 performs significantly better in extracting features from CLAHE CXR images, reaching reaching 95.6% accuracy, 94.9% sensitivity, and 96.2% specificity. It is essential to use the ROC curve to estimate overall accuracy and the precision–recall curve to measure the mean average precision of the proposed model when identifying sensitive conditions like pneumonia. The ROC curves for single- and dual-based channels, as well as the proposed model on the pneumonia dataset, are shown in Figure 12. Similarly, Figure 13 shows the precision–recall curve for single- and dual-based channels, as well as the proposed model.
Furthermore, many of the CXR images were blurry and lacking in details, which may have hampered the proposed model’s ability to extract and train relevant features. Interestingly, the benefits of utilizing LBP, CLAHE, and CECED preprocessing approaches to improve the low quality of CXR images, identify high representation details of the CXR images with visible trainable features. The proposed multi-channel weighted fusion performed admirably in identifying various kinds of pneumonia.
In terms of ROC and precision–recall curves, the proposed multi-channel weighted fusion model appears to outperform the other single and dual-based channels, particularly when dealing with low-quality CXR images. Our proposed model’s curve is closest to the upper left corner of the graph and has the largest area under curve, indicating that it has higher sensitivity and specificity according to the ROC graph in Figure 12. Similarly, our proposed model’s curve is closest to the graph’s upper right corner with the largest area, meaning that it has higher precision and sensitivity in Figure 13. More importantly, as explained above, the stated result in terms of Receiver Operating Characteristic (ROC) and precision–recall can aid expert radiologist in striking a balance between accuracy and precision.
Nevertheless, other methods have shown better performance for example, PneumoniaNet model [41]. PneumoniaNet model proposed by Alsharif et al. [41] uses CXR images to distinguish normal radiographic images from those with features consistent with viral or bacterial pneumonia in the pediatric group aged one–five years with a 99.72% accuracy, 99.74% sensitivity, 99.85% specificity, 99.7% precision, and 98.12% AUC. It achieved satisfactory performance because it combines the extracted feature of three subsequent convolution layers separated by ReLU and the batch normalization layer, whereas the general feature extraction uses only one convolution and batch normalization layer, hence permitting the CNN model to utilize the overall and marginal difference in CXR images.
The authors of PneumoniaNet seeks to optimize the information flow and gradient within the network, hence making the optimization of deep learning network tractable. This model improves the feature propagation, encourages feature re-useability and reduction in parameters. The distinguishing of this model and ResNet50 is that the latter learns from the reference input layers by using the output of the existing layer as the subsequent input layer as compared to the former, whereas comparing with VGG model, the PneumoniaNet model uses a deeper structure of small receptive subsequent 3 × 3 filters.
However, the authors did not consider the low quality of CXR images which is a major concern in real-life application. It is well known that data quality can be affected by factors such as noise, resolution, and artifacts, hence, using such data directly in an algorithm may affect model adaptability in medical application. A data preprocessing phase could help in eliminating or reducing the noise and increase the data quality, thereby improving model performance. The novelty of our methodology takes full advantage of the complementing integration of feature maps from multi-channels of data preprocessing technique in a weighted fusion manner, thereby enhancing the overall performance of our model.

5.1. Ablation Study

5.1.1. Hyperparameter Tuning

We conducted another ablation study to examine if hyperparameter tuning may improve the performance of our proposed multi-channel model. The results of different optimizers at different learning rates and dropouts are shown in Table 8, Table 9 and Table 10. Our proposed multi-channel weighted fusion model obtains the best performance utilizing Adam optimizer with a learning rate of 0.0001 and dropout of 0.50 , attaining 98.3% accuracy, according to Table 8. It is worth mentioning that the Adam optimizer is substantially more robust than the other optimizers (RMSProp and SGD) due to its computational efficiency.

5.1.2. Raw X-ray Image Feature

Further experiment is conducted to examine and validate the contribution of LBP, CECED and CLAHE features to model performance over raw X-ray image feature. Figure 16 shows the accuracy results using the raw chest X-ray images while Figure 17 depicts the performance evaluation of the models across different matrices using the raw chest X-ray images. Table 11 shows the performance of the various pre-trained models and shallow CNN using the preprocessed images including the raw X-ray image. To further validate the contribution of LBP, CECED, and CLAHE image features to the overall performance the proposed ensemble model, we conducted another experiment with the same selected pretrained models using the raw chest X-ray images, as depicted in Table 12. From all indication, the ensemble of LBP, CECED, and CLAHE yielded much better results compared to the ensemble of the raw chest X-ray image.

6. Conclusions

In this study, we proposed pneumonia identification method based on weighted fusion capable of processing LBP, CLAHE, and CECED CXR images simultaneously. We mentioned that the three channels are fused to complementary capture meaningful details from CXR images and achieved higher identification accuracy. The strategy of weighted fusion is utilized to take complete advantage of the visual features that have been captured from the different channels. The proposed weighted fusion model handles the problem of low-quality CXR images by fusing the weighted features generated from LBP, CLAHE, and CECED preprocessing stages. The shallow CNN, MobileNet-V3, and Inception-V3 models are employed to extract both healthy and non-healthy (COVID-19, virus, and bacteria pneumonia) features from LBP, CLAHE, and CECED CXR images.
Furthermore, these features are merged by utilizing the weighted fusion strategy in order to take advantage of the complementary lungs information. Softmax was introduced as the classifier to obtain the fused features. By fusing channels of complementary attributes in a weighted manner, our proposed model outweighs several state-of-the-art methods. The evaluation results show that the proposed model achieves better performance with accuracy of 98.3%, sensitivity of 98.9%, specificity of 99.2%, precision of 98.8% and f1-score of 99.0% than just using the single- and dual-channels. From the comparative results of the other established methods, it is confirmed that the proposed weighted fusion model achieved state-of-the-art identification accuracy, which makes it robust and efficient identification solution for pneumonia based on low quality CXR images. These results could effectively assist the expert radiologist to diagnose what type of pneumonia disease is present in a patient’s lungs while saving screening time.

Author Contributions

G.U.N.: Conceptualization, Methodology, Resources, Data curation, Formal analysis, Writing—original draft. J.C., J.D.: Funding acquisition, Project administration, Supervision. H.N.M.: Investigation, Software, Validation, Visualization, Writing—review and editing. E.C.J., C.C.U.: Software, Validation, Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study, as this study only makes use of publicly available data.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The dataset used in this study can be found in references [33,34].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ibrahim, U.; Ozsoz, M.; Serte, S.; Al-Turjman, F.; Yakoi, P.S. Pneumonia classification using deep learning from chest X-ray images during COVID-19. Cognit. Comput. 2021, 1–13. [Google Scholar] [CrossRef] [PubMed]
  2. Luján-Garc, J.E.; Yáñez-Márquez, C.; Villuendas-Rey, Y.; Camacho-Nieto, O. A transfer learning method for pneumonia classification and visualization. Appl. Sci. 2020, 10, 2908. [Google Scholar] [CrossRef] [Green Version]
  3. Stephen, O.; Sain, M.; Maduh, U.J.; Jeong, D.-U. An efficient deep learning approach to pneumonia classification in healthcare. J. Healthc. Eng. 2019, 2019, 4180949. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Acharya, A.K.; Satapathy, R. A deep learning based approach towards the automatic diagnosis of pneumonia from chest radio-graphs. Biomed. Pharmacol. J. 2020, 13, 449–455. [Google Scholar]
  5. Shi, F.; Xia, L.; Shan, F.; Song, B.; Wu, D.; Wei, Y.; Yuan, H.; Jiang, H.; He, Y.; Gao, Y.; et al. Large-scale screening to distinguish between COVID-19 and community-acquired pneumonia using infection size-aware classification. Phys. Med. Biol. 2021, 66, 65031. [Google Scholar] [CrossRef] [PubMed]
  6. El Asnaoui, K. Design ensemble deep learning model for pneumonia disease classification. Int. J. Multimed. Inf. Retr. 2021, 10, 55–68. [Google Scholar] [CrossRef] [PubMed]
  7. Nneji, G.U.; Cai, J.; Deng, J.; Monday, H.N.; Nahar, S.; Mgbejime, G.T.; James, E.C.; Woldeyes, S.K. A Dual Weighted Shared Capsule Network for Diabetic Retinopathy Fundus Classification. In Proceedings of the 2021 International Conference on High Performance Big Data and Intelligent Systems, Macau, China, 5–7 December 2021; pp. 297–302. [Google Scholar] [CrossRef]
  8. Nneji, G.U.; Cai, J.; Jianhua, D.; Monday, H.N.; Chikwendu, I.A.; Oluwasanmi, A.; James, E.C.; Mgbejime, G.T. Enhancing Low Quality in Radiograph Datasets Using Wavelet Transform Convolutional Neural Network and Generative Adversarial Network for COVID-19 Identification. In Proceedings of the 2021 the 4th International Conference on Pattern Recognition and Artificial Intelligence, Yibin, China, 20–22 August 2021; pp. 146–151. [Google Scholar] [CrossRef]
  9. Nneji, G.U.; Cai, J.; Jianhua, D.; Monday, H.N.; Ejiyi, C.J.; James, E.C.; Mgbejime, G.T.; Oluwasanmi, A. A Super-Resolution Generative Adversarial Network with Siamese CNN Based on Low Quality for Breast Cancer Identification. In Proceedings of the 2021 4th International Conference on Pattern Recognition and Artificial Intelligence, Yibin, China, 20–22 August 2021; pp. 218–223. [Google Scholar] [CrossRef]
  10. Monday, H.N.; Li, J.P.; Nneji, G.U.; Oluwasanmi, A.; Mgbejime, G.T.; Ejiyi, C.J.; Chikwendu, I.A.; James, E.C. Improved Convolutional Neural Multi-Resolution Wavelet Network for COVID-19 Pneumonia Classification. In Proceedings of the 2021 the 4th International Conference on Pattern Recognition and Artificial Intelligence, Yibin, China, 20–22 August 2021; pp. 267–273. [Google Scholar] [CrossRef]
  11. Monday, H.N.; Li, J.P.; Nneji, G.U.; James, E.C.; Chikwendu, I.A.; Ejiyi, C.J.; Oluwasanmi, A.; Mgbejime, G.T. The Capability of Multi Resolution Analysis: A Case Study of COVID-19 Diagnosis. In Proceedings of the 2021 the 4th International Conference on Pattern Recognition and Artificial Intelligence, Yibin, China, 20–22 August 2021; pp. 236–242. [Google Scholar] [CrossRef]
  12. Zilly, J.; Buhmann, J.M.; Mahapatra, D. Glaucoma detection using entropy sampling and ensemble learning for automatic optic cup and disc segmentation. Comput. Med. Imaging Graph. 2017, 55, 28–41. [Google Scholar] [CrossRef]
  13. Shah, V.; Keniya, R.; Shridharani, A.; Punjabi, M.; Shah, J.; Mehendale, N. Diagnosis of COVID-19 using CT scan images and deep learning techniques. Emerg. Radiol. 2021, 28, 497–505. [Google Scholar] [CrossRef]
  14. Maruthamuthu, M.K.; Raffiee, A.H.; De Oliveira, D.M.; Ardekani, A.M.; Verma, M.S. Raman spectra-based deep learning: A tool to identify microbial contamination. MicrobiologyOpen 2020, 9, e1122. [Google Scholar] [CrossRef]
  15. Maruthamuthu, M.K.; Rudge, S.R.; Ardekani, A.M.; Ladisch, M.R.; Verma, M.S. Process analytical technologies and data analytics for the manufacture of monoclonal antibodies. Trends Biotechnol. 2020, 38, 1169–1186. [Google Scholar] [CrossRef]
  16. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.S.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying Medical Diagnoses and Treatable Diseases by Image- Based Deep Learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef] [PubMed]
  17. Santosh, C.; Antani, S. Automated chest X-ray screening: Can lung region symmetry help detect pulmonary abnormalities. IEEE Trans. Med. Imaging 2018, 37, 1168–1177. [Google Scholar] [CrossRef] [PubMed]
  18. Le Roux, M.; Myer, L.; Nicol, M.P. Incidence and severity of childhood pneumonia in the first year of life in a South African birth cohort. The Drakenstein Child Health Study. Lancet Glob. Health 2015, 3, e95–e103. [Google Scholar] [CrossRef] [Green Version]
  19. Cicero, M.; Bilbily, A.; Colak, E.; Dowdell, T.; Gray, B.; Perampaladas, K.; Barfett, J. Training and validating a deep convolutional neural network for computer-aided detection and classification of abnormalities on frontal chest radiographs. Investig. Radiol. 2017, 52, 281–287. [Google Scholar] [CrossRef]
  20. Guendel, S.; Grbic, S.; Georgescu, B.; Liu, S.; Maier, A.; Comaniciu, D. Learning to Recognize Abnormalities in Chest X-rays with Location-Aware Dense Networks. In Iberoamerican Congress on Pattern Recognition; Springer: Cham, Switzerland, 2018; pp. 757–765. [Google Scholar]
  21. Rajaraman, S.; Candemir, S.; Kim, I.; Thoma, G.; Antani, S. Visualization and interpretation of convolutional neural network predictions in detecting pneumonia in pediatric chest radiographs. Appl. Sci. 2018, 8, 1715. [Google Scholar] [CrossRef] [Green Version]
  22. Correa, M.; Zimic, M.; Barrientos, F.; Barrientos, R.; Román-Gonzalez, A.; Pajuelo, M.J.; Anticona, C.; Mayta, H.; Alva, A.; Solis-Vasquez, L.; et al. Automatic classification of pediatric pneumonia based on lung ultrasound pattern recognition. PLoS ONE 2018, 13, e0206410. [Google Scholar] [CrossRef]
  23. Ke, Q.; Zhang, J.; Wei, W.; Połap, D.; Woźniak, M.; Kośmider, L.; Damaševĭcius, R. A neuro-heuristic approach for recognition of lung diseases from X-ray images. Expert. Syst. Appl. 2019, 126, 218–232. [Google Scholar] [CrossRef]
  24. Saraiva, A.A.; Ferreira, N.M.F.; de Sousa, L.L.; Costa, N.J.C.; Sousa, J.V.M.; Santos, D.B.S.; Valente, A.; Soares, S. Classification of Images of Childhood Pneumonia Using Convolutional Neural Networks; BIOIMAGING: Geneva, Switzerland, 2019; pp. 112–119. [Google Scholar]
  25. Sirazitdinov, I.; Kholiavchenko, M.; Mustafaev, T.; Yixuan, Y.; Kuleev, R.; Ibragimov, B. Deep neural network ensemble for pneumonia localization from a large-scale chest X-ray database. Comput. Electr. Eng. 2019, 78, 388–399. [Google Scholar] [CrossRef]
  26. Liang, G.; Zheng, L. A transfer learning method with deep residual network for pediatric pneumonia diagnosis. Comput. Methods Programs Biomed. 2020, 187, 104964. [Google Scholar] [CrossRef]
  27. Apostolopoulos, I.D.; Mpesiana, T.A. COVID-19: Automatic detection from X-ray images utilizing transfer learning with convolutional neural networks. Phys. Eng. Sci. Med. 2020, 43, 635–640. [Google Scholar] [CrossRef] [Green Version]
  28. Xua, X.; Jiang, X.; Ma, C.; Du, P.; Li, X.; Lv, S.; Yu, L.; Ni, Q.; Chen, Y.; Su, J.; et al. A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering 2020, 6, 1122–1129. [Google Scholar] [CrossRef] [PubMed]
  29. Habib, N.; Hasan, M.M.; Reza, M.M.; Rahman, M.M. Ensemble of CheXNet and VGG-19 feature extractor with random forest classifier for pediatric pneumonia detection. SN Comput. Sci. 2020, 1, 1–9. [Google Scholar] [CrossRef] [PubMed]
  30. Chouhan, V.; Singh, S.K.; Khamparia, A.; Gupta, D.; Tiwari, P.; Moreira, C.; Damaševičius, R.; de Albuquerque, V.H.C. A novel transfer learning based approach for pneumonia detection in chest X-ray images. Appl. Sci. 2020, 10, 559. [Google Scholar] [CrossRef] [Green Version]
  31. El Asnaoui, K.; Chawki, Y.; Idri, A. Automated Methods for Detection and Classification Pneumonia Based on X-ray Images Using Deep Learning. In Artificial Intelligence and Blockchain for Future Cybersecurity Applications; Springer: Berlin/Heidelberg, Germany, 2021; pp. 257–284. [Google Scholar]
  32. El Asnaoui, K.; Chawki, Y. Using X-ray images and deep learning for automated detection of coronavirus disease. J. Biomol. Struct. Dyn. 2020, 39, 3615–3626. [Google Scholar] [CrossRef] [PubMed]
  33. RSNA Pneumonia Detection Challenge | Kaggle [Internet]. Available online: https://www.kaggle.com/c/rsna-pneumonia-detection-challenge/data (accessed on 1 May 2021).
  34. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Kashem, S.B.A.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S.; et al. Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef] [PubMed]
  35. Yamaç, M.; Ahishali, M.; Degerli, A.; Kiranyaz, S.; Chowdhury, M.E.H.; Gabbouj, M. Convolutional Sparse Support Estimator-Based COVID-19 Recognition From X-Ray Images. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 1810–1820. [Google Scholar] [CrossRef] [PubMed]
  36. Wang, J.; Bao, Y.; Wen, Y.; Lu, H.; Luo, H.; Xiang, Y.; Li, X.; Liu, C.; Qian, D. Prior-Attention Residual Learning for More Discriminative COVID-19 Screening in CT Images. IEEE Trans. Med. Imaging 2020, 39, 2572–2583. [Google Scholar] [CrossRef] [PubMed]
  37. Li, J.; Wang, Y.; Wang, S.; Wang, J.; Liu, J.; Jin, Q.; Sun, L. Multiscale Attention Guided Network for COVID-19 Diagnosis Using Chest X-Ray Images. IEEE J. Biomed. Health Inform. 2021, 25, 1336–1346. [Google Scholar] [CrossRef]
  38. Singh, K.K.; Singh, A. Diagnosis of COVID-19 from chest X-ray images using wavelets-based depthwise convolution network. Big Data Min. Anal. 2021, 4, 84–93. [Google Scholar] [CrossRef]
  39. Yang, Z.; Zhao, L.; Wu, S.; Chen, C.Y.-C. Lung Lesion Localization of COVID-19 From Chest CT Image: A Novel Weakly Supervised Learning Method. IEEE J. Biomed. Health Inform. 2021, 25, 1864–1872. [Google Scholar] [CrossRef]
  40. Wang, X.; Jiang, L.; Li, L.; Xu, M.; Deng, X.; Dai, L.; Xu, X.; Li, T.; Guo, Y.; Wang, Z.; et al. Joint Learning of 3D Lesion Segmentation and Classification for Explainable COVID-19 Diagnosis. IEEE Trans. Med Imaging 2021, 40, 2463–2476. [Google Scholar] [CrossRef]
  41. Alsharif, R.; Al-Issa, Y.; Alqudah, A.M.; Qasmieh, I.A.; Mustafa, W.A.; Alquran, H. PneumoniaNet: Automated Detection and Classification of Pediatric Pneumonia Using Chest X-ray Images and CNN Approach. Electronics 2021, 10, 2949. [Google Scholar] [CrossRef]
  42. Alqudah, A.M.; Qazan, S.; Masad, I.S. Artificial Intelligence Framework for Efficient Detection and Classification of Pneumonia Using Chest Radiography Images. J. Med. Biol. Eng. 2021, 41, 599–609. [Google Scholar] [CrossRef]
  43. Alquran, H.; Alsleti, M.; Alsharif, R.; Abu Qasmieh, I.; Alqudah, A.; Binti Harun, N.H. Employing Texture Features of Chest X-Ray Images and Machine Learning in COVID-19 Detection and Classification. Mendel 2021, 27, 9–17. [Google Scholar] [CrossRef]
  44. Masad, I.S.; Alqudah, A.; Alqudah, A.M.; Almashaqbeh, S. A hybrid deep learning approach towards building an intelligent system for pneumonia detection in chest X-ray images. Int. J. Electr. Comput. Eng. 2021, 11, 5530–5540. [Google Scholar] [CrossRef]
Figure 1. LBP coding and calculation illustration.
Figure 1. LBP coding and calculation illustration.
Diagnostics 12 00325 g001
Figure 2. LBP pre-processing.
Figure 2. LBP pre-processing.
Diagnostics 12 00325 g002
Figure 3. CLAHE Pre-processing.
Figure 3. CLAHE Pre-processing.
Diagnostics 12 00325 g003
Figure 4. CECED Pre-processing.
Figure 4. CECED Pre-processing.
Diagnostics 12 00325 g004
Figure 5. Shallow CNN structure applied for the feature extraction of LBP CXR images.
Figure 5. Shallow CNN structure applied for the feature extraction of LBP CXR images.
Diagnostics 12 00325 g005
Figure 6. Framework of the pretrained MobileNet-V3 utilized for the features extraction of CECED CXR images.
Figure 6. Framework of the pretrained MobileNet-V3 utilized for the features extraction of CECED CXR images.
Diagnostics 12 00325 g006
Figure 7. Framework of the pretrained Inception-V3 utilized for the features extraction of CLAHE CXR images.
Figure 7. Framework of the pretrained Inception-V3 utilized for the features extraction of CLAHE CXR images.
Diagnostics 12 00325 g007
Figure 8. Our proposed multi-channel scheme for pneumonia identification.
Figure 8. Our proposed multi-channel scheme for pneumonia identification.
Diagnostics 12 00325 g008
Figure 9. Performance of the proposed multi-channel in comparison with the single channels and dual channels across the different evaluation metrics.
Figure 9. Performance of the proposed multi-channel in comparison with the single channels and dual channels across the different evaluation metrics.
Diagnostics 12 00325 g009
Figure 10. Accuracy curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Figure 10. Accuracy curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Diagnostics 12 00325 g010
Figure 11. Loss curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Figure 11. Loss curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Diagnostics 12 00325 g011
Figure 12. ROC curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Figure 12. ROC curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Diagnostics 12 00325 g012
Figure 13. Precision–recall curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Figure 13. Precision–recall curves for the proposed multi-channel in comparison with the single-channels and dual-channels.
Diagnostics 12 00325 g013
Figure 14. Performance evaluation for some selected state of the art models using the same dataset.
Figure 14. Performance evaluation for some selected state of the art models using the same dataset.
Diagnostics 12 00325 g014
Figure 15. Accuracy performance for some selected state of the art models using the same dataset.
Figure 15. Accuracy performance for some selected state of the art models using the same dataset.
Diagnostics 12 00325 g015
Figure 16. Accuracy results using the raw chest X-ray images.
Figure 16. Accuracy results using the raw chest X-ray images.
Diagnostics 12 00325 g016
Figure 17. Performance evaluations using the raw chest X-ray images.
Figure 17. Performance evaluations using the raw chest X-ray images.
Diagnostics 12 00325 g017
Table 1. Summary of the Related Works.
Table 1. Summary of the Related Works.
AuthorsYearMode of ImagingDL TechniquesClassification TaskEvaluation Results
Cicero et al. [19]2017X-ray ImageGoogLeNet is utilized to classify two classes - normal and abnormal imagesBinary classFor normal class:
SEN = 91%,
SPE = 91%, and
UC = 96.4%
For abnormal class:
SEN (within 74% to 91%),
SPE (within 75% to 91%), and
AUC (within 85% to 96.2%)
Guendel et al. [20]2018X-ray ImageUsed location-aware dense networks technique to identify anomalies in chest X-raysMultiple classPLCO dataset,
mean AUC = 87.4%,
Chest X-ray 14 dataset,
mean AUC = 84.1%
Rajaraman et al. [21]2018X-ray ImageA modified VGG16 is employed for the identification and detection of viral and bacterial pneumoniaBinary classACC (within 91.8% to 96.2%)
Correa et al. [22]2018Ultrasound ImageDetection of pneumonia using 3 layers feed-forward neural networkBinary classSEN = 90.9%
SPE = 100%
Ke et al. [23]2019X-ray ImageDetection of lung diseases using an approach called neuroheuristicMultiple classSensitivity = 84.22%,
Accuracy = 79.06%,
Specificity = 66.7%
Saraiva et al. [24]2019X-ray ImageA CNN model was applied on a dataset of 5863 images and cross-validation was used for the validation of the modelBinary classAccuracy = 95.30%
Sirazitdinov et al. [25]2019X-ray ImageAn emsemble of RetinaNet and Mask RCNN was appliedBinary classPrecision = 75.0%,
Recall = 79%,
F1-score = 77.0%
Liang and Zheng [26]2020X-ray ImageA modified 49 convolutional and 2 fully connected layer of a CNN model was used for the classification of children’s lung regionsBinary classF1-score = 92.7%
Apostolopoulos et al. [27]2020X-ray ImageDifferent fine-tuning approaches were evaluated for the automatic detection of pneumoniaBinary classVGG19 has the highest value of:
Sensitivity = 92.85%,
Specificity = 98.75%,
Accuracy = 98.75%
Xu et al. [28]2020X-ray ImageMultiple CNN models were compared in order to categorize the classes of CT scansMultiple classAccuracy = 86.7%
Habib et al. [29]2020X-ray ImageDetection of pneumonia using an ensemble of VGG-19 and CheXNet for the extraction of features and random forest as the classifierBinary classAccuracy = 98.93%
Chouhan et al. [30]2020X-ray ImageA transfer learning technique is applied for the detection of pneumoniaBinary classAccuracy = 96.4%
Sensitivity 99.0%
El Asnaoui et al. [31]2020X-ray ImageA fine-tuned of eight different models for the detection and classification of pneumoniaBinary classHighest accuracy is the fine-tubed ResNet50 (>96%)
El Asnaoui et al. [32]2020X-ray ImageA comparative findings of seven DL models for the classification and detection of pneumonia (including COVID-19)Multiple classAccuracy Evaluations:
InceptionResNet-V2 = 92.18%,
DenseNet201 = 88.09%
Table 2. Description of the Dataset.
Table 2. Description of the Dataset.
DatasetPneumonia CategoryValueSelected Amount Used
Kaggle database of RSNA [33]Bacterial30291000
Viral29831000
Normal88511000
Rahman et al. [34]COVID-1936161000
Table 3. Parameter for the modified MobileNet-V3. bneck represents bottleneck convolution, SE depicts whether there is a Squeeze-and-Excite in that block, NL represents the type of non-linearity utilized, HS represents h-swish, RE denotes ReLU and S represents stride.
Table 3. Parameter for the modified MobileNet-V3. bneck represents bottleneck convolution, SE depicts whether there is a Squeeze-and-Excite in that block, NL represents the type of non-linearity utilized, HS represents h-swish, RE denotes ReLU and S represents stride.
InputOperatorExpansion
Size
OutputSENLStride
224 × 224 × 3Conv2d, 3 × 3-16NoHS2
112 × 112 × 16bneck, 3 × 31616YesRE2
56 × 56 × 16bneck, 3 × 37224NoRE2
28 × 28 × 24bneck, 3 × 38624NoRE1
28 × 28 × 24bneck, 5 × 59640YesHS2
14 × 14 × 40bneck, 5 × 524040YesHS1
14 × 14 × 40bneck, 5 × 524040YesHS1
14 × 14 × 40bneck, 5 × 512048YesHS1
14 × 14 × 48bneck, 5 × 514448YesHS1
7 × 7 × 96bneck, 5 × 528896YesHS2
7 × 7 × 96bneck, 5 × 557696YesHS1
7 × 7 × 96bneck, 5 × 557696YesHS1
7 × 7 × 256Conv2d, 1 × 1-256YesHS1
1 × 1 × 256Avg pool, 7 × 7--No-1
1 × 1 × 512Conv2d, 1 × 1-512NoHS1
Table 4. Comparison of our proposed model with single channels and dual channels.
Table 4. Comparison of our proposed model with single channels and dual channels.
ModelACC (%)SEN (%)SPE (%)PRE (%)F1-s (%)Time (min)
LBP-Channel Shallow CNN (LCSC)90.992.393.191.292.73.2
CECED-Channel MobileNet-V3 (CCM)93.795.495.794.395.518.6
CLAHE-Channel Inception-V3 (CCI)95.694.996.295.895.321.8
LBP-Channel Shallow CNN + CECED-channel MobileNet-V3 (LCSC + CCM)92.293.794.592.794.323.4
LBP-Channel Shallow CNN + CLAHE-channel Inception-V3 (LCSC + CCI)94.495.596.895.196.622.7
CLAHE-Channel inception-V3 + CECED-channel MobileNet-V3 (CCI + CCM)97.597.398.397.898.126.8
LBP-Channel Shallow CNN + CLAHE-channel Inception-V3 + CECED-channel MobileNet-V3 (LCSC + CCI + CCM)98.398.999.298.899.030.3
Table 5. Result comparison of our proposed model with state-of-the-art methods for pneumonia classification.
Table 5. Result comparison of our proposed model with state-of-the-art methods for pneumonia classification.
AuthorsACC (%)SEN (%)SPE (%)
Cicero et al. [19]91.091.091.0
Correa et al. [22]-90.9100.0
Apostolopoulos et al. [27]98.092.998.8
Xu et al. [28]86.786.9-
Habib et al. [29]98.93--
Chouchan et al.  [30]96.499.6-
Yamaç et al. [35]86.579.290.7
Wang et al. [36]93.390.795.5
Li et al. [37]96.997.894.9
J.K. K. Singh and A. Singh [38]95.896.195.7
Yang et al. [39]88.464.792.9
Wang et al. [40]94.594.797.3
Alsharif et al. [41]99.799.799.8
Alqudah et al. [42]93.993.296.6
Alquran et al. [43]93.192.996.4
Masad et al. [44]98.998.399.2
Our Model98.398.999.2
Table 6. Comparison table for the selected state of the art models using the same dataset.
Table 6. Comparison table for the selected state of the art models using the same dataset.
ModelACC (%)SEN (%)SPE (%)
Xu et al. [28]91.291.893.1
Wang et al. [36]94.092.994.2
Li et al. [37]96.194.495.5
Yang et al. [39]93.592.494.7
Wang et al. [40]95.895.496.4
Our Model98.398.999.2
Table 7. Results obtained on our dataset using different pretrained models on our proposed model.
Table 7. Results obtained on our dataset using different pretrained models on our proposed model.
ModelLBP-Based ChannelCECED-Based ChannelCLAHE-Based Channel
ACC (%)SEN (%)SPE (%)ACC (%)SEN (%)SPE (%)ACC (%)SEN (%)SPE (%)
AlexNet89.291.492.689.387.590.292.594.393.7
VGG-1688.990.791.490.990.191.891.492.891.9
ResNet-15284.686.287.991.492.393.187.888.187.6
MobileNet-V387.789.490.593.795.495.790.491.690.8
DenseNet-12185.387.188.792.892.893.388.489.288.7
Inception-V386.388.689.493.191.593.795.694.996.2
Shallow CNN90.992.393.187.286.188.485.986.285.7
Table 8. Performance evaluation of our proposed model based on different hyperparameter tuning on our dataset with Adam optimizer.
Table 8. Performance evaluation of our proposed model based on different hyperparameter tuning on our dataset with Adam optimizer.
Hyperparameters(LCSC + Adam)(CCI + Adam)(CCM + Adam)(LCSC + CCI + Adam)(LCSC + CCM + Adam)(CCI + CCM + Adam)(LCSC + CCI + CCM + Adam)
Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)
LR (0.1) + Dropout (0.25)87.581.689.687.385.690.288.1
LR (0.1) + Dropout (0.50)86.987.390.789.687.489.587.6
LR (0.1) + Dropout (0.75)83.682.687.989.989.391.789.2
LR (0.01) + Dropout (0.25)81.485.191.190.491.790.487.1
LR (0.01) + Dropout (0.50)89.884.986.391.190.885.588.4
LR (0.01) + Dropout (0.75)84.790.792.885.789.287.889.7
LR (0.001) + Dropout (0.25)82.295.688.284.692.192.191.5
LR (0.001) + Dropout (0.50)80.791.393.486.589.794.792.6
LR (0.001) + Dropout (0.75)88.392.793.791.490.889.693.8
LR (0.0001) + Dropout (0.25)85.983.485.687.992.496.397.4
LR (0.0001) + Dropout (0.50)90.980.687.194.492.297.598.3
LR (0.0001) + Dropout (0.75)79.586.288.989.893.395.795.9
Table 9. Performance evaluation of our proposed model based on different hyperparameter tuning on our dataset with RMSProp optimizer.
Table 9. Performance evaluation of our proposed model based on different hyperparameter tuning on our dataset with RMSProp optimizer.
Hyperparameters(LCSC + RMSProp)(CCI + RMSProp)(CCM + RMSProp)(LCSC + CCI + RMSProp)(LCSC + CCM + RMSProp)(CCI + CCM + RMSProp)(LCSC + CCI + CCM + RMSProp)
Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)
LR (0.1) + Dropout (0.25)88.589.188.987.690.789.790.1
LR (0.1) + Dropout (0.50)89.787.489.587.289.590.391.3
LR (0.1) + Dropout (0.75)86.390.788.189.389.691.689.8
LR (0.01) + Dropout (0.25)87.788.289.688.488.492.591.6
LR (0.01) + Dropout (0.50)81.491.591.290.690.990.993.4
LR (0.01) + Dropout (0.75)81.189.890.489.791.692.192.8
LR (0.001) + Dropout (0.25)83.892.392.288.990.191.494.4
LR (0.001) + Dropout (0.50)86.589.694.387.393.793.796.5
LR (0.001) + Dropout (0.75)84.290.993.590.492.495.395.9
LR (0.0001) + Dropout (0.25)85.991.194.791.893.694.597.4
LR (0.0001) + Dropout (0.50)88.689.593.990.692.993.996.2
LR (0.0001) + Dropout (0.75)85.388.994.689.791.592.695.5
Table 10. Performance evaluation of our proposed model based on different hyperparameter tuning on our dataset with SGD optimizer.
Table 10. Performance evaluation of our proposed model based on different hyperparameter tuning on our dataset with SGD optimizer.
Hyperparameters(LCSC + SGD)(CCI + SGD)(CCM + SGD)(LCSC + CCI + SGD)(LCSC + CCM + SGD)(CCI + CCM + SGD)(LCSC + CCI + CCM + SGD)
Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)Accuracy (%)
LR (0.1) + Dropout (0.25)87.288.887.786.187.190.889.7
LR (0.1) + Dropout (0.50)85.589.388.188.589.689.190.5
LR (0.1) + Dropout (0.75)87.997.589.587.988.290.991.6
LR (0.01) + Dropout (0.25)89.190.991.989.789.791.789.1
LR (0.01) + Dropout (0.50)82.392.390.391.591.492.391.3
LR (0.01) + Dropout (0.75)83.688.789.590.390.190.893.5
LR (0.001) + Dropout (0.25)84.991.991.789.191.292.592.7
LR (0.001) + Dropout (0.50)85.790.492.490.592.591.695.9
LR (0.001) + Dropout (0.75)83.591.694.191.993.894.994.4
LR (0.0001) + Dropout (0.25)86.389.393.389.792.995.396.2
LR (0.0001) + Dropout (0.50)87.190.894.691.593.692.595.6
LR (0.0001) + Dropout (0.75)88.889.293.090.392.393.796.8
Table 11. Performance evaluation of the proposed model on the single and ensemble models using raw CXR image.
Table 11. Performance evaluation of the proposed model on the single and ensemble models using raw CXR image.
ModelACC (%)SEN (%)SPE (%)PRE (%)F1 (%)
Raw image Shallow CNN (RISC)82.183.680.882.983.3
Raw image MobileNet-V3 (RIM)88.089.290.589.789.0
Raw image Inception-V3 (RII)93.3491.592.793.091.9
Raw image Shallow CNN + Raw image Inception-V3 (RISC + RII)91.789.491.090.191.5
Raw image Shallow CNN + Raw image MobileNet-V3 (RISC + RIM)85.886.187.389.688.9
Raw image Inception-V3 + Raw image MobileNet-V3 (RII + RIM)95.294.696.194.295.7
Raw image Shallow CNN + Raw image Inception-V3 + Raw image MobileNet-V3 (RISC + RII + RIM)96.996.095.496.595.0
Table 12. Performance results obtained using the raw chest X-ray images on different pretrained models on our proposed model.
Table 12. Performance results obtained using the raw chest X-ray images on different pretrained models on our proposed model.
ModelRaw Image
ACC (%)SEN (%)SPE (%)
AlexNet90.989.191.0
VGG-1689.690.389.2
ResNet-15290.288.589.0
MobileNet-V388.089.290.5
DenseNet-12187.789.188.3
Inception-V393.391.592.7
Shallow CNN82.183.680.8
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nneji, G.U.; Cai, J.; Deng, J.; Monday, H.N.; James, E.C.; Ukwuoma, C.C. Multi-Channel Based Image Processing Scheme for Pneumonia Identification. Diagnostics 2022, 12, 325. https://doi.org/10.3390/diagnostics12020325

AMA Style

Nneji GU, Cai J, Deng J, Monday HN, James EC, Ukwuoma CC. Multi-Channel Based Image Processing Scheme for Pneumonia Identification. Diagnostics. 2022; 12(2):325. https://doi.org/10.3390/diagnostics12020325

Chicago/Turabian Style

Nneji, Grace Ugochi, Jingye Cai, Jianhua Deng, Happy Nkanta Monday, Edidiong Christopher James, and Chiagoziem Chima Ukwuoma. 2022. "Multi-Channel Based Image Processing Scheme for Pneumonia Identification" Diagnostics 12, no. 2: 325. https://doi.org/10.3390/diagnostics12020325

APA Style

Nneji, G. U., Cai, J., Deng, J., Monday, H. N., James, E. C., & Ukwuoma, C. C. (2022). Multi-Channel Based Image Processing Scheme for Pneumonia Identification. Diagnostics, 12(2), 325. https://doi.org/10.3390/diagnostics12020325

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop