Next Article in Journal
Automated Quantitative Analysis of Anterior Segment Inflammation Using Swept-Source Anterior Segment Optical Coherence Tomography: A Pilot Study
Next Article in Special Issue
Cardiac Involvement in Children Affected by COVID-19: Clinical Features and Diagnosis
Previous Article in Journal
Evaluating the Performance of Deep Learning Frameworks for Malaria Parasite Detection Using Microscopic Images of Peripheral Blood Smears
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Recommender System for the Efficient Treatment of COVID-19 Using a Convolutional Neural Network Model and Image Similarity

1
Department of Computer Science and Engineering, IIIT, Bhubaneswar 751003, India
2
Stroke Monitoring and Diagnostic Division, AtheroPointTM, Roseville, CA 95661, USA
3
Department of Electrical and Computer Engineering, Idaho State University, Pocatello, ID 83209, USA
4
Department of Radiology, University of Cagliari, 09123 Cagliari, Italy
5
Knowledge Engineering Center, Global Biomedical Technologies, Inc., Roseville, CA 95661, USA
*
Authors to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2700; https://doi.org/10.3390/diagnostics12112700
Submission received: 9 October 2022 / Revised: 30 October 2022 / Accepted: 3 November 2022 / Published: 5 November 2022
(This article belongs to the Special Issue AI-Based COVID-19 Detection)

Abstract

:
Background: Hospitals face a significant problem meeting patients’ medical needs during epidemics, especially when the number of patients increases rapidly, as seen during the recent COVID-19 pandemic. This study designs a treatment recommender system (RS) for the efficient management of human capital and resources such as doctors, medicines, and resources in hospitals. We hypothesize that a deep learning framework, when combined with search paradigms in an image framework, can make the RS very efficient. Methodology: This study uses a Convolutional neural network (CNN) model for the feature extraction of the images and discovers the most similar patients. The input queries patients from the hospital database with similar chest X-ray images. It uses a similarity metric for the similarity computation of the images. Results: This methodology recommends the doctors, medicines, and resources associated with similar patients to a COVID-19 patients being admitted to the hospital. The performance of the proposed RS is verified with five different feature extraction CNN models and four similarity measures. The proposed RS with a ResNet-50 CNN feature extraction model and Maxwell–Boltzmann similarity is found to be a proper framework for treatment recommendation with a mean average precision of more than 0.90 for threshold similarities in the range of 0.7 to 0.9 and an average highest cosine similarity of more than 0.95. Conclusions: Overall, an RS with a CNN model and image similarity is proven as an efficient tool for the proper management of resources during the peak period of pandemics and can be adopted in clinical settings.

1. Introduction

SARS-CoV-2 coronavirus was first discovered and reported in Wuhan, China, in 2019 and has spread globally, causing a health hazard [1,2,3]. On 30 January 2020, the World Health Organization labelled the outbreak a Public Health Emergency of International Concern, and on 11 March 2020, it was declared a pandemic. COVID-19 has varied effects on different people. The majority of infected patients experience mild to moderate symptoms and do not require hospitalization. Fever, exhaustion, cough, and a loss of taste or smell are all common COVID-19 symptoms [4]. Loss of smell, confusion, trouble breathing or shortness of breath, and chest discomfort are some of the major symptoms that lead to serious pneumonia in both lungs [1,4,5,6]. COVID-19 pneumonia is a serious infection with a high mortality rate. The signs of a COVID-19 infection progressing into dangerous pneumonia include a fast pulse, dyspnea, confusion, rapid breathing, heavy sweating, and pulmonary embolism [7,8]. It induces serious lung inflammation, as seen in lung microscopy [9]. It puts strain on the cells and tissue that cover the lungs’ air sacs. The oxygen for breathing is collected and supplied to the bloodstream through these sacs. Due to injury, tissue breaks off and blocks the lungs [10]. The sacs’ walls might thicken, making breathing extremely difficult.
The most prevalent method of diagnosing individuals with respiratory disorders is chest radiography imaging [11,12,13]. At the beginning of COVID-19, a chest radiography image appeared normal, but it gradually altered in a fashion that may be associated with pneumonia or acute respiratory distress syndrome (ARDS) [11]. Figure 1 depicts the progression of chest X-ray images for a 45-year-old person infected with COVID-19. Roughly 15% of COVID-19 patients require hospitalization and oxygen therapy. Approximately 5% of people develop serious infections and require a ventilator.
During the peak period of infection transmission, having enough oxygen and a ventilator is also a major challenge for hospitals [14,15]. As a result, hospitals and medical practitioners are under a lot of stress trying to deal with critical patients who have been admitted to hospitals [16]. They concentrate on providing good care to individuals who are hospitalized so that the mortality rate can be lowered, and the patients can recover quickly. However, hospitals’ capability to provide adequate treatments to hospitalized patients is sometimes limited by the availability of doctors and resources. In this scenario, a recommender system (RS) using machine learning (ML) approaches might be used to administer the best treatment while working with limited resources [17,18]. As the mortality rate and recovery rate of seriously hospitalized COVID-19 patients generally depend upon the amount of infection in the lungs [19,20,21], the radiographic lung images of those patients can be used to recommend proper treatment in terms of a doctor, medicine, and other related resources.
From the perspective of the RS’s implementation, a new patient’s chest X-ray image is sent to the proposed system, and doctors, medicines, and resources are recommended for that patient. The proposed system assumes that the database consists of lung images and other information such as the name of the doctor assigned, medicines, and resources provided, such as intensive care unit (ICU), oxygen therapy, and ventilators. The COVID-19 patients who were admitted to the hospital in the past successfully recovered from the hospital. It uses a collaborative filtering method to find similar COVID-19 patients to new COVID-19 patients using image similarity. The proposed approach uses convolutional neural networks (CNN) for feature extraction [22,23,24] from images and utilizes those feature vectors for similarity computation.
The proposed collaborative RS uses image similarity to produce recommendations, as image similarity is a popular and efficient technique in image-based RS [25,26,27]. Traditional image-based RS recommends images for a given input query image. The novelty of the proposed RS is that it recommends some metadata information such as doctors, medicines, and resources for a given input query such as a chest X-ray image. The proposed system is built around two hypotheses. The first hypothesis states that the proposed system’s performance is dependent on the feature extraction technique used by CNN models. The second hypothesis states that the proposed system’s performance is also affected by the similarity measure used for similarity computation. Higher similarity assures a more accurate recommendation. The chest X-ray images are compared based on their feature vectors. The CNN model is used to effectively extract feature vectors from chest X-ray images. The combination of a robust search strategy and the best feature selection approach may make the RS more powerful for efficient and accurate recommendations. Figure 2 represents the global system representation of the proposed approach. The proposed system can be analyzed as a combination of an online and an offline system. The offline system is responsible for the feature extraction process from the images, and the online system handles the recommendation process. The web-based system is equipped with a performance module that calculates accuracy based on the known reference values in the test dataset.
In particular, the objectives of this article include: (i) to propose an efficient RS system for COVID-19 based on chest X-ray images to address the impact of an RS on the efficient handling of situations in hospitals during the peak period of a pandemic with limited resources; (ii) to use multiple CNN models to construct an RS using COVID-19 chest X-ray images; (iii) to propose a unique design by embedding four kinds of search paradigm in the CNN-based framework; (iv) comparative data analysis of different similarity measure in the RS framework, providing metadata which includes doctors, medicines, and resources; (v) finally, to mitigate the impact of an RS in the healthcare domain through improved services and efficient resource management.
The remainder of this study is as follows: Section 2 includes the discussion of RS, CNN, the feature extraction process, measures for similarity computation, and studies related to the proposed work. The proposed model is explained in Section 3, and the experimental evaluation is discussed in Section 4. Section 5 focuses on future scope, and the article is concluded in Section 4.

2. Background Literature

Many researchers have presented various models employing traditional machine learning approaches in the past for the identification of COVID-19 using radiography images [2,28]. Zimmerman et al. [29] reviewed many cardiovascular uses of machine learning algorithms, as well as their applications to COVID-19 diagnosis and therapy. The authors in refs. [30,31] proposed image analysis tools to classify lung infection in COVID-19 based on chest X-ray images and claimed that artificial intelligence (AI) methods have the potential to improve diagnostic efficiency and accuracy when reading portable chest X-rays. In ref. [19], the authors established an ensemble framework of five classifiers such as K-nearest neighbors (KNN), naive Bayes, decision tree, support vector machines (SVM), and artificial neural network for the detection of COVID-19 using chest X-ray images. Ref. [32] describes a method for detecting SARS-CoV-2 precursor-miRNAs (pre-miRNAs) that aids in the identification of specific ribonucleic acids (RNAs). The method employs an artificial neural network and proposes a model with an estimated accuracy of 98.24%. The proposed method would be useful in identifying RNA target regions and improving recognition of the SARS-CoV-2 genome sequence in order to design oligonucleotide-based drugs against the virus’s genetic structure.
Due to the unprecedented benefits of a deep CNN in image processing, it has been successfully utilized by various researchers for the identification and accurate diagnosis of COVID-19. In ref. [20], the authors proposed a deep learning (DL) model for the detection of COVID-19 by annotating computed tomography (CT) and X-ray chest images of patients. In ref. [33], various DL models such as ResNet-152, VGG-16, ResNet-50, and DenseNet-121 were applied to radiographic medical images for the identification of COVID-19 and were compared and analyzed. To overcome the lack of information and enhance the training time, the authors also applied transfer learning (TL) techniques to the proposed system. A voting-based approach using DL for the identification of COVID-19 was proposed in ref. [34]. The proposed method makes use of CT scan chest images of patients and utilizes a voting mechanism to classify a CT scan image of a new patient. Various DL algorithms for identifying COVID-19 infections from lung ultrasound imaging were reviewed and compared by the authors in ref. [35]. The proposed method adopts four pre-trained models of DL such as InceptionV3, VGG-19, Xception, and ResNet50, for the classification of a lung ultrasound image. In ref. [36], the authors compared the results of using CNNs pre-trained with ML-based classification algorithms. The major purpose of this research was to see how CNN-extracted features affect the construction of COVID-19 and non-COVID-19 classifiers. The usefulness of DL learning algorithms for the detection of COVID-19 using chest X-ray images is demonstrated in ref. [37]. The proposed approach was implemented using 15 different pre-trained CNN models, and VGG-19 showed a maximum classification accuracy of 89.3%. In ref. [38], an object detection approach using DL for the identification of COVID-19 in chest X-ray images was presented. The suggested method claims a sensitivity of 94.92% and a specificity of 92%.
Many kinds of research have also been conducted in the past using image segmentation, image regrouping, and other hybrid techniques for accurate diagnosis of COVID-19 [39]. In ref. [40], the authors proposed an innovative model using multiple segmentation methods on CT scan chest images to determine the area of pulmonary parenchyma by identifying pulmonary infiltrates (PIs) and ground-glass opacity (GGO). In ref. [41], the authors proposed a hybrid model for the detection of COVID-19 using feature extraction and image segmentation techniques to improve the classification accuracy in the detection of COVID-19. In ref. [42], a hybrid approach of feature extraction and CNN on chest X-ray images for the detection of COVID-19 using a histogram-oriented gradient (HOG) algorithm and watershed segmentation methodology was proposed. This proposed hybrid technique showed satisfactory results in the detection of COVID-19 with an accuracy of 99.49%, sensitivity of 93.65%, and specificity of 95.7%. In ref. [43], the authors came up with a new way to determine COVID-19 in images of chest X-rays using image segmentation and image regrouping. The proposed approach was found to outperform the existing models for the identification of COVID-19 in terms of classification accuracy with a lower amount of training data. In ref. [44], the transfer learning technique was used in conjunction with image augmentation to train and validate several pretrained deep Convolutional Neural Networks (CNNs). The networks were trained to classify two different schemes: (i) normal and COVID-19 pneumonia and (ii) normal, viral, and COVID-19 pneumonia with and without image augmentation. The classification accuracy, precision, sensitivity, and specificity for both schemes were 99.7%, 99.7%, 99.7%, and 99.55% and 97.9%, 97.95%, 97.9%, and 98.8%, respectively. The high accuracy of this computer-aided diagnostic tool can significantly improve the speed and accuracy of COVID-19 diagnosis. A systematic and unified approach for lung segmentation and COVID-19 localization with infection quantification from CXR images was proposed in ref. [45] for accurate COVID-19 diagnosis. The proposed method demonstrated exceptional COVID-19 detection performance, with sensitivity and specificity values exceeding 99%.
RS has also been useful in combating the COVID-19 pandemic by making recommendations such as medical therapies for self-care [46], wearable gadgets to prevent the COVID-19 outbreak [47], and unreported people to reduce infection rates by contact tracing [48], among others. An RS based on image content was proposed in ref. [25] that employed a random forest classifier to determine the product’s class or category in the first phase and employed the JPEG coefficients measure to extract the feature vectors of the photos in the second phase to generate recommendations using feature vector similarity. A neural network-based framework for product selection based on a specific input query image was provided by ref. [26]. The suggested system employed a neural network to classify the supplied input query image, followed by another neural network that used the Jaccard similarity measure to find the most comparable product image to that input image. In ref. [27], the authors developed a two-stage DL framework using a neural network classifier and a ranking algorithm for recommending fashion images based on similar input images. Traditional RS frequently faces a significant challenge in learning relevant features of both users and images in big social networks with sparse relationships between users and images, as well as the widely different visual contents of images. Refs. [49,50,51] presented a strategy for solving this data sparsity problem in content-based and collaborative filtering RS by importing additional latent information to identify users’ probable preferences.
The majority of previous research in RS based on computer vision was conducted for the e-commerce domain, with only a few works carried out for the healthcare domain, according to the literature. It was also revealed from the literature that image similarity is one of the successful techniques used for designing RS in computer vision. Furthermore, the efficacy of computer vision in RS in providing solutions for combating the COVID-19 pandemic has yet to be investigated. In this context, we suggest a health recommender system (HRS) that uses image similarity and collaborative filtering to provide treatment suggestions for COVID-19.

3. Methodology

3.1. Recommender System

RS is a software program that aids in the personalization of users and is based on the principle of information filtering [52]. RS has the following formal definition: Let P represent the set of all users, and Q represent the set of all items that can be recommended. Let t be a utility function that measures the usefulness of item q to user p, i.e., t: p × q → S, where S is an ordered set. The items q′ ∈ Q that maximize the user’s utility will then be recommended for each user p ∈ P. As a result, ∀p ∈ P, qp′ = arg maxq∈Q t (p, q) may be stated more formally. RS can be broadly divided into four types, as shown in Figure 3.
Content-based RS or cognitive RS provides recommendations based on a comparison of the items’ content with a user profile [53,54]. Collaborative RS collects preferences or taste information from the collaborated users to produce automatic predictions regarding the user’s interests [55,56]. Memory-based and model-based are the two different categories of a collaborative RS. A memory-based collaborative RS makes use of all the data to provide recommendations based on user or item similarity, whereas model-based collaborative filtering RS entails creating a model based on all of the data in order to detect similarities between users or items for recommendation purposes. Hybrid RS combines two or more recommendation algorithms in different ways to take advantage of their different strengths [57,58]. A knowledge-based RS intelligently filters a group of targets to fulfil the user’s preferences. It assists in overcoming the difficulties of both collaborative and content-based RSs [59,60].
An RS used in health applications to analyze patients’ digital data and filter out the best information according to their profile is known as a health recommender system (HRS) [61,62]. HRS can be thought of as a decision-making system that plays a big role in society by advising patients on suitable disease treatments and doctors on good disease diagnoses.

3.2. Convolutional Neural Network

A CNN is a powerful DL tool for image processing and recognition [63]. The fundamental architecture of a CNN contains three distinct types of layers: convolutional, pooling, and fully connected, as shown in Figure 4.

3.2.1. Convolutional Layer

The basic layer of CNN is the convolutional layer, which has the responsibility of extracting features and recognizing patterns in input images. The CNN extracts low-level and high-level features by passing images from the training dataset through a filter comprised of feature maps and kernels [23]. The convolutional layer’s output can be expressed in the following way:
q s r m , n = c   x , y   p c x , y g r s u , v ,  
where qsr (m, n) is the convolution layer, and pc (x, y) is an element of the input image tensor pc, multiplied by the g r s u , v index of the sth convolutional kernel sr of the rth layer element-wise.

3.2.2. Pooling Layer

The pooling layer, or the down-sampling layer, gathers comparable data in the vicinity of the feature layer and generates the dominating response inside this layer. The pooling process aids in the extraction of a group of features that are invariant to xi, and it can be defined as Equation (2).
YrS = tj (QSr),
where for the sth input feature-map Qsr, YrSconveys the pooled feature-map of the rth layer.

3.2.3. Fully Connected Layer

The fully connected layer is utilized for classification at the end of the CNN network. This layer takes the features that have been collected at different stages of the network as the input and then analyses and compares those features to the results from all the other layers.

3.2.4. Activation Function

The activation function modifies the weighted sum input of one node for a given layer and uses it to activate that node for a certain input. The activation function assists in the learning of feature patterns by acting as a decision function. ReLU is one of the most widely used activation functions due to its ability to handle the gradient problem in CNN models. Mathematically, the ReLU activation function can be defined as follows:
R e L U x =   0 ,     x < 0 x ,     x   0

3.2.5. Batch Normalization

Batch normalization is used for the normalization of the output of the preceding layers, which can assist with issues such as internal covariance shifts in feature maps. The equation for the batch normalization of transformed feature map Qrs can be defined as shown in Equation (4).
B r s = Q r s µ b σ b 2 +
where B r s denotes the normalized feature map, and Q r s represents the input feature map. The mean and the variance of the feature map are represented by µb and σb2, respectively. ϵ is used to deal with the numerical instability caused by division by zero.

3.2.6. Dropout and Flatten Layer

Dropout is a technique for adding regularization to a CNN network, which establishes generalization by omitting some connections at random. After removing some random connections, the network design with the lowest weight value is chosen as a close approximation of all the suggested networks. The Flatten layer transforms the pooled feature map into a one-dimensional array that is passed as a single feature vector to the next layer.

3.3. Feature Extraction Methods

CNN is widely used in computer vision for feature extraction because it can discover relevant features from images without requiring human interaction and is computationally efficient. There are various models of CNN for the feature extraction process. In this study, we tested the performance of our proposed system with two specific CNN models, namely: the residual neural network (ResNet) and the visual geometry group (VGG). We used three different versions of ResNet, namely ResNet-50, ResNet-101, and ResNet-152, and two versions of VGG, namely VGG-16 and VGG-19. The detailed architectures of both ResNet and VGG are described in the following subsections.

3.3.1. ResNet

ResNet is an artificial neural network that can solve the problem of training very deep networks using residual blocks [64]. Several COVID-19-related publications have been tried using ResNet or the Hybrid nature of ResNet [65,66,67]. The basic architecture of a ResNet network is shown in Figure 5.
A ResNet model with these residual blocks is shown in Figure 6. A direct connection in the ResNet model can skip some layers and is known as a “skip connection”, which is the heart of the model. The model produces a different output due to this skipped connection. When the connection is not skipped, the input X is multiplied by the weights of the following layer, and a bias term is added to this. Therefore, Equation (5) can be used to describe the model’s output function.
H x =   f x + b ,         w i t h o u t   s k i p   c o n n e c t i o n f x ,                   w i t h   s k i p   c o n n e c t i o n .
Double or triple-layer skips with nonlinearities (ReLU) and batch normalization are used in most ResNet models [64]. An additional weight matrix can be utilized, and such models are termed “Highway Nets.” We used three variations of ResNet, namely ResNet-50 [68], ResNet-101 [69], and ResNet-152 [70].

3.3.2. VGG Net

VGG Net is a traditional CNN model composed of blocks, each consisting of 2D convolution and max pooling layers. The basic architecture of a VGG Net is shown in Figure 7. It was created to improve the performance of a CNN model by increasing the depth of the model. The VGG16 and VGG-19 are the two versions available. VGG-16 and VGG-19 include 16 and 19 layers of convolution layers, respectively. The VGG Net architecture serves as the foundation for cutting-edge object recognition models. The VGG Net, which was created as a deep neural network (DNN), outperforms baselines on a variety of tasks and datasets. Small convolutional filters are used to build the VGG network. VGG-16 [71] and VGG-19 [72] are two different versions of VGG that we used to test our proposed system.

3.3.3. Similarity Measures

The similarity measure is a means of determining how closely data samples are related. It plays an important role in computer vision by aiding in the comparison of two images by determining their feature vector similarity [64,74]. The proposed model uses the cosine similarity measure to compute the similarity between two feature vectors to find the most similar images to the input image, which are further utilized for the recommendation process.

3.3.4. Cosine Similarity Measure

The similarity between two vectors using cosine similarity can be calculated as follows:
CosSim U ,   V = U · V U V = i = 1 n U i V i i = 1 n U i 2     i = 1 n V i 2    
where U and V represent two vector components. The cosine similarity is measured on a scale of 0 to 1, with 0 representing no similarity and 1 representing 100% similarity. All the other values in the range [0, 1] show the equivalent percentage of similarity.

3.3.5. Euclidean Distance Similarity Measure

The Euclidean similarity between image vectors U and V can be calculated as follows:
E u c S i m U ,   V = i = 1 n U i V i 2 .
The Euclidean similarity is also measured on a scale of 0 to 1, with 0 representing no similarity and 1 representing 100% similarity. All other values in the [0, 1] range reflect the equivalent percentage of similarity.

3.3.6. Jaccard Similarity Measure

Jaccard similarity is a popular proximity measurement that is used to determine the similarity between two objects. The Jaccard similarity is calculated by dividing the number of observations in both sets by the number of observations in either set. It is also graded on a scale of 0 to 1, with 0 indicating no similarity and 1 indicating complete similarity. All other values in the [0, 1] range correspond to the same percentage of similarity. The similarity between two vectors using Jaccard similarity can be calculated as:
J a c S i m U ,   V = U V U V .

3.3.7. Maxwell–Boltzmann Similarity Measure

The Maxwell–Boltzmann similarity is a popular similarity measure for document classification and clustering [75]. It is calculated using the overall distribution of feature values and the total number of nonzero features found in the documents. The difference between the two documents is represented by the following:
D U ,   V l o g N D G D ,
where,
N D = 0.5   U k V k 2                       i f   U k   a n d   V k > 0 λ k q 0.5   U k 2         i f   U k > 0   a n d   V k = 0 λ k q 0.5   V k 2         i f   V k > 0   a n d   U k = 0 0                                     O t h e r w i s e
and
G D = t n z                                               i f   U k   a n d   V k > 0   t n z u                         i f   U k > 0   a n d   V k = 0 t n z v                       i f   V k > 0   a n d   U k = 0 0                                     O t h e r w i s e ,
where,
  • tnz = the total number of nonzero attributes,
  • tnzu = the total number of features of U having nonzero items,
  • tnzv = the total number of features of V having nonzero items,
  • 0 < λ < 1, k denotes features, and q denotes the number of present–absent pairs.
The Maxwell–Boltzmann similarity is calculated from the value of D as follows:
M a x w e l l B o l t z S i m U ,   V = 1 1 + D .

3.4. Proposed Model

In general, computer vision-based RSs are based upon the assumption that a user submits or picks an image of a product, and the user is then provided with similar products/images [25]. The proposed method slightly deviates from this assumption as it extracts features from past COVID-19 patients’ chest X-ray images and recommends some metadata information related to treatment alternatives based on these images.
The proposed framework is aimed at providing emergency solutions to hospitals during the COVID-19 pandemic using the information of past COVID-19 patients who have successfully recovered from the hospital. Therefore, it also assumes that the hospital database used to implement the proposed RS contains the chest X-ray images of COVID-19 patients who have recovered from the hospital along with the metadata (associated information) such as doctors who have investigated the patient, medicine, and resources (ICU, oxygen mask, and ventilator) provided to the patient. The architecture or the local system for the proposed RS is shown in Figure 8 and consists of two major phases: fine-tuning CNN models for feature learning (Phase-1) and recommendation (Phase-2). Phase 1 of the system is an offline system, while phase 2 is an online system.
The overall algorithm of the proposed framework is also provided in Algorithm 1. The algorithm illustrates the basic workflow of the proposed system. The system takes the chest X-ray image of a new patient as the input and recommends doctors, medicine, and hospital resources as the output. It uses a CNN model to extract the feature vectors of the input chest X-ray image and all the chest X-ray images of past COVID-19 patients stored in the hospital database. It uses a similarity measure to compute the most similar COVID-19 patients to a new patient and utilizes the metadata associated with them for the recommendations, which is represented in the testing protocol from step 5 to step 11 in the algorithm. The pre-processing and the training of the chest X-ray images are explained in the training protocol with steps from 1 to 4 in the algorithm.
Algorithm 1: The Overall Algorithm of the Proposed System
Training Protocol for Feature Extraction using deep learning
  • Obtain the chest X-ray images of COVID-19 as training data.
  • Crop those chest X-ray images at random to 224 × 224 and rotate them at random by 30°.
  • Input the transformed chest X-ray images obtained in step 2 into the CNN classifier for fine-tuning and begin the training of the model.
  • When training is completed, extract the desired output layer features and save the model.
Testing Protocol using Similarity Measure
5.
Obtain the chest X-ray images from the database of previous COVID-19 cases.
6.
Resize the chest X-ray images from the COVID-19 database to 225 × 225 and perform a centre crop of 224 × 224.
7.
Extract and store the feature vectors of chest X-ray images from the database using the pre-trained CNN model.
8.
Calculate the similarity of the query image feature vector with all the stored database feature vectors.
9.
Find the top-k similar feature vectors in the database, where k is a positive integer.
10.
Retrieve the chest X-ray images with their records of meta-data from the database, corresponding to the top-k similar feature vectors obtained in step 6.
11.
Recommend the doctors, medicines, and resources present in the retrieved meta-data records to the new patient as the output.
(A)
Phase 1 (Offline System): Fine Tuning for Feature Learning
In phase 1, the proposed method learns to extract infection features from COVID-19 patients’ chest X-ray images based on image characteristics. A CNN model is trained to learn these features by classifying these chest X-ray images into respective lung condition categories (one of which should be COVID-19) as present in the training data. The architecture of a CNN model consists of two components: (1) feature vector extractor and (2) classifier [24,76], as shown in Figure 9.
Several convolution layers are followed by max pooling and an activation function in the feature extraction process. Typically, the classifier is made up of fully connected layers. The proposed approach uses a fine-tuning method, which is more commonly used in radiology research. It involves not only replacing the pre-trained model’s fully connected layers with a fresh set to retrain them on the given dataset but also using backpropagation to fine-tune some of the layers in the pretrained convolutional base. The binary cross-entropy loss function was used for optimization in the training of the CNN models [67,77]. The binary cross entropy loss can be defined by the following equation:
Loss = 1 N i = 1 N y i . log p y i . + 1 y i . . log 1 p y i . .
During training, the ReLU activation function and its variations are also used because they can solve the problem of vanishing gradients, which often happens in CNN models. Before training the model for feature learning, the suggested method utilizes specific image transformations or augmentation, as shown in phase 1 [78,79]. This allows the model to be more adaptable to the huge variation in the region of interest (lungs) within the image, with less emphasis on its location, orientation, and size. Models that are trained with data transformations are more likely to improve CNN’s performance on image datasets and make them more general. In this phase, any efficient CNN model, such as VGG or ResNet, may be trained. The trained model weights are then saved, and the fine-tuned convolutional base is then employed in phase 2 to extract features. Steps 1 to 4 of the proposed algorithm shown in Figure 9 describe phase 1.
(B)
Phase 2 (Online System): Recommendation
Phase 2 of the proposed approach is used for providing recommendations based on the features obtained from X-ray images using the fine-tuned convolutional base from phase 1, which acts as a feature extractor in phase 2. The metadata associated with each image in the database is utilized to provide recommendations such as doctors, medicines, and resources. For recommendation, the system utilizes similar patients from the database who have the same type of infection in the chest due to COVID-19 as that of the patient corresponding to the input query chest X-ray image. In doctor recommendation, it recommends doctors who have already successfully treated similar patients to the patient corresponding to the input query chest X-ray image. In medicine recommendation, the system recommends medicines that have already been consumed by previously recovered patients who had similar chest infections. For resource recommendations, it recommends emergency resources such as oxygen masks, ventilators, and ICU if required by the patient in the future so that the hospital can arrange those resources beforehand. Phase 2 of the proposed method is again divided into two sub-phases: (1) feature vector extraction and (2) similarity-based retrieval.

3.4.1. Feature Vector Extraction

The elements or patterns of an object in an image that assist in identifying it are called “features.” Feature extraction is a step in the dimensionality reduction process, which divides and reduces a large collection of raw data into smaller groupings. It aids in extracting the most useful information from higher-dimensional data by choosing and merging variables into features, hence minimizing the amount of data. These features are easy to use and describe real data collection uniquely and accurately.
CNNs excel in extracting complex features in the form of feature vectors that depict the image in great detail, learning task-specific features while being extremely efficient [80]. Therefore, the proposed method uses CNN-based feature extractors obtained from phase 1 to extract features of the infection present inside the chest X-ray images of COVID-19 patients.
Feature vector extraction is applied both to the input query image and the chest X-ray images of COVID-19 patients present in the hospital database. Steps 5 to 10 of the proposed algorithm describe the feature vector extraction process. However, the extracted feature vectors are further exploited for similarity-based retrieval.

3.4.2. Similarity-Based Retrieval

The extracted feature vectors of the input query image and the chest X-ray images of recovered COVID-19 patients present in the database obtained from the previous step are further utilized to retrieve similar images for a given input query image. The system utilizes the cosine similarity measure to find the top-k similar patients for a given query patient, where k is a positive integer. The system further utilizes those top-k similar patients to provide various recommendations such as doctors, medicines, and resources to the patient corresponding to the given input chest X-ray image. The doctors, medicines, and resources allotted to those similar patients are recommended to the query patient. Steps 11 to 14 summarize the workflow of the proposed system (Figure 10). In the results section, we present the results for the two hypotheses of our proposed system.

4. Experimental Protocol

To verify the efficacy of the proposed approach, the experimental environment, dataset description, pre-processing of the datasets, and the related results of the experiments are discussed in this section.

4.1. Experimental Environment

The details of the computing resources used for the implementation of the proposed system are shown in Table 1.

4.2. Dataset Description

We employed two datasets, including the chest X-ray images of COVID-19 patients, for the implementation and performance evaluation of our proposed model. The detailed descriptions of the datasets are provided in Table 2. The “Dataset for Training and Verification (DFTV)” was used for to train the CNN models and was also used for the analysis of the model’s performance. It was split into training, validation, and test sets using the K5 protocol in the ratio of 8:1:1 before training the models. In total, 16,932 images are used for training, and 2116 images are used for testing. For the performance analysis of our proposed recommendation model, all the images of the COVID class in this dataset were also taken again separately and split into five different subsets of equal size, namely DFTV-1, DFTV-2, DFTV-3, DFTV-4, and DFTV-5. The second dataset, “Dataset for Cross Verification (DFCV)”, was used as a dataset for cross-verification of the system’s performance on completely new data unseen by the CNN models. It was also split into five subsets of equal size, namely DFCV-1, DFCV-2, DFCV-3, DFCV-4, and DFCV-5.

4.3. Data Pre-Processing

In phase 1 of the proposed approach, the images underwent certain image transformations, as mentioned in Step 2 of the algorithm. First, the images were randomly cropped and resized to 224 × 224 and then randomly rotated by 30 degrees before going into the CNN model for training. In phase 2 of the approach, the query and the database images were pre-processed before feature extraction took place. The chest X-ray images were first resized to 225 × 225 and then cropped to size 224 × 224, facilitating the input to ResNet and VGG architectures.

5. Results and Performance Evaluation

The results of the proposed system were compartmentalized based on the system’s two phases. The performance of offline (Phase-1) and online (Phase-2) systems was assessed using different CNN models and similarity measures.

5.1. Results

The results were obtained for the two phases of the proposed system. In phase 1, the results were determined by fine-tuning the CNN model. In phase 2, the results were recorded and obtained from the recommendation process. The results were obtained by considering both the DFTV and DFCV datasets.
Phase 1: Fine Tuning for Feature Learning—Offline System
In phase 1 of the proposed system, the CNN model was fine-tuned, and the model was saved for further use in phase 2. Training was optimized using the stochastic gradient descent (SGD) optimizer through a binary cross-entropy loss function on the DFTV dataset. Figure 11 depicts the result of these fine-tuned CNN models, as found in the test split of the dataset DFTV. The metrics used in the results are defined in the following equations.
Precision = T P   T P + F P ,
R e c a l l =   T P   T P + F N ,
F 1 s c o r e = 2   P r e c i s i o n R e c a l l   P r e c i s i o n + R e c a l l ,
A c c u r a c y = T P + T N S i z e   o f   t h e   d a t a s e t ,
where,
  • TP is the true positive, and this is when the model correctly predicts the positive class.
  • TN is the true negative, and this is when the model correctly predicts the negative class.
  • FP is the false positive, and this is when the model incorrectly predicts the positive class.
  • FN is the false negative, and this is when the model incorrectly predicts the negative class.
From Figure 11, it is found that the weighted precision, recall, and f1-score of all the CNN models are between 0.90 and 0.95. The average weighted precision, recall, and f1-score of all five CNN models are 0.938, 0.936, and 0.936, respectively.
Phase 2: Recommendation—Online System
The final weights of the fine-tuned CNN model obtained from phase 1 were used in phase 2 for feature extraction. The weights were used for the feature extraction of both the query image and the database images. The similarity of the feature vector corresponding to the input query image was determined concerning all the feature vectors corresponding to the database images. The four similarity measures, namely, cosine similarity, Euclidean distance similarity, Jaccard similarity, and Maxwell–Boltzmann similarity, were considered for evaluating the performance of the proposed system. The results were obtained by taking all the images of the COVID-19 class present in each dataset. For each dataset, 80% of these images were considered hospital database images, i.e., chest X-ray images of past recovered COVID-19 patients of the hospital. The feature vectors were already extracted and stored in the backend, and the remaining 20% of images were considered new input query images, i.e., chest X-ray images of new patients. For each query image, the similarity with every database image was calculated, and the top-k images from the hospital database having the highest similarity were retrieved, where k is the number of recommendations. A” threshold value (T)” of similarity was decided to identify relevant similar images for each query image. A retrieved database image was considered relevant when it had a similarity greater than or equal to the threshold value, as defined in Equation (18).
Relevant recommendation = retrieved database image with cosine similarity ≥ T.
We varied this threshold value between 0.7 and 0.95 to analyze different scenarios. This threshold value represents the minimum similarity of image features in the chest X-ray needed for the recommended medicines, doctors, and resources to be considered valid. This threshold value may be fixed after consulting a medical professional for practical use.
For the input query set, the average of the highest similarity corresponding to the most similar image (top-1) concerning each query was calculated and was referred to as the average highest similarity (AHS) of our proposed method, as shown in Appendix A. Table A1 and Table A2 depict the average highest similarity as observed on various datasets using different similarity measures and CNN models for feature extraction.
The performance of the different similarity measures can be analyzed from the graphs shown in Figure 12. From Figure 12, it is observed that the mean of AHS of all the datasets for Maxwell–Boltzmann similarity is maximum. The performance of the Cosine similarity measure is nearer to Maxwell–Boltzmann similarity. The composite means and standard deviation of AHS considering all the datasets for all the models are represented in Figure 13. The performance of the different CNN models was further analyzed, considering Maxwell–Boltzmann similarity.
The mean average precision (MAP@k) metric was used to evaluate the performance of phase 2 (online system) of the proposed RS, which is defined in the following Equation (19).
M A P @ k = 1 N n = 1 N 1 k m = 1 n P m × r e l m ,
where N denotes the total number of users or the length of the input query set, k denotes the number of recommendations made by the recommender system, and P(m) denotes the precision up to the first m recommendations. rel(m) is a relevance indicator function for each recommended item, returning 1 if the mth item is a relevant recommended chest X-ray image, with a similarity higher than the threshold value T, and 0 otherwise. To check the performance of our proposed RS, we determined the MAP@k for k = 5 and k = 10. The values obtained for MAP@k for k = 5 and k = 10 using the five CNN feature extraction models are listed in Appendix A and are shown in Table A3 and Table A4, respectively.
The performance of the models was analyzed through the graphs represented in Figure 14 and Figure 15. From the graphs, it was observed that the performance of the proposed RS varies according to the different feature extraction methods through the different CNN models. The proposed RS implemented with the ResNet-50 feature extraction model provided the highest MAP@k with k = 5 and k = 10 for all the datasets with higher threshold values of similarity. The proposed RS with the ResNet-50 feature extraction CNN model had the highest MAP of more than 0.90 for the threshold similarities in the range of 0.7 to 0.9. Therefore, it confirmed the first part of the hypothesis that the performance of the proposed RS depends upon the feature extraction technique through CNN models. It was also found that this framework provides better performance for the DFCV, which follows the performance obtained from DFTV.
To analyze the effect of the similarity measures on the performance of the proposed system, we found the MAP@5 for both the DFTV and DFCV datasets using the Resnet-50 CNN model. We used the Resnet-50 CNN model as it is the best-performing model for our datasets. The results obtained are shown in Table 3. It was also observed from Table 3 that the MAP@5 was at its maximum using the Maxwell–Boltzmann similarity. Hence, the MAP@k depends upon the similarity measure used for similarity computation, which reveals another part of the hypothesis of our proposed system. We also validated our two hypotheses in the performance evaluation section.

5.2. Performance Evaluation

The proposed study used two performance metrics, (i) the ROC curve and (ii) the figure of merit (FoM), to validate the performance of the proposed system. The ROC curves for the performance of the different CNN models are shown in Figure 16. The ROC curve represents the ability of the CNN models in feature extraction so that the predicted value of the recommended image matches the gold standard. The performance of the CNN models was analyzed with the area under the curve (AUC) and the corresponding p-value, as shown in Table 4. The Resnet-50 model was found to outperform other CNN models with an AUC.
The performance of the CNN models was also analyzed through FoM. The FoM is defined as the error’s central tendency and can be defined as follows:
F o M = 100 W N × 100 ,
where W is the number of images incorrectly classified according to the GT, and N is the total number of images present in the test sample. Table 5 displays the FoM values for the proposed RS with different CNN models and Maxwell–Boltzmann similarity measures in comparison to the GT. Hence, the two results of the ROC curve and FoM values found in Table 4 and Table 5 validate that the performance of the proposed RS varies according to the different CNN models used for feature extraction. This, in turn, validates the first part of our hypothesis.
We also determined the FoM values considering the four similarity measures keeping the CNN model fixed. We considered Resnet-50 as the best-performing CNN model observed from the previous results. The FoM values obtained using Resnet-50 and the four similarity measures are shown in Table 6. It was found that the FoM was at its maximum for Maxwell–Boltzmann similarity and varied according to the similarity measure used in the system. This result validates the second part of our hypothesis that the performance of the proposed RS depends upon the similarity measure used for similarity computation.

5.3. Running Time Comparison

Table 7 shows the time consumed by the proposed RS with each of the five CNN feature extraction algorithms. It is the average of multiple runs that have been expressed in seconds. The working setup to conduct experimentation is shown in Table 1. The running time of the proposed RS was calculated as the time required for the feature extraction of an input query image supplied to the proposed RS, its similarity calculation with all the images in the hospital database, and the retrieval of top-k similar images. The average running time of the proposed RS implemented with the CNN feature extraction models was compared and is represented in Figure 17. The bars in the figure represent the average running time of each RS with different CNN models. It was observed that the running time of the proposed RS is primarily dependent upon the size of the hospital database and is also affected by the type of feature extraction model.

5.4. Statistical Tests

The proposed study performed the validations of the two hypotheses designed for the proposed system. To assess the system’s reliability and stability, the standard Mann–Whitney, paired t-test, and Wilcoxon tests were used. When the distribution was not normal, the Wilcoxon test was used instead of the paired t-test to determine whether there was sufficient evidence to support the hypothesis. MedCalc software (Osteen, Belgium) was used for the statistical analysis. To validate the system proposed in the study, we provided all of the MAP@k values for k = 5 and k = 10 against various models of RS with different CNNs. The results of the Mann–Whitney, paired t-test, and Wilcoxon test are shown in Table 8.

6. Discussion

6.1. Principal Findings

The proposed study presented an image-based health RS for the efficient management of resources in hospitals such as doctors, medicine, ICUs, ventilators, and oxygen masks during the peak period of the COVID-19 pandemic. The proposed system recommends these resources to a new patient according to his or her current health condition. It is defined as a hybrid of an offline and an online system. The offline system is in charge of extracting feature vectors from images. Using a similarity measure, the online system compares the feature vectors of the image being queried and the image in the database. The top-k most similar images are then found, as shown in Figure 18.
The test was carried out on 20,000 COVID-19 patients’ chest X-ray images. The following similarity measures were used to select the best one for the system based on the AHS value: (i) cosine similarity, (ii) Maxwell–Boltzmann similarity, (iii) Euclidean similarity, and (iv) Jaccard similarity. With a similarity value of more than 94%, the Maxwell–Boltzmann similarity outperformed all other similarity measures. The proposed RS’ performance was validated using the following CNN models: (i) Resnet-50, (ii) Resnet-101, (iii) Resnet-152, (iv) VGG-16, and (v) VGG-19. The performance of the CNN models was validated using parameters such as the ROC curve and FoM value. The AUC and p-values obtained from the ROC curve indicate the ability of the CNN models to correctly predict the GT of the input image. The Resnet-50 model was found to outperform other CNN models with an AUC greater than 0.98 (p < 0.0001). The performance of the CNN models was also analyzed through FoM. The FoM was defined as the error’s central tendency. The Resnet-50 CNN model was found to have a maximum FoM value of 98.38. The performance of the similarity measures was also validated using the FoM value, and Maxwell–Boltzmann similarity outperformed the other three similarity measures; the overall performance of the proposed RS was evaluated using MAP@k. The MAP@k was determined using different CNN models for the threshold similarity in the range of 0.7 to 0.95. The proposed RS with the Resnet-50 CNN model showed the best result with a MAP@k value of 0.98014 and 0.98861 for k = 5 and k = 10, respectively. Finally, the system recommended meta-data information regarding hospital resources to a new COVID-19 patient admitted to the hospital based on his or her chest X-ray image.

6.2. Benchmarking

We considered various papers related to RS based on image similarity in our benchmarking strategy. This included Ullah et al. [17], Chen et al. [18], Tuinhof et al. [19], and Geng et al. [40]. In ref. [17], an RS based on image content was proposed and divided into two phases. The RS used a random forest classifier in the first phase to determine the product’s class or category. The system then used the JPEG coefficients measure to extract the feature vectors of the photos, which were then used to provide recommendations based on feature vector similarity in the second phase. The proposed method produced correct recommendations with a 98% accuracy rate, indicating its efficacy in real-world applications. Ref. [18] provided a neural network-based framework for product selection based on a specific input query image. A neural network was used in the proposed system to classify the supplied input query image, followed by another neural network that used the Jaccard similarity measure to determine the most comparable product image to that input image. The approach had a classification accuracy of 0.5. It offered quick and accurate online purchasing assistance and recommended products with a similarity of more than 0.5. Ref. [19] describes a two-stage deep learning framework for recommending fashion images based on similar input images. The authors proposed using a neural network classifier as a data-driven, visually aware feature extractor. The data were then fed into ranking algorithms, which generated suggestions based on similarities. The proposed method was validated using the fashion dataset, which was made public. The proposed framework, when combined with other types of content-based recommendation systems, can improve the system’s stability and effectiveness. Ref. [40] proposed a framework for combining an RS with visual product attributes by employing a deep architecture and a series of convolution operations that result in the overlapping of edges and blobs in images. The benchmarking table for the proposed study is shown in Table 9.
The proposed framework for developing an entirely image-based recommendation model compares various linear and nonlinear reduction approaches to the properties of a CNN. Ref. [82] presented an RS framework that uses chest X-ray images to predict whether a person needs COVID-19 testing. It implemented the same datasets used by the proposed method but with a different objective. None of these studies proposed any hypothesis for their proposed systems.
In contrast, we proposed two hypotheses for our system and also evaluated and validated them in the result and performance evaluation sections, respectively.

6.3. Special Note on Searching for RS

RS works on the principle of information filtering, and the searching strategy plays an important role in finding the relevant items to produce efficient and useful recommendations. The proposed RS utilizes image similarity to find the most relevant chest X-ray images with similar infections for a new COVID-19 patient with a chest X-ray image. Although CNN models play a vital role in producing accurate feature vectors, the quality of the recommendation mainly depends on the similarity measure. A proper similarity measure producing a high similarity value can produce more accurate recommendations. The four similarity measures considered for this study were analyzed based on AHS. In this study, the AHS was determined by averaging the similarity value of the most similar image to every input image present in the test set. The similarity measure with the highest AHS was considered for the RS. The performance of the proposed RS was determined in terms of MAP@k for a top-k recommendation. To identify relevant similar images for each query image, a “threshold value (T)” of similarity was also considered in the system. A retrieved database image was considered relevant when it had a similarity greater than or equal to the threshold value. This threshold value was found to affect the overall performance of the system in terms of MAP@k for a top-k recommendation.
The input images in both the training set and the testing sets were large images. These large images had many pixels to process. Further, the method we adopted reduced the computational complexity. The similarity measure strategy was very fast, quick, and low in complexity, one reason being there was no special optimization protocol and iteration adopted. Thus, overall, there was simplicity, speed, and low complexity. Such benefits overrule direct image comparison. Note that the top-n similar images obtained from the similarity computation were used for the recommendation. The proposed RS using CNN for feature extraction and similarity measurement can be an efficient tool to produce recommendations in the healthcare domain. The recommendations can be utilized for the proper allocation of doctors, medicine, and hospital resources to new patients.

6.4. Strengths, Weaknesses, and Extensions

The proposed method shows that the RS using a CNN for feature extraction and similarity measure can be an efficient tool for producing recommendations in the healthcare domain. The recommendations can be utilized for the proper allocation of doctors, medicine, and hospital resources to new patients. The proposed study proposed two hypotheses and also evaluated and validated them in the paper.
The results of the current pilot study are encouraging. However, due to the unavailability of the denoising technique in the proposed RS, the quality of the recommendation may be affected due to the presence of noise in the chest X-ray images. Denoising can be conducted in the offline and online systems. Denoising is an expensive operation in terms of computations. Therefore, offline denoising does not hurt the system that much, but the online system must be hardware interactive. The low resolution of chest X-ray images may also affect the quality of recommendations. Due to the limited number of images available for similarity calculation, a small database size may result in incorrect recommendations. A large database size may result in longer training time. While the study used basic ResNet-based systems, this can be extended to hybrid ResNet systems [83,84].
In the future, we could apply more sophisticated feature extraction techniques by fusing the different deep-learning models to achieve accurate recommendations. Better similarity methods can be explored to increase the efficiency of the proposed system. It could also be enhanced by applying segmentation techniques to make the system more robust. It can also be extended to cloud settings and big data platforms.

7. Conclusions

Through this study, we offered an RS for treating COVID-19 patients based on X-ray images of the chest. The proposed RS was divided into two phases. In phase 1, the proposed system fine-tuned the CNN models for feature extraction in phase 2. In phase 2, the finely tuned CNN model was used to extract features from both the chest X-ray of a new COVID-19 patient and the chest X-rays of COVID-19 patients present in the hospital database who were already treated successfully. The top-k similar images to the input query image of a new COVID-19 patient were determined further utilized for recommendation. In its recommendation, the proposed RS recommends doctors, medicines, and resources for new COVID-19 patients according to the metadata information of similar patients.
The proposed RS implemented with the ResNet-50 feature extraction CNN model provides the highest MAP@k with k = 5 (top-5) and k = 10 (top-10) for all the datasets with higher threshold values of similarity. The proposed RS with ResNet-50 CNN feature extraction model was found to be a proper framework for the treatment recommendation with a mean average precision (MAP) of more than 0.90 for the threshold similarities in the range of 0.7 to 0.9. The results of the proposed study were hypothesized and validated using various parameters. The proposed RS in this paper assumes that the hospital database contains related metadata, such as information about the doctors investigated, medicines, and resources allocated to a patient. The major limitation of our proposed system is that we did not consider the related physiological parameters such as sugar level, blood pressure, and other associated parameters that may affect the condition of a COVID-19 patient having similar chest infections. In the future, the proposed RS can be enhanced by considering these parameters for better recommendations.

Author Contributions

Conceptualization, M.K., P.M., J.S.S. and S.S.; Methodology and software, M.K., S.M., M.M.F. and P.M.; Validation, M.K., L.S., J.S.S. and P.M.; Investigation, M.K., P.M., J.S.S., M.M.F. and L.S.; Writing—original draft preparation, M.K., P.M., S.M. and J.S.S.; Writing—review and editing, M.K., P.M., J.S.S., M.M. and S.S.; M.K., M.M. and P.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Used data is available in Kaggle website.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

SNAbb *Definition
1AHSAverage Highest Similarity
2AUCArea-under-the-curve
3AIArtificial Intelligence
4ARDSAcute Respiratory Distress Syndrome
5BMIBody mass index
6CSSMCosine Similarity Measure
7CNNConvolution neural network
8EDSMEuclidean Distance Similarity Measure
9GGOGround Glass Opacities
10VGGVisual Geometry Group
11DLDeep learning
12DMDiabetes mellitus
13DensenetDense Convolutional Network
14MBSMMaxwell–Boltzmann Similarity Measure
15MAPMean Average Precision
16MLMachine Learning
17KNNK- nearest neighbor
18FoMFigure of Merit
19HDLHybrid deep learning
20HRSHealth Recommender System
21NBNaive Bayes
22NRNot reported
23PCAPrincipal component analysis
24PIPulmonary Infiltrates
25ResnetResidual Neural Network
26PRThe period measured in milliseconds
27RFRandom forest
28ReLURectified Linear Activation Unit
29RoBRisk of bias
30ROCReceiver operating-characteristics
31TLTransfer learning
32DFTVDataset for Training and Verification
33DFCVDataset for Training and Cross Validation

Appendix A

Table A1. Average highest cosine similarity (AHS) obtained for the CNN models with DFTV dataset.
Table A1. Average highest cosine similarity (AHS) obtained for the CNN models with DFTV dataset.
DatasetSimilarity MeasureFeature Vector Extractor Architecture
ResNet-50ResNet-101ResNet-152VGG-16VGG-19
DFTV-1CSSM0.95570.93830.92980.92330.9109
EDSM0.68340.65170.63260.63380.6319
JCSM0.43690.42100.40980.40490.4016
MBSM0.96510.95880.94890.94570.9421
DFTV-2CSSM0.95870.93760.93000.92020.9090
EDSM0.68580.65110.63740.63040.6308
JCSM0.43940.42070.41170.41250.4039
MBSM0.96760.95640.95100.94280.9392
DFTV-3CSSM0.95720.93920.93270.92160.9089
EDSM0.68500.65360.63880.63210.6306
JCSM0.43830.42280.41200.41070.4041
MBSM0.96680.95790.95330.94450.9385
DFTV-4CSSM0.95660.93810.93240.92440.9114
EDSM0.68590.65410.63760.63460.6328
JCSM0.43760.42340.41220.40860.4066
MBSM0.96700.95810.95280.94400.9419
DFTV-5CSSM0.95720.93850.93190.92380.9100
EDSM0.68540.65480.63800.63350.6318
JCSM0.43700.42300.41180.40270.4006
MBSM0.96650.95870.95300.94310.9416
Mean0.7620.7430.7330.7270.721
Standard Deviation0.2240.2270.2290.2270.225
Table A2. Average highest cosine similarity (AHS) obtained for the CNN models with DFCV dataset.
Table A2. Average highest cosine similarity (AHS) obtained for the CNN models with DFCV dataset.
DatasetSimilarity MeasureFeature Vector Extractor Architecture
ResNet-50ResNet-101ResNet-152VGG-16VGG-19
DFCV-1CSSM0.95820.93790.92810.92560.9121
EDSM0.68740.65280.63320.63470.6323
JCSM0.43840.42180.40980.40690.4011
MBSM0.96770.95800.94650.94720.9437
DFCV-2CSSM0.95700.93590.92540.92630.9103
EDSM0.68610.65140.63260.63570.6308
JCSM0.43680.42060.40660.40740.4006
MBSM0.96630.95640.94330.94820.9418
DFCV-3CSSM0.95870.93630.92800.92520.9122
EDSM0.68820.65220.63350.63450.6319
JCSM0.43700.42200.40830.40540.4018
MBSM0.96700.95590.94570.94680.9435
DFCV-4CSSM0.95650.93510.92500.92620.9104
EDSM0.68660.65190.63170.63600.6312
JCSM0.43550.42150.40580.40680.4013
MBSM0.96570.95520.94360.94740.9415
DFCV-5CSSM0.95800.93670.92730.92590.9126
EDSM0.68710.65180.63290.63520.6327
JCSM0.43620.42210.40740.40740.4027
MBSM0.96680.95670.94490.94660.9428
Mean0.7620.7420.7280.7290.722
Standard Deviation0.2250.2260.2280.2290.227
Table A3. Mean average precision (MAP) @K = 5 of CNN model.
Table A3. Mean average precision (MAP) @K = 5 of CNN model.
DatasetsThreshold Similarity (T)ResNet-50ResNet-101ResNet-152VGG-16VGG-19
DFTV-1T = 0.701110.997241
T = 0.750.99793110.995860.99655
T = 0.800.9896510.994480.990340.97724
T = 0.850.969650.986200.956550.955860.88551
T = 0.880.932410.962750.879310.877930.69724
T = 0.900.882750.867580.770340.728270.45379
T = 0.950.485510.119310.112410.035170.02551
DFTV-2T = 0.700.99931110.995860.99586
T = 0.750.99793110.995170.99172
T = 0.800.993790.994480.994480.989650.97862
T = 0.850.985510.987580.962060.942750.88896
T = 0.880.966890.953790.903440.831030.70275
T = 0.900.925510.873790.777930.691030.47379
T = 0.950.515170.147580.119310.0420680.02137
DFTV-3T = 0.7011111
T = 0.750.998621110.99379
T = 0.800.99793110.992410.97379
T = 0.850.980680.995860.971030.953100.86482
T = 0.880.955860.972410.909650.843440.68275
T = 0.900.924130.917930.801370.695170.47379
T = 0.950.509650.150340.152410.033790.02206
DFTV-4T = 0.700.99862110.998620.99862
T = 0.750.99862110.998620.99241
T = 0.800.997240.9965510.997930.97172
T = 0.850.980.986200.967580.961370.86551
T = 0.880.946890.955860.897930.868270.70413
T = 0.900.902060.899310.784820.721370.48551
T = 0.950.518620.143440.139310.029650.02482
DFTV-5T = 0.700.99586110.997241
T = 0.750.99586110.995860.99517
T = 0.800.9958610.997240.991720.97241
T = 0.850.981370.987580.967580.957240.86896
T = 0.880.946890.961370.882060.859310.70689
T = 0.900.890340.883440.786200.700680.51034
T = 0.950.513100.142060.149650.034480.02137
DFCV-1T = 0.701110.998740.99791
T = 0.750.99874110.996860.99289
T = 0.800.996650.999160.996860.992470.97262
T = 0.850.991430.986200.976170.959030.89655
T = 0.880.969070.956110.887770.886310.74670
T = 0.900.931870.894870.773240.772200.58098
T = 0.950.568230.154230.094040.066450.02884
DFCV-2T = 0.701110.999790.99811
T = 0.750.99958110.998950.99247
T = 0.800.998950.999580.997490.993310.97471
T = 0.850.987660.990380.971150.964890.89299
T = 0.880.969070.952760.889650.887140.74461
T = 0.900.932070.882340.761540.782650.53939
T = 0.950.545030.1262270.070420.056420.02549
DFCV-3T = 0.700.999160.99916110.99937
T = 0.750.998950.998950.999370.999370.99686
T = 0.800.998530.998530.998320.994980.98202
T = 0.850.989960.985370.981400.961960.88965
T = 0.880.974710.952760.908040.875020.74775
T = 0.900.948580.876280.778680.764260.56907
T = 0.950.566770.139390.082960.056210.02737
DFCV-4T = 0.701110.999581
T = 0.750.99895110.997910.99832
T = 0.800.998530.999160.998110.991220.97763
T = 0.850.992470.988290.967600.969070.89153
T = 0.880.971780.945030.877110.898220.74252
T = 0.900.927890.860600.741690.768230.54670
T = 0.950.537720.129360.070210.062690.02110
DFCV-5T = 0.7011110.99811
T = 0.751110.998740.99477
T = 0.800.9985310.998110.996440.98244
T = 0.850.992260.990170.978680.973870.91055
T = 0.880.980140.951930.899260.908880.76781
T = 0.900.950470.887560.776800.788710.57199
T = 0.950.559450.142110.074390.044720.02110
Table A4. Mean average precision (MAP) @K = 10 of CNN model.
Table A4. Mean average precision (MAP) @K = 10 of CNN model.
DatasetsThreshold Similarity (T)ResNet-50ResNet-101ResNet-152VGG-16VGG-19
DFTV-1T = 0.700.99827110.996890.99965
T = 0.750.99310110.992750.99206
T = 0.800.982750.998960.986200.986890.96206
T = 0.850.948270.980340.935860.931030.82586
T = 0.880.902060.933440.841370.819650.57241
T = 0.900.831720.807930.678960.645170.32862
T = 0.950.351030.068270.067240.017580.01275
DFTV-2T = 0.700.99655110.993440.99448
T = 0.750.995510.9982710.991030.98724
T = 0.800.989650.993790.988620.979650.96586
T = 0.850.973100.978270.947580.919310.83413
T = 0.880.941370.926200.856890.767930.59172
T = 0.900.874480.817580.699310.623790.35103
T = 0.950.402750.095170.079310.022750.01137
DFTV-3T = 0.7011110.99724
T = 0.750.997581110.98931
T = 0.800.9962010.998620.988620.95655
T = 0.850.970680.993790.960.925170.82206
T = 0.880.935510.951720.879650.778270.58724
T = 0.900.885860.864820.717580.608960.34931
T = 0.950.391720.087580.103100.016890.01103
DFTV-4T = 0.700.99758110.997580.99586
T = 0.750.99689110.997580.99034
T = 0.800.993440.994480.995510.993100.95586
T = 0.850.963790.978620.942750.941030.80931
T = 0.880.912410.938270.858620.815860.59827
T = 0.900.853440.836200.706550.633100.35172
T = 0.950.394820.093440.088620.014820.01241
DFTV-5T = 0.700.99275110.993440.99896
T = 0.750.99275110.992750.98862
T = 0.800.989650.998960.996550.987580.95206
T = 0.850.964130.983790.943100.933440.81103
T = 0.880.908960.933790.844480.803790.62068
T = 0.900.846200.834130.706200.623100.38310
T = 0.950.387930.087930.106200.017240.01068
DFCV-1T = 0.700.99947110.997910.99655
T = 0.750.99759110.995710.98819
T = 0.800.995190.998010.995500.986520.96112
T = 0.850.983280.984530.961330.942940.85862
T = 0.880.951410.939180.851930.849940.67920
T = 0.900.895400.847750.709400.723920.48192
T = 0.950.478160.095400.059660.037300.01462
DFCV-2T = 0.700.99989110.999050.99613
T = 0.750.99926110.997280.98923
T = 0.800.998220.998320.996130.988500.96426
T = 0.850.981290.983900.955170.945870.85454
T = 0.880.946810.932910.852030.857570.65956
T = 0.900.894980.841900.684010.732810.45256
T = 0.950.444820.073560.041790.031130.01400
DFCV-3T = 0.700.999050.9990510.999890.99853
T = 0.750.998950.998950.999160.998740.99362
T = 0.800.997280.997700.996650.990900.97126
T = 0.850.984950.978570.969170.944400.85464
T = 0.880.961650.931660.868540.836990.68317
T = 0.900.919740.834790.707520.709610.47513
T = 0.950.474500.083170.050570.033120.01400
DFCV-4T = 0.700.99989110.999260.99937
T = 0.750.99885110.997280.99571
T = 0.800.997910.998740.996860.987870.96468
T = 0.850.983900.979310.948060.953080.85318
T = 0.880.952240.917970.836670.860500.66112
T = 0.900.895920.816190.674500.717650.44524
T = 0.950.443570.074810.044090.034480.01076
DFCV-5T = 0.7011110.99738
T = 0.750.99979110.998320.99258
T = 0.800.9977010.997590.993100.97272
T = 0.850.988610.981500.963940.961230.87617
T = 0.880.966560.933640.866030.873980.69717
T = 0.900.918390.845350.715250.731030.47387
T = 0.950.470210.081500.046600.023190.01076

References

  1. Suri, J.S.; Puvvula, A.; Biswas, M.; Majhail, M.; Saba, L.; Faa, G.; Singh, I.M.; Oberleitner, R.; Turk, M.; Chadha, P.S. COVID-19 pathways for brain and heart injury in comorbidity patients: A role of medical imaging and artificial intelligence-based COVID severity classification: A review. Comput. Biol. Med. 2020, 124, 103960. [Google Scholar] [CrossRef] [PubMed]
  2. Suri, J.S.; Agarwal, S.; Gupta, S.K.; Puvvula, A.; Biswas, M.; Saba, L. A narrative review on characterization of acute respiratory distress syndrome in COVID-19-infected lungs using artificial intelligence. Comput. Biol. Med. 2021, 130, 104210. [Google Scholar] [CrossRef]
  3. Taquet, M.; Geddes, J.R.; Husain, M.; Luciano, S.; Harrison, P.J. 6-month neurological and psychiatric outcomes in 236 379 survivors of COVID-19: A retrospective cohort study using electronic health records. Lancet Psychiatry 2021, 8, 416–427. [Google Scholar] [CrossRef]
  4. To, K.K.-W.; Tsang, O.T.-Y.; Leung, W.-S.; Tam, A.R.; Wu, T.-C.; Lung, D.C.; Yip, C.C.-Y.; Cai, J.-P.; Chan, J.M.-C.; Chik, T.S.-H. Temporal profiles of viral load in posterior oropharyngeal saliva samples and serum antibody responses during infection by SARS-CoV-2: An observational cohort study. Lancet Infect. Dis. 2020, 20, 565–574. [Google Scholar] [CrossRef] [Green Version]
  5. Saba, L.; Gerosa, C.; Fanni, D.; Marongiu, F.; La Nasa, G.; Caocci, G.; Barcellona, D.; Coghe, F.; Orru, G.; Coni, P.; et al. Molecular pathways triggered by COVID-19 in different organs: ACE2 receptor-expressing cells under attack? A review. Eur. Rev. Med. Pharmacol. Sci. 2020, 24, 12609–12622. [Google Scholar]
  6. Viswanathan, V.; Puvvula, A.; Jamthikar, A.D.; Saba, L.; Johri, A.M.; Kotsis, V.; Khanna, N.N.; Dhanjil, S.K.; Majhail, M.; Misra, D.P. Bidirectional link between diabetes mellitus and coronavirus disease 2019 leading to cardiovascular disease: A narrative review. World J. Diabetes 2021, 12, 215–222. [Google Scholar] [CrossRef]
  7. Cau, R.; Falaschi, Z.; Paschè, A.; Danna, P.; Arioli, R.; Arru, C.D.; Zagaria, D.; Tricca, S.; Suri, J.S.; Kalra, M.K. CT findings of COVID-19 pneumonia in ICU-patients. J. Public Health Res. 2021, 10, jphr. 2021.2270. [Google Scholar] [CrossRef]
  8. Cau, R.; Pacielli, A.; Fatemeh, H.; Vaudano, P.; Arru, C.; Crivelli, P.; Stranieri, G.; Suri, J.S.; Mannelli, L.; Conti, M. Complications in COVID-19 patients: Characteristics of pulmonary embolism. Clin. Imaging 2021, 77, 244–249. [Google Scholar] [CrossRef]
  9. Congiu, T.; Demontis, R.; Cau, F.; Piras, M.; Fanni, D.; Gerosa, C.; Botta, C.; Scano, A.; Chighine, A.; Faedda, E. Scanning electron microscopy of lung disease due to COVID-19–a case report and a review of the literature. Eur. Rev. Med. Pharmacol. Sci 2021, 25, 7997–8003. [Google Scholar]
  10. Schoene, D.; Schnekenberg, L.G.; Pallesen, L.P.; Barlinn, J.; Puetz, V.; Barlinn, K.; Siepmann, T. Pathophysiology of Cardiac Injury in COVID-19 Patients with Acute Ischaemic Stroke: What Do We Know So Far?—A Review of the Current Literature. Life 2022, 12, 75. [Google Scholar]
  11. El-Rashidy, N.; Abdelrazik, S.; Abuhmed, T.; Amer, E.; Ali, F.; Hu, J.-W.; El-Sappagh, S. Comprehensive survey of using machine learning in the COVID-19 pandemic. Diagnostics 2021, 11, 1155. [Google Scholar] [CrossRef] [PubMed]
  12. Sharma, N.; Saba, L.; Khanna, N.N.; Kalra, M.K.; Fouda, M.M.; Suri, J.S. Segmentation-Based Classification Deep Learning Model Embedded with Explainable AI for COVID-19 Detection in Chest X-ray Scans. J. Diagn. 2022, 12, 2132. [Google Scholar]
  13. Khanna, N.N.; Maindarkar, M.; Puvvula, A.; Paul, S.; Bhagawati, M.; Ahluwalia, P.; Ruzsa, Z.; Sharma, A.; Munjral, S.; Kolluri, R. Vascular Implications of COVID-19: Role of Radiological Imaging, Artificial Intelligence, and Tissue Characterization: A Special Report. J. Cardiovasc. Dev. Dis. 2022, 9, 268. [Google Scholar] [CrossRef]
  14. Chatburn, R.L.; Branson, R.D. Shortages and Vulnerabilities of Hospital Oxygen Systems. Respir. Care 2022. [Google Scholar] [CrossRef] [PubMed]
  15. Mahase, E. COVID-19: Oxygen shortages two years into pandemic highlight pre-covid failures, says WHO. Br. Med. J. Publ. Group 2022, 376, o829. [Google Scholar] [CrossRef] [PubMed]
  16. DePuccio, M.J.; Gaughan, A.A.; Shiu-Yee, K.; McAlearney, A.S. Doctoring from home: Physicians’ perspectives on the advantages of remote care delivery during the COVID-19 pandemic. PLoS ONE 2022, 17, e0269264. [Google Scholar] [CrossRef]
  17. Cau, R.; Faa, G.; Nardi, V.; Balestrieri, A.; Puig, J.; Suri, J.S.; SanFilippo, R.; Saba, L. Long-COVID diagnosis: From diagnostic to advanced AI-driven models. Eur. J. Radiol. 2022, 148, 110164. [Google Scholar] [CrossRef]
  18. Munjral, S.; Maindarkar, M.; Ahluwalia, P.; Puvvula, A.; Jamthikar, A.; Jujaray, T.; Suri, N.; Paul, S.; Pathak, R.; Saba, L. Cardiovascular Risk Stratification in Diabetic Retinopathy via Atherosclerotic Pathway in COVID-19/Non-COVID-19 Frameworks Using Artificial Intelligence Paradigm: A Narrative Review. Diagnostics 2022, 12, 1234. [Google Scholar] [CrossRef]
  19. Chandra, T.B.; Verma, K.; Singh, B.K.; Jain, D.; Netam, S.S. Coronavirus disease (COVID-19) detection in chest X-ray images using majority voting based classifier ensemble. Expert Syst. Appl. 2021, 165, 113909. [Google Scholar] [CrossRef]
  20. Liang, S.; Liu, H.; Gu, Y.; Guo, X.; Li, H.; Li, L.; Wu, Z.; Liu, M.; Tao, L. Fast automated detection of COVID-19 from medical images using convolutional neural networks. Commun. Biol. 2021, 4, 35. [Google Scholar] [CrossRef]
  21. Suri, J.S.; Maindarkar, M.A.; Paul, S.; Ahluwalia, P.; Bhagawati, M.; Saba, L.; Faa, G.; Saxena, S.; Singh, I.M.; Chadha, P.S. Deep Learning Paradigm for Cardiovascular Disease/Stroke Risk Stratification in Parkinson’s Disease Affected by COVID-19: A Narrative Review. J. Diagn. 2022, 12, 1543. [Google Scholar] [CrossRef] [PubMed]
  22. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef] [PubMed]
  23. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  24. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [PubMed]
  25. Ullah, F.; Zhang, B.; Khan, R.U. Image-based service recommendation system: A JPEG-coefficient RFs approach. IEEE Access 2019, 8, 3308–3318. [Google Scholar] [CrossRef]
  26. Chen, L.; Yang, F.; Yang, H. Image-Based Product Recommendation System with Convolutional Neural Networks. Stanford University. Available online: http://cs231n.stanford.edu/reports/2017/pdfs/105.pdf:2017 (accessed on 12 January 2021).
  27. Tuinhof, H.; Pirker, C.; Haltmeier, M. Image-based fashion product recommendation with deep learning. In Proceedings of the International Conference on Machine Learning, Optimization, and Data Science, Volterra, Italy, 3–16 September 2018; pp. 472–481. [Google Scholar]
  28. Saba, L.; Agarwal, M.; Patrick, A.; Puvvula, A.; Gupta, S.K.; Carriero, A.; Laird, J.R.; Kitas, G.D.; Johri, A.M.; Balestrieri, A. Six artificial intelligence paradigms for tissue characterisation and classification of non-COVID-19 pneumonia against COVID-19 pneumonia in computed tomography lungs. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 423–434. [Google Scholar] [CrossRef]
  29. Zimmerman, A.; Kalra, D. Usefulness of machine learning in COVID-19 for the detection and prognosis of cardiovascular complications. Rev. Cardiovasc. Med. 2020, 21, 345–352. [Google Scholar] [CrossRef]
  30. Hussain, L.; Nguyen, T.; Li, H.; Abbasi, A.A.; Lone, K.J.; Zhao, Z.; Zaib, M.; Chen, A.; Duong, T.Q. Machine-learning classification of texture features of portable chest X-ray accurately classifies COVID-19 lung infection. BioMedical Eng. OnLine 2020, 19, 88. [Google Scholar] [CrossRef]
  31. Agarwal, M.; Agarwal, S.; Saba, L.; Chabert, G.L.; Gupta, S.; Carriero, A.; Pasche, A.; Danna, P.; Mehmedovic, A.; Faa, G. Eight pruning deep learning models for low storage and high-speed COVID-19 computed tomography lung segmentation and heatmap-based lesion localization: A multicenter study using COVLIAS 2.0. Comput. Biol. Med. 2022, 105571. [Google Scholar] [CrossRef]
  32. Hasan, M.M.; Murtaz, S.B.; Islam, M.U.; Sadeq, M.J.; Uddin, J. Robust and efficient COVID-19 detection techniques: A machine learning approach. PLoS ONE 2022, 17, e0274538. [Google Scholar] [CrossRef]
  33. Yang, D.; Martinez, C.; Visuña, L.; Khandhar, H.; Bhatt, C.; Carretero, J. Detection and analysis of COVID-19 in medical images using deep learning techniques. Sci. Rep. 2021, 11, 19638. [Google Scholar] [CrossRef] [PubMed]
  34. Silva, P.; Luz, E.; Silva, G.; Moreira, G.; Silva, R.; Lucio, D.; Menotti, D. COVID-19 detection in CT images with deep learning: A voting-based scheme and cross-datasets analysis. Inform. Med. Unlocked 2020, 20, 100427. [Google Scholar] [CrossRef] [PubMed]
  35. Diaz-Escobar, J.; Ordóñez-Guillén, N.E.; Villarreal-Reyes, S.; Galaviz-Mosqueda, A.; Kober, V.; Rivera-Rodriguez, R.; Rizk, J.E.L. Deep-learning based detection of COVID-19 using lung ultrasound imagery. PLoS ONE 2021, 16, e0255886. [Google Scholar] [CrossRef] [PubMed]
  36. Teodoro, A.A.; Silva, D.H.; Saadi, M.; Okey, O.D.; Rosa, R.L.; Otaibi, S.A.; Rodríguez, D.Z. An analysis of image features extracted by cnns to design classification models for covid-19 and non-COVID-19. J. Signal Process. Syst. 2021, 1–13. [Google Scholar] [CrossRef] [PubMed]
  37. Rahaman, M.M.; Li, C.; Yao, Y.; Kulwa, F.; Rahman, M.A.; Wang, Q.; Qi, S.; Kong, F.; Zhu, X.; Zhao, X. Identification of COVID-19 samples from chest X-Ray images using deep learning: A comparison of transfer learning approaches. J. X-Ray Sci. Technol. 2020, 28, 821–839. [Google Scholar] [CrossRef]
  38. Saiz, F.; Barandiaran, I. COVID-19 detection in chest X-ray images using a deep learning approach. 2020; 112–118. [Google Scholar]
  39. Samy, B.; Al-Hamadi, A. Automatic detection of COVID-19 using pruned GLCM-Based texture features and LDCRF classification. Comput. Biol. Med. 2021, 37, 104781. [Google Scholar]
  40. Tello-Mijares, S.; Woo, L. Computed tomography image processing analysis in COVID-19 patient follow-up assessment. J. Healthc. Eng. 2021, 2021, 8869372. [Google Scholar] [CrossRef]
  41. Subramaniam, U.; Subashini, M.M.; Almakhles, D.; Karthick, A.; Manoharan, S. An expert system for COVID-19 infection tracking in lungs using image processing and deep learning techniques. BioMed Res. Int. 2021, 2021, 123–127. [Google Scholar] [CrossRef]
  42. Ahsan, M.; Based, M.A.; Haider, J.; Kowalski, M. COVID-19 detection from chest X-ray images using feature fusion and deep learning. Sensors 2021, 21, 1480. [Google Scholar]
  43. Zhou, C.; Song, J.; Zhou, S.; Zhang, Z.; Xing, J. COVID-19 detection based on image regrouping and ResNet-SVM using chest X-ray images. IEEE Access 2021, 9, 81902–81912. [Google Scholar] [CrossRef]
  44. Chowdhury, M.E.; Rahman, T.; Khandakar, A.; Mazhar, R.; Kadir, M.A.; Mahbub, Z.B.; Islam, K.R.; Khan, M.S.; Iqbal, A.; Al Emadi, N. Can AI help in screening viral and COVID-19 pneumonia? IEEE Access 2020, 8, 132665–132676. [Google Scholar] [CrossRef]
  45. Tahir, A.M.; Chowdhury, M.E.; Khandakar, A.; Rahman, T.; Qiblawey, Y.; Khurshid, U.; Kiranyaz, S.; Ibtehaz, N.; Rahman, M.S.; Al-Maadeed, S. COVID-19 infection localization and severity grading from chest X-ray images. Comput. Biol. Med. 2021, 139, 105002. [Google Scholar] [CrossRef]
  46. Othman, M.; Zain, N.M.; Paidi, Z.; Pauzi, F.A. Framework of Health Recommender System for COVID-19 Self-assessment and Treatments: A Case Study in Malaysia. Int. J. Comput. Sci. Netw. Secur. 2021, 21, 12–18. [Google Scholar]
  47. Nilashi, M.; Asadi, S.; Abumalloh, R.A.; Samad, S.; Ibrahim, O. Intelligent recommender systems in the COVID-19 outbreak: The case of wearable healthcare devices. J. Soft Comput. Decis. Support Syst. 2020, 7, 8–12. [Google Scholar]
  48. Asabere, N.Y.; Acakpovi, A.; Ofori, E.K.; Torgby, W.; Kuuboore, M.; Lawson, G.; Adjaloko, E. SARPPIC: Exploiting COVID-19 Contact Tracing Recommendation through Social Awareness. Comput. Math. Methods Med. 2020, 2020, 3460130. [Google Scholar] [CrossRef] [PubMed]
  49. Deng, F.; Ren, P.; Qin, Z.; Huang, G.; Qin, Z. Leveraging Image Visual Features in Content-Based Recommender System. Sci. Program. 2018, 2018, 5497070. [Google Scholar] [CrossRef] [Green Version]
  50. Sulthana, A.R.; Gupta, M.; Subramanian, S.; Mirza, S. Improvising the performance of image-based recommendation system using convolution neural networks and deep learning. Soft Comput. 2020, 24, 14531–14544. [Google Scholar] [CrossRef]
  51. Geng, X.; Zhang, H.; Bian, J.; Chua, T.-S. Learning image and user features for recommendation in social networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4274–4282. [Google Scholar]
  52. Kuanr, M.; Mohapatra, P. Recent challenges in recommender systems: A survey. In Progress in Advanced Computing and Intelligent Engineering; Springer: Berlin/Heidelberg, Germany, 2021; pp. 353–365. [Google Scholar]
  53. Walek, B.; Spackova, P. Content-based recommender system for online stores using expert system. In Proceedings of the 2018 IEEE First International Conference on Artificial Intelligence and Knowledge Engineering (AIKE), Laguna Hills, CA, USA, 26–28 September 2018; pp. 164–165. [Google Scholar]
  54. Pal, A.; Parhi, P.; Aggarwal, M. An improved content based collaborative filtering algorithm for movie recommendations. In Proceedings of the 2017 Tenth International Conference on Contemporary Computing (IC3), Noida, India, 10–12 August 2017; pp. 1–3. [Google Scholar]
  55. Yu, C.; Tang, Q.; Liu, Z.; Dong, B.; Wei, Z. A recommender system for ordering platform based on an improved collaborative filtering algorithm. In Proceedings of the 2018 International Conference on Audio, Language and Image Processing (ICALIP), Shanghai, China, 16–17 July 2018; pp. 298–302. [Google Scholar]
  56. Shakirova, E. Collaborative filtering for music recommender system. In Proceedings of the 2017 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus), Moscow, Russia, 2–7 February 2017; pp. 548–550. [Google Scholar]
  57. Devika, R.; Subramaniyaswamy, V. A novel model for hospital recommender system using hybrid filtering and big data techniques. In Proceedings of the 2018 2nd International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud) (I-SMAC) I-SMAC (IoT in Social, Mobile, Analytics and Cloud)(I-SMAC), Palladam, India, 30–31 August 2018; pp. 267–271. [Google Scholar]
  58. Kbaier, M.E.B.H.; Masri, H.; Krichen, S. A personalized hybrid tourism recommender system. In Proceedings of the 2017 IEEE/ACS 14th International Conference on Computer Systems and Applications (AICCSA), Hammamet, Tunisia, 30 October–3 November 2017; pp. 244–250. [Google Scholar]
  59. Subbotin, S.; Gladkova, O.; Parkhomenko, A. Knowledge-based recommendation system for embedded systems platform-oriented design. In Proceedings of the 2018 IEEE 13th International Scientific and Technical Conference on Computer Sciences and Information Technologies (CSIT), Lviv, Ukraine, 11–14 September 2018; pp. 368–373. [Google Scholar]
  60. Wonoseto, M.G.; Rosmansyah, Y. Knowledge based recommender system and web 2.0 to enhance learning model in junior high school. In Proceedings of the 2017 International Conference on Information Technology Systems and Innovation (ICITSI), Bandung, Indonesia, 23–24 October 2017; pp. 168–171. [Google Scholar]
  61. Wiesner, M.; Pfeifer, D. Health recommender systems: Concepts, requirements, technical basics and challenges. Int. J. Environ. Res. Public Health 2014, 11, 2580. [Google Scholar] [CrossRef]
  62. Kuanr, M.; Mohapatra, P.; Piri, J. Health recommender system for cervical cancer prognosis in women. In Proceedings of the 2021 6th International Conference on Inventive Computation Technologies (ICICT), Lalitpur, Nepal, 20–22 July 2022; pp. 673–679. [Google Scholar]
  63. Tarasenko, A.O.; Yakimov, Y.V. Convolutional neural networks for image classification. 2020; 12, 123–128. [Google Scholar]
  64. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  65. Pathak, Y.; Shukla, P.K.; Tiwari, A.; Stalin, S.; Singh, S. Deep transfer learning based classification model for COVID-19 disease. J. Irbm 2020, 43, 87–92. [Google Scholar] [CrossRef]
  66. Suri, J.S.; Agarwal, S.; Elavarthi, P.; Pathak, R.; Ketireddy, V.; Columbu, M.; Saba, L.; Gupta, S.K.; Faa, G.; Singh, I.M. Inter-Variability Study of COVLIAS 1.0: Hybrid Deep Learning Models for COVID-19 Lung Segmentation in Computed Tomography. J. Diagn. 2021, 11, 2025. [Google Scholar] [CrossRef]
  67. Skandha, S.S.; Nicolaides, A.; Gupta, S.K.; Koppula, V.K.; Saba, L.; Johri, A.M.; Kalra, M.S.; Suri, J.S. A hybrid deep learning paradigm for carotid plaque tissue characterization and its validation in multicenter cohorts using a supercomputer framework. J. Comput. Biol. Med. 2021, 105–119. [Google Scholar] [CrossRef] [PubMed]
  68. Chu, Y.; Yue, X.; Yu, L.; Sergei, M.; Wang, Z. Automatic image captioning based on ResNet50 and LSTM with soft attention. Wirel. Commun. Mob. Comput. 2020, 2020, 8909458. [Google Scholar] [CrossRef]
  69. Zhang, Q. A novel ResNet101 model based on dense dilated convolution for image classification. SN Appl. Sci. 2022, 4, 9. [Google Scholar] [CrossRef]
  70. Pustokhin, D.A.; Pustokhina, I.V.; Dinh, P.N.; Phan, S.V.; Nguyen, G.N.; Joshi, G.P. An effective deep residual network based class attention layer with bidirectional LSTM for diagnosis and classification of COVID-19. J. Appl. Stat. 2020, 1–18. [Google Scholar] [CrossRef]
  71. Jiang, Z.-P.; Liu, Y.-Y.; Shao, Z.-E.; Huang, K.-W. An Improved VGG16 Model for Pneumonia Image Classification. Appl. Sci. 2021, 11, 11185. [Google Scholar] [CrossRef]
  72. Mateen, M.; Wen, J.; Song, S.; Huang, Z. Fundus image classification using VGG-19 architecture with PCA and SVD. Symmetry 2018, 11, 1. [Google Scholar] [CrossRef] [Green Version]
  73. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  74. Kaur, S.; Aggarwal, D. Image content based retrieval system using cosine similarity for skin disease images. Adv. Comput. Sci. Int. J. 2013, 2, 89–95. [Google Scholar]
  75. Kuppili, V.; Biswas, M.; Edla, D.R.; Prasad, K.R.; Suri, J.S. A mechanics-based similarity measure for text classification in machine learning paradigm. IEEE Trans. Emerg. Top. Comput. Intell. 2018, 4, 180–200. [Google Scholar] [CrossRef]
  76. Khoshdeli, M.; Cong, R.; Parvin, B. Detection of nuclei in H&E stained sections using convolutional neural networks. In Proceedings of the 2017 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Orland, FL, USA, 16–19 February 2017; pp. 105–108. [Google Scholar]
  77. Jain, P.K.; Sharma, N.; Saba, L.; Paraskevas, K.I.; Kalra, M.K.; Johri, A.; Laird, J.R.; Nicolaides, A.N.; Suri, J.S. Unseen artificial intelligence—Deep learning paradigm for segmentation of low atherosclerotic plaque in carotid ultrasound: A multicenter cardiovascular study. J. Diagn. 2021, 11, 2257. [Google Scholar] [CrossRef]
  78. Sanagala, S.S.; Nicolaides, A.; Gupta, S.K.; Koppula, V.K.; Saba, L.; Agarwal, S.; Johri, A.M.; Kalra, M.S.; Suri, J.S. Ten Fast Transfer Learning Models for Carotid Ultrasound Plaque Tissue Characterization in Augmentation Framework Embedded with Heatmaps for Stroke Risk Stratification. J. Diagn. 2021, 11, 2109. [Google Scholar] [CrossRef] [PubMed]
  79. Agarwal, M.; Saba, L.; Gupta, S.K.; Johri, A.M.; Khanna, N.N.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; Miner, M.; Sfikakis, P.P. Wilson disease tissue classification and characterization using seven artificial intelligence models embedded with 3D optimization paradigm on a weak training brain magnetic resonance imaging datasets: A supercomputer application. J. Med. Biol. Eng. Comput. 2021, 59, 511–533. [Google Scholar] [CrossRef]
  80. DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar]
  81. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Kashem, S.B.A.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S. Exploring the effect of image enhancement techniques on COVID-19 detection using chest X-ray images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef] [PubMed]
  82. Chen, T.; Wu, D.; Chen, H.; Yan, W.; Yang, D.; Chen, G.; Ma, K.; Xu, D.; Yu, H.; Wang, H. Clinical characteristics of 113 deceased patients with coronavirus disease 2019: Retrospective study. J. BMJ 2020, 368, m1091. [Google Scholar] [CrossRef] [Green Version]
  83. Dufan, W.; Gong, K.; Arru, C.D.; Homayounieh, F.; Bizzo, B.; Buch, V.; Ren, H.; Kim, K.; Neumark, N.; Xu, P.; et al. Severity and consolidation quantification of COVID-19 from CT images using deep learning based on hybrid weak labels. IEEE J. Biomed. Health Inform. 2020, 24, 3529–3538. [Google Scholar]
  84. Suri, J.S.; Agarwal, S.; Chabert, G.L.; Carriero, A.; Paschè, A.; Danna, P.S.C.; Saba, L.; Mehmedović, A.; Faa, G.; Singh, I.M.; et al. COVLIAS 1.0Lesion vs. MedSeg: An Artificial Intelligence Framework for Automated Lesion Segmentation in COVID-19 Lung Computed Tomography Scans. Diagnostics 2022, 12, 1283. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Evolution of an X-ray image for a COVID-19 patient over days (1 (A), 3 (B), 6 (C), 7 (D), 8 (E), and 10 (F), respectively) [11] (reproduced with permission).
Figure 1. Evolution of an X-ray image for a COVID-19 patient over days (1 (A), 3 (B), 6 (C), 7 (D), 8 (E), and 10 (F), respectively) [11] (reproduced with permission).
Diagnostics 12 02700 g001
Figure 2. The global treatment recommender system for COVID-19.
Figure 2. The global treatment recommender system for COVID-19.
Diagnostics 12 02700 g002
Figure 3. Various types of recommender systems.
Figure 3. Various types of recommender systems.
Diagnostics 12 02700 g003
Figure 4. The basic architecture of convolutional neural network [22] (reproduced with permission).
Figure 4. The basic architecture of convolutional neural network [22] (reproduced with permission).
Diagnostics 12 02700 g004
Figure 5. The basic architecture of ResNet.
Figure 5. The basic architecture of ResNet.
Diagnostics 12 02700 g005
Figure 6. ResNet with residual block.
Figure 6. ResNet with residual block.
Diagnostics 12 02700 g006
Figure 7. The basic architecture of VGG Net [73] (reproduced with permission); CONV: convolution layer and FC: fully connected network.
Figure 7. The basic architecture of VGG Net [73] (reproduced with permission); CONV: convolution layer and FC: fully connected network.
Diagnostics 12 02700 g007
Figure 8. The architecture or local system for the proposed system.
Figure 8. The architecture or local system for the proposed system.
Diagnostics 12 02700 g008
Figure 9. Phase-1 of the proposed system.
Figure 9. Phase-1 of the proposed system.
Diagnostics 12 02700 g009
Figure 10. The data flow of the proposed system.
Figure 10. The data flow of the proposed system.
Diagnostics 12 02700 g010
Figure 11. Performance evaluation of the fine-tuned CNN models.
Figure 11. Performance evaluation of the fine-tuned CNN models.
Diagnostics 12 02700 g011
Figure 12. Graph for mean/standard deviation of AHS.
Figure 12. Graph for mean/standard deviation of AHS.
Diagnostics 12 02700 g012
Figure 13. Graph for a mean of AHS.
Figure 13. Graph for a mean of AHS.
Diagnostics 12 02700 g013
Figure 14. Mean average precision @k graphs for DFTV datasets for k = 5 and k = 10.
Figure 14. Mean average precision @k graphs for DFTV datasets for k = 5 and k = 10.
Diagnostics 12 02700 g014
Figure 15. Mean Average precision @k graphs for DFCV datasets for k = 5 and k = 10.
Figure 15. Mean Average precision @k graphs for DFCV datasets for k = 5 and k = 10.
Diagnostics 12 02700 g015
Figure 16. ROC curves for the performance analysis of CNN models.
Figure 16. ROC curves for the performance analysis of CNN models.
Diagnostics 12 02700 g016
Figure 17. Graphs for average running time comparison.
Figure 17. Graphs for average running time comparison.
Diagnostics 12 02700 g017
Figure 18. Top 10 similar images for recommendation.
Figure 18. Top 10 similar images for recommendation.
Diagnostics 12 02700 g018
Table 1. Computer resources.
Table 1. Computer resources.
Software and HardwareSpecifications
Operating SystemDebian GNU/Linux 9 (stretch)
CPUIntel(R) Xeon(R) CPU @ 2.00GHz
GPUGPU—Tesla P100 16GB
LanguagePython 3.9
RAM16 GB
Disk645 GB
Table 2. Description of datasets.
Table 2. Description of datasets.
DFTVDFCV
Dataset NameCOVID-19 Radiography DatabaseCOVID-QU-Ex Dataset
Dataset linkLink1 [81]Link2 [45]
Dataset size781 MB329 MB
Dimensions299 × 299256 × 256
Number of Images21,16511,956
Number of Classes41
DFTV: Data set for Training and verification, DFCV: Data set for Cross Validation.
Table 3. Mean average precision @5 for DFTV and DFCV datasets using the Resnet-50 CNN model.
Table 3. Mean average precision @5 for DFTV and DFCV datasets using the Resnet-50 CNN model.
DatasetsEDSMJCSMCSSMMBSM
DFTV-10.501270.623460.917240.94827
DFTV-20.520370.636580.935860.9731
DFTV-30.543860.662150.950340.97068
DFTV-40.498730.613840.925860.95224
DFTV-50.500120.619310.937930.96413
DFCV-10.532470.640340.952820.98328
DFCV-20.497710.612410.935170.98129
DFCV-30.510230.620110.943870.98495
DFCV-40.533870.632860.956720.9839
DFCV-50.532180.652380.964580.98861
Table 4. AUC and p-value for the CNN models.
Table 4. AUC and p-value for the CNN models.
SNCNN ModelAUCp-Value
1VGG-190.915p < 0.001
2VGG-160.968p < 0.001
3Resnet-1520.977p < 0.001
4Resnet-1010.978p < 0.001
5Resnet-500.980p < 0.001
Table 5. FoM was observed for the CNN models.
Table 5. FoM was observed for the CNN models.
SNCNN ModelFoM (%)
1VGG-1991.12
2VGG-1696.23
3Resnet-15297.04
4Resnet-10197.84
5Resnet-5098.38
Table 6. FoM was observed for the similarity measures.
Table 6. FoM was observed for the similarity measures.
SNCNN ModelFoM (%)
1JCSM42.74
2EDSM58.39
3CSSM97.26
4MBSM98.38
Table 7. Average running time obtained for the CNN models.
Table 7. Average running time obtained for the CNN models.
Average Running Time (in Seconds)
DatasetsResNet-50ResNet-101ResNet-152VGG-16VGG-19
DFTV0.0370.0630.0610.0410.034
DFCV0.0980.1010.1030.0940.092
Table 8. Results of statistical tests.
Table 8. Results of statistical tests.
ModelsMann–WhitneyPaired t-TestWilcoxon Test
M1 vs. M2p < 0.0001p < 0.0001p < 0.0001
M1 vs. M3p < 0.0001p < 0.0001p < 0.0001
M1 vs. M4p < 0.0001p < 0.0001p < 0.0001
M1 vs. M5p < 0.0001p < 0.0001p < 0.0001
M2 vs. M3p < 0.0001p < 0.0001p < 0.0001
M2 vs. M4p < 0.0001p < 0.0001p < 0.0001
M2 vs. M5p < 0.0001p < 0.0001p < 0.0001
M3 vs. M4p < 0.0001p < 0.0001p < 0.0001
M3 vs. M5p < 0.0001p < 0.0001p < 0.0001
M4 vs. M5p < 0.0001p < 0.0001p < 0.0001
M1—RS using Resnet-50, M2—RS using Resnet-101, M3—RS using Resnet-152, M4—RS using VGG-16, M5—RS using VGG-19.
Table 9. Benchmarking table.
Table 9. Benchmarking table.
C1C2C3C4C5C6C7C8
SNAuthorNumber of ImagesTechniqueModel TypesACCMAPLossAHS
1Ullah et al. [25]2000Feature Extraction by JPEG Coefficient and Classification by RFCosineNRNRNR0.967
EuclideanNRNRNR0.961
SubjectiveNRNRNR0.93
2Chen et al. [82]10,000Neural Network and Jaccard SimilaritySVM0.26790NRNRNR
Alexnet0.39460NRNRNR
VGG0.5010NRNRNR
3Tuinhof et al. [27]11,851Feature Extraction and ranking by KNNAlexnetNRNR1.48NR
BN-InceptionNRNR1.27NR
4Geng et al. [51]686,457Feature LearningCBFNR0.098NRNR
UCFNR0.308NRNR
ICFNR0.338NRNR
WMFNR0.356NRNR
DWNR0.550NRNR
DUIFNR0.457NRNR
5Kuanr et al. (proposed)20,000Feature Extraction and Similarity by Maxwell–Boltzmann SimilarityVGG-1991.120.853181.390.9428
VGG-1696.230.953081.210.9466
Resnet-15297.040.948061.260.9449
Resnet-10197.840.979311.170.9567
Resnet-5098.380.9843751.090.9668
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kuanr, M.; Mohapatra, P.; Mittal, S.; Maindarkar, M.; Fouda, M.M.; Saba, L.; Saxena, S.; Suri, J.S. Recommender System for the Efficient Treatment of COVID-19 Using a Convolutional Neural Network Model and Image Similarity. Diagnostics 2022, 12, 2700. https://doi.org/10.3390/diagnostics12112700

AMA Style

Kuanr M, Mohapatra P, Mittal S, Maindarkar M, Fouda MM, Saba L, Saxena S, Suri JS. Recommender System for the Efficient Treatment of COVID-19 Using a Convolutional Neural Network Model and Image Similarity. Diagnostics. 2022; 12(11):2700. https://doi.org/10.3390/diagnostics12112700

Chicago/Turabian Style

Kuanr, Madhusree, Puspanjali Mohapatra, Sanchi Mittal, Mahesh Maindarkar, Mostafa M. Fouda, Luca Saba, Sanjay Saxena, and Jasjit S. Suri. 2022. "Recommender System for the Efficient Treatment of COVID-19 Using a Convolutional Neural Network Model and Image Similarity" Diagnostics 12, no. 11: 2700. https://doi.org/10.3390/diagnostics12112700

APA Style

Kuanr, M., Mohapatra, P., Mittal, S., Maindarkar, M., Fouda, M. M., Saba, L., Saxena, S., & Suri, J. S. (2022). Recommender System for the Efficient Treatment of COVID-19 Using a Convolutional Neural Network Model and Image Similarity. Diagnostics, 12(11), 2700. https://doi.org/10.3390/diagnostics12112700

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop