Next Article in Journal
Spectral and Oscillation Theory for an Unconventional Fractional Sturm–Liouville Problem
Previous Article in Journal
On Solutions of Two Post-Quantum Fractional Generalized Sequential Navier Problems: An Application on the Elastic Beam
Previous Article in Special Issue
Fractional Fuzzy Neural System: Fractional Differential-Based Compensation Prediction for Reputation Infringement Cases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

COVID-19 Diagnosis by Extracting New Features from Lung CT Images Using Fractional Fourier Transform

Medical Sciences & Technologies Faculty, Science and Research Branch, Islamic Azad University, Tehran 1477893855, Iran
*
Author to whom correspondence should be addressed.
Fractal Fract. 2024, 8(4), 237; https://doi.org/10.3390/fractalfract8040237
Submission received: 2 May 2023 / Revised: 8 June 2023 / Accepted: 16 June 2023 / Published: 18 April 2024

Abstract

:
COVID-19 is a lung disease caused by a coronavirus family virus. Due to its extraordinary prevalence and associated death rates, it has spread quickly to every country in the world. Thus, achieving peaks and outlines and curing different types of relapses is extremely important. Given the worldwide prevalence of coronavirus and the participation of physicians in all countries, information has been gathered regarding the properties of the virus, its diverse types, and the means of analyzing it. Numerous approaches have been used to identify this evolving virus. It is generally considered the most accurate and acceptable method of examining the patient’s lungs and chest through a CT scan. As part of the feature extraction process, a method known as fractional Fourier transform (FrFT) has been applied as one of the time-frequency domain transformations. The proposed method was applied to a database consisting of 2481 CT images. Following the transformation of all images into equal sizes and the removal of non-lung areas, multiple combination windows are used to reduce the number of features extracted from the images. In this paper, the results obtained for KNN and SVM classification have been obtained with accuracy values of 99.84% and 99.90%, respectively.

1. Introduction

Globally, over 660 million patients had been diagnosed with COVID-19 infection by January 2023, and about 6.6 million died from COVID-19. It is important to screen for COVID-19 as soon as possible to prevent its spread through separation and medical treatment. COVID-19 is an infection initiated via a virus called SARS-CoV-2. As a matter of fact, there are many different types of lung infections, which can range from a simple cold to a fatal condition. Coronaviruses are known to cause respiratory system symptoms, frequently mistaken for contagions caused by other viruses. In unusual cases, individuals may have slight, self-restraining infections with adversative properties such as influenza. There are several symptoms of respiratory concerns, including fever, coughing, and trouble breathing, as well as exhaustion, a sore throat, and a weak immune system [1,2,3,4,5]. The use of CT scans and X-rays is without a doubt one of the most important methods for the detection of COVID-19. There is a consensus within the medical community that chest imaging is a quick and effective method of diagnosis. Numerous papers have pointed out that it is the first screening tool for widespread use. Several different computer vision techniques are used to segment and classify objects. As an alternative to a computer-based automatic method that allows the progression of patients infected by measuring the infection region and disintegrating it every two to six days, a computer-based automatic method that can run on partial computing devices is a great solution when the task is quick and straightforward. Medical recognition is essential, as COVID-19 is a challenging syndrome for medical professionals to detect [6,7,8,9].
The WHO called it a major universal public health crisis, and the situation continued to pose a public health issue that necessitated substantial coordination. COVID-19 can be reported automatically through CT scans, making it an excellent way to extend traditional healthcare methods. However, CT has several limitations [10,11]. It should be noted that despite this, COVID-19 cases continue to rise sharply. COVID-19 patients represent a significant number of patients that must be detected and monitored manually, which is challenging. Hence, one of the most significant responsibilities of stopping the spread of COVID-19 is to come up with a fast and precise method of diagnosing the disease. In the past few years, research into artificial intelligence has grown in popularity, attracting scholars to solve complex problems in many areas, such as medicine, economics, and cyber security. Artificial intelligence offers significant advantages by substituting humans with machines capable of executing repetitive and complex tasks, resulting in notable benefits [12,13,14,15,16,17]. Artificial intelligence (AI) can solve the analytical issues relating to a rapid increase in patients. Many researchers believe machine learning can successfully detect COVID-19 patients using medical images. The development of machine learning techniques for chest CT and X-ray images of COVID-19 patients has been studied, and some have achieved outstanding results. Moreover, many innovative and inspirational image-processing algorithms have been developed for the detection of COVID-19 [18].
Other important diagnostic tools, including radiological imaging, can also be used for COVID-19. It is commonly observed that CT images of COVID-19 patients demonstrate ground-glass opacities in the early stages of the disease and consolidation of the lungs in the later stages of the disease. The morphology of the lung may also be rounded, and there may be a peripheral lung distribution [19,20,21,22,23,24,25,26,27].
In situations where the patient is suspected of having viral pneumonia, CT scans can provide an early indication of infection. While there are several causes of viral pneumonia, the images of the various causes are similar. These images also overlap with those seen in other inflammatory and infectious lung diseases. Since COVID-19 cannot be distinguished from other viral pneumonia as a result of this overlap, radiologists have difficulty diagnosing the disease. As the gold standard for diagnosing viral and bacterial infections at the molecular level, the RT-PCR method is used by most health authorities around the world to diagnose viral and bacterial infections [28]. To lower the disease curve, early detection and mass testing are necessary. To fulfill the demands on the healthcare system, all infrastructure must be expanded with the rapidly increasing number of newly diagnosed illnesses. It has been found that chest computed tomography can be useful in the early detection of the disease. In certain instances, the initial PCR test results for the patient indicated a negative outcome. However, confirmation relied on the findings from their CT scan. Moreover, a chest CT screening was advised as the patient displayed symptoms consistent with COVID-19, despite the negative results of the PCR tests [29,30,31]. Automatic detection tools will be key to preventing the spread of disease and speeding up the diagnosis process to prevent further spread and, at the same time, ensure that CT images are available to construct an AI-based device to prevent their spread. Several attempts have been made to identify alternative testing tools for COVID-19 infection to alleviate the shortage and inefficiency of current tests. Researchers have demonstrated that CT scans highly indicate COVID-19’s radiological features. CT scanners are readily available and can be used by a wide range of medical professionals, making them an efficient and useful tool for testing.
This paper investigates COVID-19 detection from CT images by fractional Fourier transform. The FrFT is a generalized form of the Fourier transform, considered a time-frequency transform, unlike the Fourier transform, which only provides frequency information. The features extracted in this study are fractional Fourier transform coefficients, which are complex numbers, obtained by applying FRFT to the image.
An overview of the literature is presented in the following section. The next section provides an overview of the proposed database. After describing the database, the next section describes the proposed method, pre-processing, feature selection, and classification. In the following section, we summarize the results of the experiments. As a final section, the main conclusions of the study are presented.

2. Literature Review

In the medical imaging field, artificial intelligence has been primarily introduced to provide improved quality and efficiency of clinical care in response to the need for better clinical treatments. It is widely believed that the amount of radiology imaging data is growing much faster than the number of qualified readers. Due to this fact, healthcare professionals are constantly required to improve their efficiency in the way they analyze images to compensate for this lack of efficiency. Kaur and Ghandi conducted a study on COVID-19 detection using transfer learning and investigated various pre-trained network architectures suitable for small medical imaging datasets. Specifically, they explored different variants of the pre-trained ResNet model, including ResNet18, ResNet50, and ResNet101. The experimental findings showed that the transfer learned ResNet50 model outperformed the other models, achieving a recall of 98.80% and an F1-score of 98.41%. To further enhance the results, the researchers examined the activations from different layers of the best performing model and applied support vector machine, logistic regression, and K-nearest neighbor classifiers for detection. Additionally, they proposed a classifier fusion strategy that combined predictions from different classifiers using majority voting. The experimental results demonstrated that by utilizing learned image features and the classification fusion strategy, the recall and F1-score were further improved to 99.20% and 99.40%, respectively [32].
Xu et al. presented a model that was evaluated on the COVID-CT and SARS-CoV-2 datasets. The proposed method was compared to a standard deep convolutional neural network (DCNN) as well as seven other variable-length models using five commonly used metrics: sensitivity, accuracy, specificity, F1-score, precision, and receiver operating curve (ROC) and precision-recall curves. The results of the study indicated that the proposed DCNN-IPSCA model outperformed the other benchmarks. It achieved a final accuracy of 98.32% and 98.01%, sensitivity of 97.22% and 96.23%, and specificity of 96.77% and 96.44% on the SARS-CoV-2 and COVID-CT datasets, respectively. These findings demonstrate the superiority of the proposed model in accurately classifying COVID-19 cases and distinguishing them from other conditions in the examined datasets [33].
According to Dansana, CT scan and X-ray image data sets containing 360 images were processed using a CNN-based approach. The data were transformed using the Inception_V2, DT, and VGG-19 methods based on a binary classification pneumonia method. We have demonstrated that a fine-tuned version of the VGG-19, Inception_V2, and DT method produces high training accuracy and validation rates [34].
In the case of feature extraction-based approaches, several frameworks have recently been developed, usually relying on a CNN. In addition to applying a 3D CNN to the whole CT volume in one stage, 2D CNNs are applied to CT slices, and the slice-level results are aggregated through an aggregation mechanism. The results obtained in a study by Wang on a dataset that included only COVID-19 cases and normal cases showed an accuracy of 90.1%, a sensitivity of 84.0%, and a specificity of 98.2% with a three-dimensional CNN-based classifier [26].
As Hu demonstrated that the same label can be applied to all slices of a CT scan. This allowed them to develop a comprehensive model that employed intermediate CNN layers for identifying classification characteristics. Final decisions are made by combining these features. Based on the three-way classification proposed by the researchers, their proposed method achieved an overall accuracy of 87.4 because each volume of the CT scan contains many slices without any visible infection area, and utilizing patient-level labels for all slices is unreasonable and will add errors to the system [35].
As described by Pathak, the proposed system is considered a pre-proposed transfer learning method for the detection of COVID-19 in CT scans. A 2D convolutional neural network was used to classify CT images using the ResNet50 algorithm. A 10-fold cross-validation procedure was used on 413 images of COVID-19 and 439 images of non-COVID-19, and the proposed system performed with an accuracy of 93.01% [9].
To classify 150 COVID-19 and non-COVID-19 images, Barstugan used machine learning algorithms instead of deep learning approaches. A support vector machine was employed to classify the extracted features, using several feature extraction methods such as the grey-level size zone matrix (GLSZM) and the discrete wavelet transform (DWT). Two, five, and ten folds of cross-validation were conducted in the experiments. As a result of the use of the GLSZM feature extraction method, an accuracy rate of 99.68% was achieved [36].
According to Nayak et al., their proposed model was evaluated using two larger datasets of chest X-ray images. Dataset-1 consisted of 2250 images, while Dataset-2 contained 15,999 images. The classification accuracy achieved by the model was 98.67% for Dataset-1 in the multi-class classification case and 99.00% in the binary classification case. For Dataset-2, the model achieved an accuracy of 95.67% in the multi-class classification and 96.25% in the binary classification. The performance of the proposed model was compared with four contemporary pre-trained convolutional neural network (CNN) models as well as state-of-the-art models. Additionally, the study investigated the impact of various hyperparameters such as different optimization techniques, batch size, and learning rate. One advantage of the proposed model is its ability to achieve high accuracy while demanding fewer parameters and requiring less memory space compared to other models [19].
Silva et al. introduced an efficient deep learning technique for COVID-19 screening, incorporating a voting-based approach. The proposed method involves classifying images from a given patient into groups using a voting system. The approach was evaluated on the two largest datasets available for COVID-19 CT analysis, with a patient-based split to ensure accurate testing. Additionally, a cross-dataset study was conducted to evaluate the models’ robustness in scenarios where data is sourced from different distributions. The cross-dataset analysis revealed that the generalization capability of deep learning models for this task is considerably inadequate. The accuracy dropped significantly from 87.68% to 56.16% in the best evaluation scenario, indicating a significant decrease in performance when applied to datasets from different distributions. This highlights the challenges in achieving robust and accurate results when deploying deep learning models in scenarios with varying data sources [37].
Compared with RT-PCR testing mechanisms, radiographic patterns on computed tomography (CT) chest scanning models provide superior sensitivity and specificity. Furthermore, a variety of methods have been developed to make use of CT and X-ray image datasets for the implementation of automated classifications. The results of the study demonstrate that CT and RT-PCR can be complementary in predicting COVID-19. CT features act as instant diagnostic indicators, while RT-PCR is used to confirm the diagnosis. Additionally, it is necessary to differentiate COVID-19 from other pneumonia infections in CT chest screening by leveraging the detective capabilities of artificial intelligence (AI). A deep learning (DL) method is particularly effective in separating COVID-19 cases from other types of pneumonia [38].
The Kassani study used several pre-trained networks to extract the features of images in the publicly available dataset so that COVID-19 could be differentiated from normal cases. The networks used were MobileNet, DenseNet, Xception, InceptionV3, InceptionResNetV2, and ResNet. After feature extraction, a series of machine learning algorithms were applied, including a decision tree, random forest, XGBoost, AdaBoost, Bagging, and LightGBM. Kassani concluded that the Bagging classifier produces the most accurate results based on features extracted from the pre-trained network DESNSEA121, with an accuracy of 99.00% [39].
Kogilavani et al. conducted a study where they utilized deep learning techniques to detect COVID-19 patients through the analysis of CT scans. Their research focused on developing deep learning methods specifically designed for COVID-19 detection. Several convolutional neural network (CNN) architectures, namely, VGG16, DenseNet121, MobileNet, NASNet, Xception, and EfficientNet, were employed in their study. The dataset consisted of a total of 3873 CT scans, encompassing both “COVID-19” and “Non-COVID-19” cases. Separate datasets were allocated for the validation and test phases. The obtained accuracy rates for the different CNN architectures were as follows: VGG16 achieved an accuracy of 96.68%, DenseNet121 achieved 97.53%, MobileNet achieved 96.38%, NASNet achieved 89.51%, Xception achieved 92.47%, and EfficientNet achieved 80.19%. The analysis of the results demonstrated that the VGG16 architecture exhibited superior accuracy compared to the other architectures [40].
Ruano et al. conducted a study where they utilized two datasets for their analysis: SARS-CoV-2 CT Scan (Set-1) and FOSCAL clinic’s dataset (Set-2). To leverage the power of deep learning, they employed supervised learning models that were pre-trained on natural image data. These models were then fine-tuned using a transfer learning approach. The deep classification was carried out using two methods: (a) an end-to-end deep learning approach; (b) random forest and support vector machine classifiers, where the deep representation embedding vectors were fed into these classifiers. For Set-1, the end-to-end deep learning approach achieved an average accuracy of 92.33%, with a precision of 89.70%. In the case of Set-2, the end-to-end approach achieved an average accuracy of 96.99%, with a precision of 96.62%. On the other hand, when utilizing deep feature embedding with a support vector machine, the average accuracy for Set-1 was 91.40%, with a precision of 95.77%. For Set-2, the average accuracy reached 96.00%, with a precision of 94.74% [41].
According to Peng et al., their study involved the utilization of three existing COVID-19-related CT image datasets, which were combined to form a larger integrated dataset. The weights of DenseNet, Swin Transformer, and RegNet were pretrained on the ImageNet dataset using transfer learning. Subsequently, these models were further trained on the integrated dataset comprising COVID-19 CT images. The classification results were obtained by aggregating the predictions from the three models using the soft voting approach. The proposed model, called DeepDSR, was compared to three state-of-the-art deep learning models (EfficientNetV2, ResNet, and Vision Transformer) as well as the individual models (DenseNet, Swin Transformer, and RegNet) for both binary and three-class classification problems. The results demonstrated that DeepDSR achieved the highest precision of 98.33%, recall of 98.95%, accuracy of 98.94%, F1-score of 98.64%, AUC of 99.91%, and AUPR of 0.9986 in the binary classification problem, significantly surpassing other methods. Furthermore, DeepDSR attained the best precision of 97.40%, recall of 96.53%, accuracy of 97.37%, and F1-score of 0.9695 in the three-class classification problem [17].

3. Materials and Methods

In this paper, the first step of the proposed method is explaining the COVID-19 dataset. In this case, the images were in png format and did not require any filtering or normalization. Once all images have been resized to 200 × 200 pixels, K-means segmentation is applied to the images to remove any non-lung parts from the images. As a result of this, FrFT coefficients are used for feature extraction with diverse fractional orders. Real, absolute, imaginary, and phase components of the complex coefficients are computed as part of the equation for the complex coefficients. Next, a novel method called adjacent rectangular windows is applied to determine which features will produce the best results. The coefficients of a 2D window containing coefficients were calculated based on the maximum, minimum, median, and mean values. In the end, K’s nearest neighbors (KNN) and support vector machines (SVM) were applied to classify the images with the most accurate features extracted by the windowing method.
These various stages are applied to achieve the best accuracy and separate COVID-19 from non-COVID-19. The structure of each step is shown in Figure 1.

3.1. Database

The suggested dataset (SARS-CoV-2) contains 2481 CT images shared among 1252 patients infected with COVID-19; and 1229 CT scans for non-COVID-19 patients that have other pulmonary diseases. Data were gathered from hospitals in Sao Paulo, Brazil. In this case, the data consist of 60 patients infected with coronavirus; 28 of them were female, and 32 were male [42]. The size of the images in this database was different; the size of all the images equalized before the pre-processing stage (Figure 2). The dimensions of the images vary; for example, the smallest size recorded in the database is 104 × 153, while the largest size is 484 × 416. Figure 3 displays several instances of CT scans from both SARS-CoV-2-infected and non-infected patients, which form part of the dataset.

3.2. Pre-Processing

To begin with, every image in the database was resized to 200 × 200 pixels. This is due to the fact that the images in the database are of different sizes, and each image has a different number of pixels. Furthermore, since coronavirus is a lung disease, it is imperative to distinguish lung areas from other body parts. In other regions of the image, white borders can be seen that were derived from the original image using K-means segmentation and morphological methods. The lung image is illustrated in Figure 4, following the application of the pre-processing method.

3.3. Fractional Fourier Transform

An alternative to the traditional Fourier transform (FT) is the fractional Fourier transform (FrFT), which extends the FT to the entire time-frequency domain using fractional powers of the Fourier operator. As a generalization of the FT, the FrFT adds an extra parameter related to the rotation of the signal in the time-frequency domain. A chirp function basis is used in such a scheme to decompose the temporal signal. Thus, the FrFT provides a unified time-frequency representation of the signal with a higher time-frequency resolution than other techniques [43]. Numerous applications of the FrFT can be found in quantum mechanics, optics, and signal and image processing. Scholars believe that the extracted features must excellently separate the desired output classes.
The extracting features can be applied in frequency, time, and time-frequency domains based on the paper’s purpose. In this paper, FrFT is used for feature extraction. In 1980, Namias offered this operative to explain Hamilton’s second-order function in quantum-mechanical systems [23]. After that, Several researchers contributed to its development in numerous applications, such as mathematic approaches to the matrix in image processing and swept-frequency filters [7].
In addition to being a linear transformation, FrFT is a generalization format of the normal Fourier transform that includes a specific order parameter α that varies from zero to one ( 0 < α < 1 ) ; FrFT is illustrated in Figure 5. There is no way of determining the most acceptable value of α to provide accurate data analysis. While α = 1, the fractional Fourier transform behaves similarly to a conventional Fourier transform. FrFT can be used in several techniques, for instance, the rotation of a function in the time-frequency domain, fractional powers of the Fourier transform, and differential equations [25]. Accordingly, the linear integral transform explanation was used as the most direct and concrete explanation based on the linear transform explanation for computing. The function usually be represented as f a u .
f a u = k a u . u f u d u .
k a u . u = A α exp [ i π ( cot ( α ) u 2 2 csc ( α ) u u + cot ( α ) u 2 ) ]
A α = 1 i cot ( α ) α = a π 2
where a is the number of rotations on the interval 0 ≤ |a| ≤ 2 and k a u . u is the kernel, which is symmetric and has the following explanation when α = π 2 .
The kernel is δ(u u ) when α = 2 nπ, and the kernel function utilizied is represented by δ(u + u ), where α + π = 2 nπ [43].
k a u . u = 1 i c o t ( α ) 2 × e j u 2 2 c o t ( α ) e j u 2 2 c o t ( α ) j u u c s c ( α )
Determining the appropriate alpha value is one of the challenges of this project, which uses fractional Fourier transform to extract features. The coefficients of the fractional Fourier transform are obtained after applying the fractional Fourier transform to extract features. It is very important to note that these coefficients are complex numbers. The concept of real and imaginary parts, absolute value, and phase are all included. We have separated and processed each of these components separately in the course of this research.

3.4. Feature Extraction and Selection

A subset of features can be derived in two ways: by feature extraction and selection. Feature extraction involves the process of extracting specific properties or characteristics from a dataset, typically with a certain level of detail or resolution. On the other hand, feature selection refers to the act of choosing or selecting a subset of features from the original feature set, often based on certain criteria or algorithms. Feature separation is impossible when a quality strongly correlates with the feature set. In some cases, a part may play a role in classification accuracy despite poor relevance [13]. According to Heisenberg’s uncertainty principle, the accuracy of time and frequency measurements cannot be increased simultaneously. In this way, by increasing the accuracy in the time domain, the accuracy of the work in the frequency domain decreases.
In the same way, if the accuracy in the frequency domain increases, the accuracy in the time domain decreases. Frequency in biological signals at different times can have very different meanings. This limitation can be used in image and signal processing. Frequency domain transformations encompass various techniques, such as the fractional Fourier transform, fractional S Transform, and fractional wavelet transform, which involve manipulating signals in the frequency domain. In this paper, after extracting the feature, due to the large volume of features extracted by fractional Fourier transform, we needed to select the feature and reduce its values. As shown in Figure 6, the proposed method is outlined step by step. A novel windowing method was used to reduce features. As a result of the pre-processing method, each image is converted to 200 by 200 pixels. The feature extraction process resulted in the extraction of 40,000 features. The number of features needed to be reduced. The features were reduced using two-dimensional windows. First, one-dimensional vector features were reconstructed into a two-dimensional matrix of 200 × 200 features. To reduce the features, a two-dimensional transition window was used. These images were scaled according to the proportions of 10 × 10, 20 × 20, to 100 × 100. In the case of a 10 × 10 window, the number of features decrease from 40,000 to 200. In the case of a 100 × 100 window, the number of features decrease from 40,000 to 20. After applying the windowing, the statistical methods of mean, median, minimum, and maximum were used.

3.5. Classification

As a result of image processing after feature extraction, the most practical features to use for classification consist of identifying COVID-19 in each image. In this paper (KNN) K-nearest neighbor and the support vector machine (SVM) are used for classification.
KNN is one of the most popular classifiers in pattern recognition. It is a popular method for supervised binary classification and two-class learning, which can be used for machine learning. This classifier is designed to distinguish classes based on the detection of hyperplanes. SVM is designed to identify distinct classes of data points by finding hyperplanes in an N-dimensional space. SVM measures the margin distance between data points of the two classes. This is because there are several options for choosing the hyperplane. This margin should be maximized to provide a clear decision boundary for classifying future data points [29].
Data are divided into two groups using the Holdout method: train and test. Several possible divisions exist, such as 40/60, 30/70, or 20/80. Therefore, training is conducted on training data, and evaluation is performed on test data to produce the desired model. This method is called Holdout validation. If the number of classes in the test and train groups does not match, the Holdout method will not be able to train the model correctly. Consequently, both the training and test groups must have the same distribution of classes. Classification is based on the division 10/90 in this project.
This method compares the training group to a test group based on similarity. The distance between the entered new sample and all training samples can then be calculated, and the K-nearest model will be chosen to classify the updated sample. After organizing non-COVID-19 and COVID-19 subjects, nominated features were classified with various properties for the train and test stage. The primary step was to allocate 10% of non-COVID-19 and the remaining for the test stage. Then, the amount of training is reduced to 10%. The performance of the suggested method is evaluated using specificity, accuracy, sensitivity, and precision, which are defined as follows:
A c c u r a c y   A c c = T P + T N T P + T N + F P + F N P r e c i s i o n   P r e = T P T P + F P S e n s i t i v i t y   S e n = T P T P + F N S p e c i f i c i t y   S p e = T N T N + F P
T N : true negative, T N : true negative, F P : false positive, F N : false negative
The performance has been stated with the performance factors via two classifiers.

4. Results

After extracting features, such as the derived phase, imaginary, real, and absolute value, a classification procedure was used. The fractional order of 0.8 and 0.7 had been achieved for the KNN and SVM classifiers, respectively. This order was achieved by trial and error; unfortunately, there are not any other techniques to determine this order. This study used the 2D windowing method to select features, and the derivative calculated the difference in features. This operator is used to convert the feature matrix into a new space, with significantly promising results. The declared average value is the average taken from the indicators (specificity, accuracy, precision, and sensitivity).
To evaluate its performance, different features of the proposed technique are examined in Figure 7, such as the absolute, real, imaginary, and phase coefficients. Figure 7 shows that the best coefficient for KNN is determined based on phase, while the best coefficient for SVM is determined based on absolute value. The average value used in all stages is calculated by considering the average value of all indicators in the real and imaginary parts, as well as the phase and amplitude parts. Based on the results depicted in the figure, the optimal alpha value for KNN classification is determined to be 0.8, while for SVM classification, it is obtained from 0.7.
At the end of the process, the features are classified. It was necessary to preprocess images before FrFT to determine the best fractional order. Each coefficient was used to specify the optimal characteristics for each step. Figure 8 shows the most accurate K-neighbor (1–9) value for KNN classifiers. With 99.80% accuracy, the best K-neighbor value is achieved at k = 1. The best features were selected, illustrated in 90% of the training and one neighborhood by using a 2D window (including mean, median, maximum, and minimum coefficients) with different dimensions. As an important point to note, FrFT with 0.8 and 0.7 fractional orders may provide more accurate results than FrFT with other fractional orders for KNN and SVM classifiers. This applies specifically when α = 1, which is equivalent to FT. From Figure 9, the Euclidean distance between Chebyshev, Hamming, and Minkowski classifiers yields the highest average and accuracy with 99.74% and 99.09%, respectively, for the KNN classifier. In comparison to the other training results for the KNN classifier, 90% of the training results have higher accuracy than the others. Median, maximum, and minimum, when compared by the mean static method, have an accuracy and average of all indicators of 99.89% and 93.96%, respectively (Figure 10).
Among the most significant parameters involved in SVM classification is the selection of the appropriate kernel. Table 1 shows that the polynomial kernel, the most accurate one, achieves 99.90% and 97.49%, respectively, at both the average and accuracy levels of SVM kernels. There is a significant difference between 90% of the training results for the SVM classifier that are more accurate in comparison to the rest of the results. As a result of using the median static method between mean, maximum, and minimum, the accuracy and average of all indicators were 99.89% and 93.96%, respectively (Table 2).
After the classification step, the following steps determine an appropriate number of optimal features. The first objective was to determine what was the average performance of classifiers across all criteria. Afterward, it was determined which features resulted in the highest level of accuracy. According to these findings, 80 and 200 features are appropriate for the detection of COVID-19 using KNN and SVM classifiers. With the use of KNN and SVM classifiers, Figure 11 displays the average accuracy of each feature. In addition, by utilizing KNN and SVM classifiers, the average accuracy achieved is approximately 95.05% and 99.90%, respectively.

5. Discussion

As a result of this experiment, it appears that fractional Fourier transforms can produce increasingly good results when used in conjunction with CT images for classification. The results presented in this paper were produced by following some steps to be as reliable as possible. To sum up, in the beginning, only lung parts were segmented from the dataset. A large feature plane was constructed using 40,000 features extracted from the fractional Fourier transform coefficients based on the size of each image (200 × 200). An innovative method, including 2D windowing, was used to select the best features. We used a 2D window with different dimensions to calculate the mean, median, maximum, and minimum coefficients. Consequently, the number of features decreased from 40,000 to 80 for KNN and 200 for SVM. The KNN and SVM classifiers yielded the most optimal outcomes. With 90% training (the training dataset is completely different from the test dataset) and k = 1, Euclidean distance for KNN, and polynomial kernel for SVM classifiers, they were able to accomplish this objective. All four coefficients, real, absolute, imaginary, and phase, were extracted as different features in each step.
A review of previous works reveals that all used CT scan images and employed supervised and unsupervised classification methods and deep learning. A clear comparison between the results of the proposed method and those of traditional methods can be achieved by comparing the results of the proposed method to those of traditional methods, such as those shown in Table 3. To obtain a reliable comparison, the conditions of the experiments should be similar. Accordingly, it is impossible to recommend one study over another. Taking a quick look at the information, the fractional Fourier transform method is found to be a sufficient method for extracting features. There are many advantages to using the fractional Fourier transform. These include its coefficients, which are the same characteristics that were examined in this study. Mixture coefficients provide both amplitude and phase as well as the absolute value and imaginary part. They are much simpler and faster than the methods currently being used. The 2D windowing method is used in this research to reduce features. This innovative method is also fast and reliable and relies on simple mathematical Algorithms like minimum and maximum have shown effective outcomes in accurately classifying between COVID-19 and non-COVID-19 cases. Despite the time-consuming nature commonly associated with 2D fractional Fourier transform, it can still be considered as a potential method for feature extraction in future applications, taking into account certain limitations.

6. Conclusions

To validate the results, our model outputs were cross-checked with healthcare professionals. Our goal is to demonstrate the potential of artificial intelligence-based methods in the fight against the current pandemic by using reliable diagnostic methods that can be obtained easily, such as chest radiographs, to aid in the fight. In this paper, the FrFT technique is suggested for COVID-19 classification. The proposed method is fast and accurate. The accuracy scores achieved for KNN and SVM classification are 99.84% and 99.90%, respectively. The results initiate the application of the presented process. This technique will help doctors and scholars with much quicker and more accurate COVID-19 detection than other previous methods. The results of this study specify the capability of FrFT and the influence of this technique in COVID-19 detection.

Author Contributions

A.N., methodology, software, formal analysis, writing—original draft, review, and editing; S.R., supervision, the idea of the research, helping in programming, helping in writing and editing of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data utilized in this article can be obtained from the following source: (https://www.kaggle.com/plameneduardo/sarscov2-ctscan-dataset, accessed on 22 June 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sandeep Kumar, M.; Jayagopal, P.; Ahmed, S.; Manivannan, S.S.; Kumar, P.J.; Raja, K.T.; Prasad, R.G. Adoption of e-learning during lockdown in India. Int. J. Syst. Assur. Eng. Manag. 2023, 14 (Suppl. S1), 575. [Google Scholar]
  2. Sandeep, K.; Maheshwari, V.; Prabhu, J.; Prasanna, M.; Jayalakshmi, P.; Suganya, P.; Jothikumar, R. Social economic impact of COVID-19 outbreak in India. Int. J. Pervasive Comput. Commun. 2020, 16, 309–319. [Google Scholar]
  3. Wang, B.; Jiansheng, T.; Fujian, Y.; Zhiyu, Z. Identification of sonar detection signal based on fractional Fourier transform. Pol. Marit. Res. 2018, 25, 125–131. [Google Scholar]
  4. Wu, G.; Kim, M.; Wang, Q.; Munsell, B.C.; Shen, D. Scalable High-Performance Image Registration Framework by Unsupervised Deep Feature Representations Learning. IEEE Trans. Biomed. Eng. 2016, 63, 1505–1516. [Google Scholar] [CrossRef]
  5. Xu, B.; Martín, D.; Khishe, M.; Boostani, R. COVID-19 diagnosis using chest CT scans and deep convolutional neural networks evolved by IP-based sine-cosine algorithm. Med. Biol. Eng. Comput. 2022, 60, 2931–2949. [Google Scholar] [CrossRef] [PubMed]
  6. Adel, O.; Agaian, S.; Trongtirakul, T.; Laouar, A.K. Automatic COVID-19 lung infected region segmentation and measurement using CT-scans images. Pattern Recognit. 2020, 114, 107747. [Google Scholar]
  7. Banu Priya, K.; Rajendran, P.; Kumar, S.; Prabhu, J.; Rajendran, S.; Kumar, P.J.; Jothikumar, R. Pediatric and geriatric immunity network mobile computational model for COVID-19. Int. J. Pervasive Comput. Commun. 2020, 16, 321–330. [Google Scholar] [CrossRef]
  8. Kumar, G.; Bhatia, P.K. A detailed review of feature extraction in image processing systems. In Proceedings of the 2014 Fourth International Conference on Advanced Computing & Communication Technologies, Washington, DC, USA, 8–9 February 2014; IEEE: Miami, FL, USA, 2014; pp. 5–12. [Google Scholar]
  9. Ouyang, X.; Huo, J.; Xia, L.; Shan, F.; Liu, J.; Mo, Z.; Shen, D. Dual-sampling attention network for diagnosis of COVID-19 from community-acquired pneumonia. IEEE Trans. Med. Imaging 2020, 39, 2595–2605. [Google Scholar] [CrossRef]
  10. Abbas, A.; Abdelsamea, M.M.; Medhat Gaber, M. Classification of COVID-19 in chest x-ray images using detract deep convolutional neural network. Appl. Intell. 2021, 51, 854–864. [Google Scholar] [CrossRef] [PubMed]
  11. Kaur, T.; Gandhi, T.K. Classifier fusion for detection of COVID-19 from CT scans. Circuits Syst. Signal Process. 2022, 41, 3397–3414. [Google Scholar] [CrossRef] [PubMed]
  12. Bansal, A.; Jain, V.; Chatterjee, J.M.; Kose, U.; Jain, A. (Eds.) Computational Intelligence in Software Modeling; Walter de Gruyter GmbH & Co., KG: Berlin, Germany, 2020; Volume 13. [Google Scholar]
  13. Gupta, S.; Goel, S.; Nijhawan, R.; Vivek, P. CNN Models and Machine Learning Classifiers for Analysis of Goiter Disease. In Proceedings of the 2022 IEEE International Students’ Conference on Electrical, Electronics and Computer Science (SCEECS), Bhopal, India, 19–20 February 2022; IEEE: Miami, FL, USA; pp. 1–6. [Google Scholar]
  14. Huang, C.; Wang, Y.; Li, X.; Ren, L.; Zhao, J.; Hu, Y.; Zhang, L.; Fan, G.; Xu, J.; Gu, X.; et al. Clinical features of patients infected with 2019 novel coronavirus in Wuhan, China. Lancet 2020, 395, 497–506. [Google Scholar] [CrossRef] [PubMed]
  15. Iwai, R.; Yoshimura, H. A new method for improving robustness of registered fingerprint data using the fractional Fourier transform. Int. J. Commun. Netw. Syst. Sci. 2010, 3, 722. [Google Scholar] [CrossRef]
  16. Pathak, Y.; Shukla, P.; Tiwari, A.; Stalin, S.; Singh, S. Deep transfer learning based classification model for COVID-19 disease. IRBM 2020, 43, 87–92. [Google Scholar] [CrossRef] [PubMed]
  17. Manjit, K.; Kumar, V.; Yadav, V.; Singh, D.; Kumar, N.; Das, N.N. Metaheuristic-based deep COVID-19 screening model from chest X-ray images. J. Healthc.Eng. 2021, 2021, 8829829. [Google Scholar]
  18. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  19. Chu, D.K.; Pan, Y.; Cheng, S.M.; Hui, K.P.; Krishnan, P.; Liu, Y.; Poon, L.L. Molecular diagnosis of a novel coronavirus (2019-nCoV) causing an outbreak of pneumonia. Clin. Chem. 2020, 66, 549–555. [Google Scholar] [CrossRef]
  20. Chung, M.; Bernheim, A.; Mei, X.; Zhang, N.; Huang, M.; Zeng, X.; Shan, H. CT imaging features of 2019 novel coronavirus (2019-nCoV). Radiology 2020, 295, 202–207. [Google Scholar] [CrossRef]
  21. Corman, V.M.; Landt, O.; Kaiser, M.; Molenkamp, R.; Meijer, A.; Chu, D.K.W.; Bleicker, T.; Brünink, S.; Schneider, J.; Schmidt, M.L.; et al. Detection of 2019 novel coronavirus (2019-nCoV) by real-time RT-PCR. Eurosurveillance 2020, 25, 2000045. [Google Scholar] [CrossRef] [PubMed]
  22. Dansana, D.; Kumar, R.; Bhattacharjee, A.; Hemanth, D.J.; Gupta, D.; Khanna, A.; Castillo, O. Early diagnosis of COVID-19-affected patients based on X-ray and computed tomography images using deep learning algorithm. Soft Comput. 2020, 27, 2635–2643. [Google Scholar] [CrossRef]
  23. Du, J.-X.; Huang, D.-S.; Wang, X.-F.; Gu, X. Shape recognition based on neural networks trained by differential evolution algorithm. Neurocomputing 2007, 70, 896–903. [Google Scholar] [CrossRef]
  24. Maheshwari, V.; Mathivanan, S.K.; Jayagopal, P.; Mani, P.; Rajendran, S.; Subramaniam, U.; Sorakaya Somanathan, M. Forecasting of the SARS-CoV-2 epidemic in India using SIR model, flatten curve and herd immunity. J. Ambient. Intell. Humaniz. Comput. 2020, 1–9, Epub ahead of print. [Google Scholar]
  25. Mendlovic, D.; Ozaktas, H.M.; Lohmann, A.W. Fourier transforms of fractional order and their optical interpretation. In Optical Computing; Optica Publishing Group: Washington, DC, USA, 1993; p. OWD–6. [Google Scholar]
  26. Wang, D.; Hu, B.; Hu, C.; Zhu, F.; Liu, X.; Zhang, J.; Peng, Z. Clinical characteristics of 138 hospitalized patients with 2019 novel coronavirus–infected pneumonia in Wuhan, China. JAMA 2020, 323, 1061–1069. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, N.; Wang, L.; Deng, X.; Liang, R.; Su, M.; He, C.; Jiang, S. Recent advances in the detection of respiratory virus infection in humans. J. Med. Virol. 2020, 92, 408–417. [Google Scholar] [CrossRef] [PubMed]
  28. Abbas, T.; Ardebili, A. Real-time RT-PCR in COVID-19 detection: Issues affecting the results. Expert Rev. Mol. Diagn. 2020, 20, 453–454. [Google Scholar]
  29. Li, Q.; Guan, X.; Wu, P.; Wang, X.; Zhou, L.; Tong, Y.; Feng, Z. Early transmission dynamics in Wuhan, China, of novel coronavirus–infected pneumonia. N. Engl. J. Med. 2020, 382, 1199. [Google Scholar] [CrossRef]
  30. Nayak, S.R.; Nayak, D.R.; Sinha, U.; Arora, V.; Pachori, R.B. An Efficient Deep Learning Method for Detection of COVID-19 Infection Using Chest X-ray Images. Diagnostics 2023, 13, 131. [Google Scholar] [CrossRef]
  31. Rawat, W.; Wang, Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef]
  32. Gómez-Echavarría, A.; Ugarte, J.P.; Tobón, C. The fractional Fourier transform as a biomedical signal and image processing tool: A review. Biocybern. Biomed. Eng. 2020, 40, 1081–1093. [Google Scholar] [CrossRef]
  33. Wu, X.; Hui, H.; Niu, M.; Li, L.; Wang, L.; He, B.; Zha, Y. Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: A multicentre study. Eur. J. Radiol. 2020, 128, 109041. [Google Scholar] [CrossRef]
  34. Hu, S.; Gao, Y.; Niu, Z.; Jiang, Y.; Li, L.; Xiao, X.; Yang, G. Weakly supervised deep learning for COVID-19 infection detection and classification from CT images. IEEE Access 2020, 8, 118869–118883. [Google Scholar] [CrossRef]
  35. Barstugan, M.; Ozkaya, U.; Ozturk, S. Coronavirus (COVID-19) classification using CT images by machine learning methods. arXiv 2020, arXiv:2003.09424. [Google Scholar]
  36. Silva, P.; Luz, E.; Silva, G.; Moreira, G.; Silva, R.; Lucio, D.; Menotti, D. COVID-19 detection in CT images with deep learning: A voting-based scheme and cross-datasets analysis. Inform. Med. Unlocked 2020, 20, 100427. [Google Scholar] [CrossRef] [PubMed]
  37. Shankar, K.; Mohanty, S.N.; Yadav, K.; Gopalakrishnan, T.; Elmisery, A.M. Automated COVID-19 diagnosis and classification using a convolutional neural network with a fusion-based feature extraction model. Cogn. Neurodyn. 2021, 17, 1–14. [Google Scholar] [CrossRef] [PubMed]
  38. Kassania, S.H.; Kassanib, P.H.; Wesolowskic, M.J.; Schneidera, K.A.; Detersa, R. Automatic detection of coronavirus disease (COVID-19) in X-ray and CT images: A machine learning based approach. Biocybern. Biomed. Eng. 2021, 41, 867–879. [Google Scholar] [CrossRef]
  39. Kogilavani, S.V.; Prabhu, J.; Sandhiya, R.; Kumar, M.S.; Subramaniam, U.; Karthick, A.; Muhibbullah, M.; Imam, S.B.S. COVID-19 detection based on lung CT scan using deep learning techniques. Comput. Math. Methods Med. 2022, 2022, 1–13. [Google Scholar] [CrossRef] [PubMed]
  40. Peng, L.; Wang, C.; Tian, G.; Liu, G.; Li, G.; Lu, Y.; Yang, J.; Chen, M.; Li, Z. Analysis of CT scan images for COVID-19 pneumonia based on a deep ensemble framework with DenseNet, Swin transformer, and RegNet. Front. Microbiol. 2022, 13, 3523. [Google Scholar] [CrossRef] [PubMed]
  41. Soares, E.; Angelov, P.; Biaso, S.; Froes, M.H.; Abe, D.K. SARS-CoV-2 CT-scan dataset: A large dataset of real patients CT scans for SARS-CoV-2 identification. MedRxiv 2020, 2020, 20078584. [Google Scholar]
  42. Victor, N. The fractional order Fourier transform and its application to quantum mechanics. IMA J. Appl. Math. 1980, 25, 241–265. [Google Scholar]
  43. Taghizadeh, Z.; Rashidi, S.; Shalbaf, A. Finger movements classification based on fractional fourier transform coefficients extracted from surface emg signals. Biomed. Signal Process. Control. 2021, 68, 102573. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed method.
Figure 1. Block diagram of the proposed method.
Fractalfract 08 00237 g001
Figure 2. Database sample images.
Figure 2. Database sample images.
Fractalfract 08 00237 g002
Figure 3. Showcases a selection of CT scans included in the dataset, demonstrating examples of both SARS-CoV-2-infected and non-infected patients.
Figure 3. Showcases a selection of CT scans included in the dataset, demonstrating examples of both SARS-CoV-2-infected and non-infected patients.
Fractalfract 08 00237 g003
Figure 4. This figure illustrates the contrast between the image prior to and following the implementation of pre-processing procedures. The raw image of the lung of a person infected with COVID-19 (A); the image of the lung after pre-processing (B).
Figure 4. This figure illustrates the contrast between the image prior to and following the implementation of pre-processing procedures. The raw image of the lung of a person infected with COVID-19 (A); the image of the lung after pre-processing (B).
Fractalfract 08 00237 g004
Figure 5. FrFT is represented by two axes u u , and those axes are oriented by ϕ.
Figure 5. FrFT is represented by two axes u u , and those axes are oriented by ϕ.
Fractalfract 08 00237 g005
Figure 6. An overview of the proposed method.
Figure 6. An overview of the proposed method.
Fractalfract 08 00237 g006
Figure 7. Classification based on average results (%) of fractional Fourier transform coefficients of imaginary, real, absolute value, and phase.
Figure 7. Classification based on average results (%) of fractional Fourier transform coefficients of imaginary, real, absolute value, and phase.
Fractalfract 08 00237 g007
Figure 8. The average accuracy result (%) of KNN classification based on the number of K-neighbors for α = 0.8.
Figure 8. The average accuracy result (%) of KNN classification based on the number of K-neighbors for α = 0.8.
Fractalfract 08 00237 g008
Figure 9. The average accuracy results (%) of the KNN classification based on distance type (Euclidean, Chebyshev, Mahalanobis, Minkowski) for α = 0.8.
Figure 9. The average accuracy results (%) of the KNN classification based on distance type (Euclidean, Chebyshev, Mahalanobis, Minkowski) for α = 0.8.
Fractalfract 08 00237 g009
Figure 10. KNN classification results (%) based on statistical method after applying 2D windowing and Euclidean distance with K-neighbor 1 (mean, median, minimum, maximum) for α = 0.8.
Figure 10. KNN classification results (%) based on statistical method after applying 2D windowing and Euclidean distance with K-neighbor 1 (mean, median, minimum, maximum) for α = 0.8.
Fractalfract 08 00237 g010
Figure 11. Average results (%) by the impacts of the number of features reduction. KNN Classifier (Green): Phase features, 1 number of neighbors, Euclidean distance with mean mathematical statistics (α = 0.8). SVM classifier (Blue): Absolute features, Polynomial kernel with median mathematical statistic (α = 0.7).
Figure 11. Average results (%) by the impacts of the number of features reduction. KNN Classifier (Green): Phase features, 1 number of neighbors, Euclidean distance with mean mathematical statistics (α = 0.8). SVM classifier (Blue): Absolute features, Polynomial kernel with median mathematical statistic (α = 0.7).
Fractalfract 08 00237 g011
Table 1. SVM classification results based on SVM kernels with median statical method (RBF, Polynomial, Linear) for α = 0.7.
Table 1. SVM classification results based on SVM kernels with median statical method (RBF, Polynomial, Linear) for α = 0.7.
SVMKernels
RBF (%)Linear (%)Polynomial (%)
Accuracy 99.59 92.53 99.86
Specificity 99.18 89.04 99.69
Sensitivity 100 100 100
Precision 99.20 91.06 99.75
Table 2. SVM classification results based on statistical method after applying 2D windowing for Polynomial kernel with 200 features (mean, median, minimum, maximum) for α = 0.7.
Table 2. SVM classification results based on statistical method after applying 2D windowing for Polynomial kernel with 200 features (mean, median, minimum, maximum) for α = 0.7.
Mathematical
Statistic
Mean (%)Median (%)Minimum (%)Maximum (%)
Accuracy96.4199.9098.3095.88
Specificity92.7599.8396.5892.69
Sensitivity100100100100
Precision93.3699.8496.7594.56
Table 3. The comparison of other research for the diagnosis of COVID-19.
Table 3. The comparison of other research for the diagnosis of COVID-19.
Model UsedDatasetResults (%)
Xu et al. [5]DCNN-IPSCA1252
COVID-19
1229
Non-COVID-19
ACC: 98.32
Peng et al. [17]DeepDSR1252
COVID-19
1229
Non-COVID-19
ACC: 98.94
Gupta et al. [22]ResNet-50, and DenseNet-121.1252
COVID-19
1229
Non-COVID-19
AUC: 85
Wu et al. [33]ResNet501252
COVID-19
1229
Non-COVID-19
AUC: 73.20
Ruano et al. [41]End-To-End
Deep learning
1252
COVID-19
1229
Non-COVID-19
ACC: 96.99
Proposed
Method
FrFT1252
COVID-19
1229
Non-COVID-19
SVM ACC: 99.90
KNN ACC: 99.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nokhostin, A.; Rashidi, S. COVID-19 Diagnosis by Extracting New Features from Lung CT Images Using Fractional Fourier Transform. Fractal Fract. 2024, 8, 237. https://doi.org/10.3390/fractalfract8040237

AMA Style

Nokhostin A, Rashidi S. COVID-19 Diagnosis by Extracting New Features from Lung CT Images Using Fractional Fourier Transform. Fractal and Fractional. 2024; 8(4):237. https://doi.org/10.3390/fractalfract8040237

Chicago/Turabian Style

Nokhostin, Ali, and Saeid Rashidi. 2024. "COVID-19 Diagnosis by Extracting New Features from Lung CT Images Using Fractional Fourier Transform" Fractal and Fractional 8, no. 4: 237. https://doi.org/10.3390/fractalfract8040237

Article Metrics

Back to TopTop