Next Article in Journal
A Machine Learning-Based Clustering Using Radiomics of F-18 Fluorodeoxyglucose Positron Emission Tomography/Computed Tomography for the Prediction of Prognosis in Patients with Intrahepatic Cholangiocarcinoma
Previous Article in Journal
Pre-Conception Androgen Levels and Obstetric Outcomes in Polycystic Ovary Syndrome: A Single-Center Retrospective Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Study of Metaheuristic Feature Selection Algorithms for Respiratory Disease Classification

1
Department of Electrical and Electronics Engineering, Dokuz Eylül University, İzmir 35160, Türkiye
2
Department of Biomedical Engineering, İskenderun Technical University, İskenderun 31200, Türkiye
3
Department of Electrical and Electronics Engineering, İzmir Institute of Technology, İzmir 35433, Türkiye
4
Ministry of Transportation, İzmir 35070, Türkiye
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(19), 2244; https://doi.org/10.3390/diagnostics14192244
Submission received: 26 August 2024 / Revised: 2 October 2024 / Accepted: 7 October 2024 / Published: 8 October 2024
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The correct diagnosis and early treatment of respiratory diseases can significantly improve the health status of patients, reduce healthcare expenses, and enhance quality of life. Therefore, there has been extensive interest in developing automatic respiratory disease detection systems. Most recent methods for detecting respiratory disease use machine and deep learning algorithms. The success of these machine learning methods depends heavily on the selection of proper features to be used in the classifier. Although metaheuristic-based feature selection methods have been successful in addressing difficulties presented by high-dimensional medical data in various biomedical classification tasks, there is not much research on the utilization of metaheuristic methods in respiratory disease classification. This paper aims to conduct a detailed and comparative analysis of six widely used metaheuristic optimization methods using eight different transfer functions in respiratory disease classification. For this purpose, two different classification cases were examined: binary and multi-class. The findings demonstrate that metaheuristic algorithms using correct transfer functions could effectively reduce data dimensionality while enhancing classification accuracy.

1. Introduction

There are many respiratory diseases, such as chronic obstructive pulmonary disease (COPD), asthma, pneumonia, bronchiectasis, bronchiolitis, and upper/lower respiratory tract infections. These diseases are at the top of the list when considering global deaths, emphasizing the importance of their accurate and early diagnosis. Correct diagnosis and early treatment of respiratory diseases can significantly improve the health status of patients, reduce healthcare costs, and improve quality of life. Among the various diagnostic tools available, analysis of respiratory sounds by auscultation is a basic method for identifying respiratory abnormalities. Respiratory sounds such as roughness, coarse crackling, monophonic wheeze, polyphonic wheeze, stridor, bronchus, and squawk provide valuable clues about the respiratory system. Upper respiratory tract infection (URTI), COPD, bronchiectasis, pneumonia, bronchiolitis, asthma, and lower respiratory tract infection (LRTI) are among the most common respiratory diseases that can be detected by auscultation methods. Traditional auscultation relies heavily on the experience and interpretation capability of the physician, which can lead to variabilities in diagnosis. Digital stethoscopes and advanced signal processing algorithms provide a more objective analysis of respiratory sounds. They have also paved the way for the use of automatic decision-making algorithms.
Recent advancements in machine and deep learning algorithms have encouraged researchers working on respiratory sound analysis to develop automated classification systems. Studies in this field fall under two main groups. The first group includes the classification of respiratory diseases, such as asthma, COPD, etc. The second group focuses on classifying respiratory sounds, such as crackle, wheeze, etc. Under these main topics, many valuable studies exist in the literature. Shuvo et al. [1] used a lightweight convolutional neural network (CNN) model, which demonstrates significant efficacy in classifying respiratory auscultation sounds. The model employs a hybrid approach utilizing empirical mode decomposition and continuous wavelet transform, achieving an accuracy of 98.92% in three-class chronic disease classification and 98.70% in six-class pathological classification. Naqvi and Choudhry [2] presented an automated low-cost diagnostic method for COPD and pneumonia, utilizing respiratory sound analysis from the International Conference on Biomedical and Health Informatics (ICBHI) open-access database. The method achieved a classification accuracy of 99.7%. García-Ordás [3] proposed a novel approach utilizing a Variational Convolutional Autoencoder (VAE) combined with a Convolutional Neural Network (CNN) to classify respiratory sounds into healthy, chronic disease, and non-chronic disease categories as well as six specific pathologies. They achieved performance improvements over state-of-the-art methods with a reported F-Score of 0.993 in the ternary classification. Fraiwan et al. [4] investigated the classification of respiratory diseases using respiratory sound signals, achieving an accuracy of 98.27% with boosted decision trees, which outperformed traditional classifiers such as support vector machines. Pham et al. [5] presented a robust deep-learning framework for the analysis of respiratory anomalies and the detection of respiratory diseases using auscultation recordings. They achieved an 84% ICHBI score that averages specificity and sensitivity metrics, which surpasses the previous state-of-the-art result of 72%. In another study by Pham et al. [6], an inception-based deep learning model was developed to detect respiratory anomalies and respiratory diseases from audio recordings, utilizing the ICBHI benchmark dataset. The model achieved 0.53/0.45 ICBHI scores (arithmetic and harmonic averages of sensitivity and specificity) for respiratory anomaly detection and 0.87/0.85 for disease prediction, outperforming several state-of-the-art systems. Kababulut et al. [7] introduced a clinical decision support system for respiratory disease identification using decision tree algorithms and a Shapley-based feature selection to improve performance. Their findings highlight that effective feature selection significantly enhances classification performance in respiratory disease detection. Sfayyih et al. [8] analyzed deep learning applications in respiratory sound analysis, focusing on the effectiveness of CNNs in classifying respiratory sounds. The authors concluded that deep learning techniques show high accuracy in diagnosing respiratory conditions, underscoring AI’s potential in medical diagnostics.
It is widely known that feature extraction has a substantial impact on the efficiency of clinical decision systems. The literature presents numerous diverse feature extraction methods. Classical methods such as Fourier Transform [9], Empirical Mode Decomposition [1], Wavelet Transform [1,6], and Mel-Frequency Cepstral Coefficients (MFCC) [10,11,12,13,14,15] are among the most commonly used methods for feature extraction in the respiratory sound classification field. The proper selection of the most descriptive feature subsets from all extracted features has also been very important for the success of the classification system. Through feature selection, the computational burden on the classifier is reduced by employing a smaller feature set, and in addition classification performance is increased. Therefore, finding effective feature selection (FS) methods has been an extensively studied topic. Filter, wrapper, and embedding techniques are the three general categories into which feature selection methods fall [16]. Kang et al. [17] provide a comprehensive overview of FS techniques, highlighting their significance in managing the challenges posed by high-dimensional datasets. Iqbal et al. [18] presented a comprehensive approach to feature extraction and selection from physiological signals.
In recent years, the application of nature-inspired metaheuristic algorithms for feature selection has gained significant attention within the machine-learning community. Metaheuristic algorithms are primarily used as wrapper-type feature selection methods. Metaheuristic methods range from well-established techniques, like genetic algorithm (GA) and particle swarm optimization (PSO), to newer and more creative approaches, such as the grey wolf optimizer (GWO), teaching learning-based optimization (TLO), Whale Optimization Algorithm (WOA), and the Equilibrium Optimizer (EO). The comprehensive review conducted by Nssibi et al. [19] evaluated various metaheuristic techniques, highlighting their effectiveness in navigating the complex search space associated with feature selection tasks. Sathiyabhama et al. [20] introduced a novel computer-aided diagnosis (CAD) system that employs a GWO and rough set-based approach to identify abnormalities in mammogram images effectively. Kang et al. introduced the Two-Stage Teaching-Learning-Based Optimization (TS-TLBO) algorithm, which demonstrates significant improvements in classification accuracy [17]. Nadimi-Shahraki et al. [21] presented an enhanced version of the Whale Optimization Algorithm (E-WOA) specifically tailored for medical feature selection with a focus on the COVID-19 case study. The experimental results demonstrated that E-WOA significantly outperforms traditional WOA variants and other well-known optimization algorithms. Chen et al. introduced a novel approach that combines particle swarm optimization (PSO) with the 1-nearest neighbor (1-NN) classifier, demonstrating its effectiveness on various life science datasets [22]. Elgamal et al. [23] introduced an enhanced version of the Harris Hawks Optimization (HHO) algorithm, termed Chaotic Harris Hawks Optimization (CHHO), which integrates chaotic maps and simulated annealing (SA) to address the limitations of the standard HHO. Rajammal et al. [24] presented a binary improved grey wolf optimizer (BIGWO) that integrates a mutation operation and an adaptive k-nearest Neighbor (AKNN) algorithm to enhance feature selection efficacy. Prabhakar and Won [25] proposed several innovative techniques, including metaheuristics feature selection methods for classification in telemedicine applications. The study highlights the effectiveness of these methods in analyzing respiratory sounds and showcases the potential for enhanced diagnostic capabilities in healthcare settings. Abedi et al. [26] developed an innovative algorithm that utilizes GA and support vector machine (SVM) classification to analyze thoracic respiratory effort and oximetric signal features. Álvarez et al. [27] conducted a comprehensive study for detecting OSA patients. The authors employed GA for feature selection, achieving very high diagnostic accuracy. All of these studies show that metaheuristic feature selection methods have been successful in addressing the difficulties presented especially by high-dimensional data. Classifier models using features selected by metaheuristic methods enhance prediction accuracy, decrease computing costs, and clarify the process by eliminating unimportant features.
Though successfully used for feature selection in many studies, there is not much research on the utilization of metaheuristic feature selection methods in respiratory disease classification. This study aims to conduct a detailed and comparative analysis of metaheuristic optimization methods in respiratory disease classification. For this purpose, various features were extracted from audio recordings obtained from the publicly available ICBHI 2017 Respiratory Sound Database [28] using 15 frequently used feature extraction techniques. Then, by employing diverse statistical metrics on the collected numerical data, a new feature set was created. Next, to determine the best features that enhance classification performance, six well-known metaheuristics methods were employed with eight transfer functions. Finally, the performances of each method were measured and compared with each other using a simple and identical KNN classifier. In this study, two different classification problems were examined. The first one was a binary classification task (respiratory disease vs healthy) while the second one was a multi-class task (healthy, chronic respiratory disease, nonchronic respiratory disease). Since the database used is highly imbalanced, MCC metric was used as the main performance metrics instead of accuracy. The findings demonstrate that metaheuristic algorithms using correct transfer functions could effectively reduce data dimensionality while enhancing classification accuracy.
This paper is organized as follows: Section 2 provides information about the materials and methods used, including the feature extraction methods from audio recordings, implementation of metaheuristic feature selection methods and transfer functions, classification stage, and evaluation of the results. Section 3 conducts a detailed comparative analysis based on the results obtained. Finally, Section 4 presents the discussion and conclusions of the study.

2. Materials and Methods

2.1. Data Source

The present study utilizes the openly accessible ICBHI 2017 Respiratory Sound Database [28]. ICBHI 2017 comprises 5.5 h of recordings obtained from seven different chest locations, namely trachea, left and right anterior, posterior, and lateral. The recordings encompass 6898 breathing cycles, labeled by respiratory specialists as containing crackles, wheezes, a combination of both, or no abnormal respiratory sounds. The database was compiled from 126 individual participants over several years by two separate study teams located in two countries. The database contains 920 annotated audio samples from 126 participants (77 adult/49 children and 79 male/46 female), which were recorded using heterogeneous types of equipment, namely Meditron, LittC2SE, Litt3200 stethoscopes, and AKGC417L microphones. The age of the participants is 43.0 ± 32.2 years. Participants included patients with lower respiratory tract infections, upper respiratory tract infections, pneumonia, COPD, asthma, bronchiolitis, bronchiectasis, and cystic fibrosis. The respiratory cycles were also categorized into eight distinct conditions by experts: URTI, Chronic Obstructive Pulmonary Disease (COPD), Bronchiectasis, Pneumonia, Bronchiolitis, Asthma, LRTI, and Healthy.

2.2. Feature Extraction

The feature extraction phase was executed utilizing Python’s Librosa library Ver.0.10 [29]. The default value of the library was utilized as the number of features to be extracted for each method. In this study, 15 distinct methods were employed for the feature extraction procedure. The specific characteristics of these methods are displayed in Table 1 [30]. Since the number of features obtained from each method is very large (several thousands), a new and reduced feature set was generated by computing five statistical measures (minimum, maximum, mean, standard deviation, and skewness) of feature values for each feature extraction method. Then, these features were concatenated to form the feature vector for each sample. This way, each sample is represented by 15 × 5 = 75 feature values.

2.3. Classification Cases Examined

Once the feature extraction step was completed, feature selection and classification stages followed. This study examines two different cases. Case 1 examines a binary classification task where respiratory sound recordings are categorized into two classes diseased and healthy. The disease class consists of recordings that belong to individuals with seven respiratory diseases specified in Section 2.1; the healthy class comprises data collected from individuals who do not have any respiratory illnesses. Case 2 pertains to a multi-class categorization with three distinct classes: Chronic Respiratory Diseased, Non-chronic Respiratory Diseased, and Healthy. The Chronic Diseased class was formed by amalgamating instances of individuals with COPD, asthma, and bronchiectasis, which are representative of chronic respiratory conditions. The Non-chronic Diseased class consists of respiratory sound recordings obtained from individuals diagnosed with pneumonia, URTI, bronchiolitis, and LRTI. The scenario for the healthy class is analogous to Case 1. Table 2 provides information about the datasets utilized.

2.4. Feature Selection by Metaheuristic Methods

This study examines six different metaheuristic methods with various transfer functions to identify the most effective approach for classifying respiratory diseases from respiratory sounds. The methods include genetic algorithm (GA), particle swarm optimization (PSO), grey wolf optimization (GWO), teaching learning-based optimization (TLO), Whale Optimization Algorithm (WOA), and the Equilibrium Optimizer (EO). Each algorithm was executed in 25 trials using identical experimental settings to achieve statistically meaningful outcomes. In each trial, the population was initialized with a size of 100 individuals, and the simulation was run for 50 iterations.
Metaheuristic algorithms inherently generate continuous solutions that lie within the upper and lower boundaries of the search space. One of the primary strategies to discretize these approaches is to utilize transfer functions. The primary objective of transfer functions is to acquire a 0–1 vector representing the characteristics to be chosen for feature selection while keeping the original procedure unchanged. The continuous values produced by the transfer function are transformed into either 0 or 1 using a thresholding operation. Thus, this approach yields a solution or possible answers to the binary optimization issue of identifying the most efficient characteristics for diagnosing respiratory disease. Transfer functions exhibit distinct characteristics, and when employed for discretization, varying outcomes are bound to arise [31]. Various transfer functions have been examined in many studies, but it remains necessary to make comparisons to determine which specific transfer functions should be employed to achieve discretization. To this end, this study utilizes a total of eight commonly employed transfer functions, which belong to two distinct families, namely S-shaped and V-shaped functions. Formulas for S-shaped transfer functions S1 to S4 and V-shaped transfer functions V1 to V4 are provided in Table 3. The function curves for them are given in Figure 1.
Our study introduces a wrapper feature selection strategy that utilizes the aforementioned metaheuristic search algorithms and the KNN classifier as the evaluator. When using feature selection methods, it is crucial to consider both a solution’s representation and the optimization process’s evaluation. The dataset used in this study exhibits a serious class imbalance issue. Thus, accurate prediction of minority classes is also crucial for a reliable model. Therefore, we used the Matthews correlation coefficient (MCC) as the fitness value for the optimization problem since the MCC metric is more suitable for distinguishing between various misclassification distributions in datasets with imbalanced class issues [32].

2.5. Classifier

When the studies in the field of feature selection using metaheuristic methods are analyzed, it is seen that the KNN algorithm has been frequently utilized as the classifier [33,34,35,36]. This is due to its simplicity and ease of application, as well as its ability to provide fairly accurate results. This is due to its simplicity, practicality, ease of application and use, as well as its ability to provide rapid and accurate results when dealing with large datasets. For the same reasons, a simple KNN (K = 5) classifier is utilized in this study. This classifier functions as a decision-maker during both the feature selection and classification phases. To make the comparisons fair, the same classifier was used in all trials.

3. Results

This study conducted experiments for comparing six different metaheuristics feature selection methods using eight different transfer functions from two families (V-shaped and S-shaped) to determine the best feature selection methods to be used for respiratory disease classification problems. This means finding the methods that lead to high classification performance while using a small number of features. Findings are examined, evaluated, and commented on below under three headings: transfer function fitness values, comparison of the classification performances of metaheuristic feature selection methods, and comparison with traditional feature selection methods.

3.1. Fitness Values of Transfer Functions for Each Feature Selection Method

During the feature selection phase, the classification model is evaluated using the test data, which is a randomly retained 20% of all data. The fitness values are computed by comparing the actual values with the predicted results. The fitness value is a numerical measure that quantifies the quality of a solution candidate with respect to the optimization problem at hand. In our study, fitness function takes into consideration both a classification performance metric, which is selected as MCC, and also the number of features selected. Therefore, it simultaneously tries to maximize MCC value while minimizing the number of features selected. This feature selection process is repeated 25 times for each transfer function for each optimization method. Since there is an issue of imbalance in the class labels for both Case 1 and Case 2, the MCC metric was used to utilize the results. Figure 2 and Figure 3 display the fitness values of eight transfer functions for each feature selection method obtained after 25 trials for Case 1 and Case 2, respectively.
The fitness values in Figure 2 and Figure 3 show the overall performance of the transfer functions. As can be seen in Section 3.3, there are significantly fewer number of features selected for V-shaped transfer functions, while MCC values are slightly higher for S-shaped transfer functions. Therefore, considering both factors, the fitness values of V-shaped functions are slightly higher. In addition, the vertical size of the plots, which represent the dispersion of fitness values in 25 trials, shows that V-shaped functions are generally more stable than S-shaped ones. This implies that the fitness values of V-shaped functions do not change much from trial to trial.

3.2. Comparison of Classification Performance of Metaheuristic Feature Selection Methods

Once the feature selection process was concluded, the KNN classifier was fed by the selected features for testing. The same test set used in the FS phase is also used for classification. However, whereas each sample is represented by 75 features in the FS phase, the number of features used for each sample in the classification stage is less than 75, depending on the transfer function used. Classification conditions are maintained consistent by using the same random state rate, which is chosen as 42, and identical test data for the evaluation of all methods. Evaluation of the classification results is made based on MCC scores since the dataset is highly imbalanced. Each metaheuristic method was executed 25 times for each of the eight different transfer functions (S1–S4, V1–V4). For each run, an MCC value was calculated, resulting in 200 MCC values per method. To compare the effects of the S and V-shaped transfer function families on classification performance we calculated the mean and standard deviation of 100 MCC scores obtained for both families for each metaheuristic method. The mean and standard deviation of MCC values for each transfer function family are presented in Table 4 for Case 1 and Table 5 for Case 2.
The bold-styled values in these tables represent the best average scores obtained for each transfer function family. The performance results of each individual transfer function are also presented in a graphical format in Figure 4 and Figure 5. It can be observed that methods using S-shaped transfer functions obtain higher classification performance than those using V-shaped ones.

3.3. Comparison of Metaheuristic Feature Selection Methods Based on Number of Features Selected

In addition to classification performance, the number of features selected by metaheuristic algorithms for different transfer functions was also investigated and compared. Every metaheuristic method for each of the eight different transfer functions is employed 25 times and statistical averages (mean and standard deviation) are calculated for each of these 25 runs for fair assessment. We first compared the effects of transfer function families. In other words, we took the average of scores obtained for all V-shaped functions; we also did the same for all S-shaped functions. The results are presented in Table 6 and Table 7 for Case 1 and Case 2, respectively.
The dispersion of selected feature size in one of 25 trials for each individual transfer function is also presented as a scatter plot in Figure 6 and Figure 7, for Case 1 and 2, respectively. It can be observed that methods using V-shaped functions use significantly fewer features than those using S-shaped ones.

3.4. Comparison of Metaheuristic Feature Selection Methods with Classical Feature Selection Methods

In this section, the effectiveness of metaheuristic feature selection methods is compared with the following classical feature selection methods: Spearman’s correlation (Spearman), mutual information (MI), relief, variance threshold (VAR), mean absolute difference (MAD), dispersion ratio (DR), lasso, tree-based, recursive feature elimination (RFE), and sequential forward selection (SFS). These classical methods were implemented using the Mafese library [37], which is an open-source Python library. The number of features to be selected was determined by ranking them based on their MCC scores. This was achieved by calculating all possible combinations of features within the range of minimum and maximum number of features for each method. The number of features that lead to the highest MCC value was taken as the optimum feature size for that method. The performance values are presented in Table 8 and Table 9 for Cases 1 and 2, respectively. The bold types represent the highest scores for each metric. In order to compare the results of classical feature selection methods with metaheuristic ones, the highest average performance values, obtained over 25 trials, for each metaheuristic method are presented in Table 10 and Table 11 for Cases 1 and 2, respectively. These tables also show with which transfer functions (one for S-shaped functions and the other for V-shaped functions), these values are obtained. Table 8, Table 9, Table 10 and Table 11 also represent the number of features selected by each method and their computation time.

4. Discussion

In this study, we explored and compared the performances of nature-inspired metaheuristic algorithms for feature selection in respiratory disease classification. The main objective is to enhance classification accuracy while reducing the feature set size, thereby improving the model’s computational efficiency. The experiments were conducted on a dataset consisting of respiratory sound recordings belonging to different diseases. Our approach utilized a variety of metaheuristic algorithms and two popular families of transfer functions to select the most relevant features. We first analyzed the fitness values of eight different transfer functions belonging to two families for each metaheuristic feature selection algorithm. It is seen that fitness values of V-shaped functions exhibit slightly higher values than S-shaped ones. In addition, the vertical size of the plots, which represent the dispersion of fitness values obtained in 25 trials (Figure 6 and Figure 7), shows that V-shaped functions are generally more stable than S-shaped ones. In other words, fitness values of V-shaped functions do not change much from trial to trial. Then we analyzed the effects of these MHA and transfer functions on the classification performance for two different cases. It reveals that MHA methods using S-shaped transfer functions obtain slightly better classification performance than the same methods when they use V-shaped functions. However, using V-shaped transfer functions has two important advantages. First, they are much more stable, which means their fitness values do not change much from trial to trial. Second, the number of features selected by MHA methods when employing V-shaped transfer functions is significantly less than the same methods when they use S-shaped functions.
Next, we compared the performances of MHA-based feature selection methods with classical feature selection methods based on MCC metric and number of features selected. Upon analysis of the results, it is seen that the proposed MHA-based feature selection methods obtain better results than most, if not all, of the classical methods in both binary and multi-class cases. The “Sequential” feature selection method, which is a classical one, obtained the highest scores in both cases. In addition, two other classical methods, “Spearman” and “Recursive,” obtained the second highest scores in Case 1 and Case 2, respectively. However, these methods select a large number of features, increasing the computational burden on the classification system. We also measured computational time of each method for the feature selection process. The MHA-based methods’ computational time is found to be significantly lower than FFS (Sequential) method. These results show that if, in addition to classification performance, small number of features and low computational time are required, then MHA feature selection methods (in particular, GWO and PSO) using V-shaped transfer functions should be chosen.
There is one point that needs explanation here. The accuracy values obtained by classical feature selection methods are higher than MHA-based methods. This result can lead to the conclusion that metaheuristic methods are worse. However, we want to emphasize that the Accuracy metric could be misleading in datasets with class imbalances. Therefore, in such datasets, metrics such as MCC, F1 or Balanced Accuracy should be used. Our study uses a highly imbalanced dataset. Thus, we used one of these metrics, i.e., MCC, for measuring classification performances. This way, we have derived models that can also accurately forecast minority classes.
A possible limitation of this study is that it was conducted using a single database. The results that would be obtained with other databases could be different than the ones obtained here. Another limitation is due to the database used since some disease classes are represented by very small numbers of data. Still another drawback is the use of only six MHA; although these are the most-used ones in the literature, there are other MHA that are not included in this study.

5. Conclusions

The main goal of this detailed comparison study is to examine the performance of different nature-inspired MHA for feature selection in respiratory disease classification. The results show that MHA-based feature selection algorithms provide robust mechanisms for navigating the high-dimensional feature space, ensuring that the selected features contribute maximally to the classification task. Furthermore, the reduction in the number of features, especially by using V-shaped transfer functions, not only decreases the computational load but also minimizes the risk of overfitting. This is particularly important in the context of respiratory disease classification, where the diversity and variability of the data can lead to complex decision boundaries. The overall improvement in classification accuracy and computational efficiency underscores the effectiveness of these algorithms in feature selection for respiratory disease classification. Future work could explore the combination of these algorithms with other advanced machine learning techniques, such as ensemble learning, to further enhance the robustness and generalizability of the models.

Author Contributions

Conceptualization, D.G.K., N.Ö. and O.D.; methodology, F.Y.K. and N.Ö.; software, N.Ö. and O.D.; validation, N.Ö., O.D. and F.Y.K.; formal analysis, D.G.K.; investigation, D.G.K. and M.K.; resources, F.Y.K.; data curation, N.Ö.; writing—original draft preparation, O.D., N.Ö. and F.Y.K.; writing—review and editing, D.G.K. and M.K.; visualization, N.Ö. and O.D.; supervision, D.G.K. and M.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are openly available in [ICBHI 2017] at [https://doi.org/10.1007/978-981-10-7419-6_6], reference number [28].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shuvo, S.B.; Ali, S.N.; Swapnil, S.I.; Hasan, T.; Bhuiyan, M.I.H. A Lightweight CNN Model for Detecting Respiratory Diseases from Lung Auscultation Sounds Using EMD-CWT-Based Hybrid Scalogram. IEEE J. Biomed. Health Inform. 2021, 25, 2595–2603. [Google Scholar] [CrossRef] [PubMed]
  2. Naqvi, S.Z.H.; Choudhry, M.A. An automated system for classification of chronic obstructive pulmonary disease and pneumonia patients using lung sound analysis. Sensors 2020, 20, 6512. [Google Scholar] [CrossRef] [PubMed]
  3. García-Ordás, M.T.; Benítez-Andrades, J.A.; García-Rodríguez, I.; Benavides, C.; Alaiz-Moretón, H. Detecting respiratory pathologies using convolutional neural networks and variational autoencoders for unbalancing data. Sensors 2020, 20, 1214. [Google Scholar] [CrossRef] [PubMed]
  4. Fraiwan, L.; Hassanin, O.; Fraiwan, M.; Khassawneh, B.; Ibnian, A.M.; Alkhodari, M. Automatic identification of respiratory diseases from stethoscopic lung sound signals using ensemble classifiers. Biocybern. Biomed. Eng. 2020, 41, 1–14. [Google Scholar] [CrossRef]
  5. Pham, L.; Phan, H.; Palaniappan, R.; Mertins, A.; McLoughlin, I. CNN-MoE Based Framework for Classification of Respiratory Anomalies and Lung Disease Detection. April 2020. Available online: http://arxiv.org/abs/2004.04072 (accessed on 1 September 2024).
  6. Pham, L.; Phan, H.; Schindler, A.; King, R.; Mertins, A.; McLoughlin, I. Inception-Based Network and Multi-Spectrogram Ensemble Applied For Predicting Respiratory Anomalies and Lung Diseases. 2020. Available online: http://arxiv.org/abs/2012.13699 (accessed on 1 September 2024).
  7. Kababulut, F.Y.; Kuntalp, D.G.; Düzyel, O.; Özcan, N.; Kuntalp, M. A New Shapley-Based Feature Selection Method in a Clinical Decision Support System for the Identification of Lung Diseases. Diagnostics 2023, 13, 3558. [Google Scholar] [CrossRef]
  8. Sfayyih, A.H.; Sabry, A.H.; Jameel, S.M.; Sulaiman, N.; Raafat, S.M.; Humaidi, A.J.; Al Kubaiaisi, Y.M. Acoustic-Based Deep Learning Architectures for Lung Disease Diagnosis: A Comprehensive Overview. Diagnostics 2023, 13, 1748. [Google Scholar] [CrossRef]
  9. Manir, S.B.; Karim, M.; Kiber, A. Assessment of Lung Diseases from Features Extraction of Breath Sounds Using Digital Signal Processing Methods. In Proceedings of the 2020 Emerging Technology in Computing, Communication and Electronics (ETCCE), Dhaka, Bangladesh, 21–22 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
  10. Bardou, D.; Zhang, K.; Ahmad, S.M. Lung sounds classification using convolutional neural networks. Artif. Intell. Med. 2018, 88, 58–69. [Google Scholar] [CrossRef]
  11. Jaber, M.M.; Abd, S.K.; Shakeel, P.; Burhanuddin, M.; Mohammed, M.A.; Yussof, S. A telemedicine tool framework for lung sounds classification using ensemble classifier algorithms. Measurement 2020, 162, 107883. [Google Scholar] [CrossRef]
  12. Engin, M.A.; Aras, S.; Gangal, A. Extraction of low-dimensional features for single-channel common lung sound classification. Med. Biol. Eng. Comput. 2022, 60, 1555–1568. [Google Scholar] [CrossRef]
  13. Rocha, B.M.; Pessoa, D.; Marques, A.; Carvalho, P.; Paiva, R.P. Automatic classification of adventitious respiratory sounds: A (un)solved problem? Sensors 2021, 21, 57. [Google Scholar] [CrossRef]
  14. Sengupta, N.; Sahidullah; Saha, G. Lung sound classification using cepstral-based statistical features. Comput. Biol. Med. 2016, 75, 118–129. [Google Scholar] [CrossRef] [PubMed]
  15. Srivastava, A.; Jain, S.; Miranda, R.; Patil, S.; Pandya, S.; Kotecha, K. Deep learning based respiratory sound analysis for detection of chronic obstructive pulmonary disease. PeerJ Comput. Sci. 2021, 7, e369. [Google Scholar] [CrossRef] [PubMed]
  16. Wang, L.; Wang, Y.; Chang, Q. Feature selection methods for big data bioinformatics: A survey from the search perspective. Methods 2016, 111, 21–31. [Google Scholar] [CrossRef] [PubMed]
  17. Kang, Y.; Wang, H.; Pu, B.; Tao, L.; Chen, J.; Yu, P.S. A Hybrid Two-Stage Teaching-Learning-Based Optimization Algorithm for Feature Selection in Bioinformatics. IEEE/ACM Trans. Comput. Biol. Bioinform. 2022, 20, 1746–1760. [Google Scholar] [CrossRef] [PubMed]
  18. Iqbal, T.; Elahi, A.; Wijns, W.; Amin, B.; Shahzad, A. Improved Stress Classification Using Automatic Feature Selection from Heart Rate and Respiratory Rate Time Signals. Appl. Sci. 2023, 13, 2950. [Google Scholar] [CrossRef]
  19. Nssibi, M.; Manita, G.; Korbaa, O. Advances in nature-inspired metaheuristic optimization for feature selection problem: A comprehensive survey. Comput. Sci. Rev. 2023, 49, 100559. [Google Scholar] [CrossRef]
  20. Sathiyabhama, B.; Kumar, S.U.; Jayanthi, J.; Sathiya, T.; Ilavarasi, A.K.; Yuvarajan, V.; Gopikrishna, K. A novel feature selection framework based on grey wolf optimizer for mammogram image analysis. Neural Comput. Appl. 2021, 33, 14583–14602. [Google Scholar] [CrossRef]
  21. Nadimi-Shahraki, M.H.; Zamani, H.; Mirjalili, S. Enhanced whale optimization algorithm for medical feature selection: A COVID-19 case study. Comput. Biol. Med. 2022, 148, 105858. [Google Scholar] [CrossRef]
  22. Chen, L.-F.; Su, C.-T.; Chen, K.-H.; Wang, P.-C. Particle swarm optimization for feature selection with application in obstructive sleep apnea diagnosis. Neural Comput. Appl. 2012, 21, 2087–2096. [Google Scholar] [CrossRef]
  23. Elgamal, Z.M.; Yasin, N.B.M.; Tubishat, M.; Alswaitti, M.; Mirjalili, S. An improved harris hawks optimization algorithm with simulated annealing for feature selection in the medical field. IEEE Access 2020, 8, 186638–186652. [Google Scholar] [CrossRef]
  24. Rajammal, R.R.; Mirjalili, S.; Ekambaram, G.; Palanisamy, N. Binary Grey Wolf Optimizer with Mutation and Adaptive K-nearest Neighbour for Feature Selection in Parkinson’s Disease Diagnosis. Knowl.-Based Syst. 2022, 246, 108701. [Google Scholar] [CrossRef]
  25. SPrabhakar, S.K.; Won, D.-O. HISET: Hybrid interpretable strategies with ensemble techniques for respiratory sound classification. Heliyon 2023, 9, e18466. [Google Scholar] [CrossRef] [PubMed]
  26. Abedi, Z.; Naghavi, N.; Rezaeitalab, F. Detection and classification of sleep apnea using genetic algorithms and SVM-based classification of thoracic respiratory effort and oximetric signal features. Comput. Intell. 2017, 33, 1005–1018. [Google Scholar] [CrossRef]
  27. Álvarez, D.; Hornero, R.; Marcos, J.V.; del Campo, F. Feature selection from nocturnal oximetry using genetic algorithms to assist in obstructive sleep apnoea diagnosis. Med. Eng. Phys. 2012, 34, 1049–1057. [Google Scholar] [CrossRef]
  28. Rocha, B.M.; Filos, D.; Mendes, L.; Vogiatzis, I.; Perantoni, E.; Kaimakamis, E.; Natsiavas, P.; Oliveira, A.; Jácome, C.; Marques, A.; et al. A Respiratory Sound Database for the Development of Automated Classification. Precision Medicine Powered by PHealth and Connected Health. ICBHI. 2017. Available online: https://bhichallenge.med.auth.gr/ (accessed on 1 September 2024).
  29. Librosa. GitHub—Librosa/Librosa: Python Library for Audio and Music Analysis. GitHub. Available online: https://github.com/librosa/librosa (accessed on 1 September 2024).
  30. Feature Extraction—Librosa 0.10.2 Documentation. Available online: https://librosa.org/doc/latest/feature.html (accessed on 1 September 2024).
  31. Wang, L.; He, Y.; Wang, X.; Zhou, Z.; Ouyang, H.; Mirjalili, S. A novel discrete differential evolution algorithm combining transfer function with modulo operation for solving the multiple knapsack problem. Inf. Sci. 2024, 680, 121170. [Google Scholar] [CrossRef]
  32. Jurman, G.; Riccadonna, S.; Furlanello, C. A comparison of MCC and CEN error measures in multi-class prediction. PLoS ONE 2012, 7, e41882. [Google Scholar] [CrossRef]
  33. Alweshah, M.; Alkhalaileh, S.; Al-Betar, M.A.; Abu Bakar, A. Coronavirus herd immunity optimizer with greedy crossover for feature selection in medical diagnosis. Knowl.-Based Syst. 2022, 235, 107629. [Google Scholar] [CrossRef]
  34. Nourmohammadi-Khiarak, J.; Feizi-Derakhshi, M.-R.; Behrouzi, K.; Mazaheri, S.; Zamani-Harghalani, Y.; Tayebi, R.M. New hybrid method for heart disease diagnosis utilizing optimization algorithm in feature selection. Health Technol. 2020, 10, 667–678. [Google Scholar] [CrossRef]
  35. Oğur, N.B.; Kotan, M.; Balta, D.; Yavuz, B.; Oğur, Y.S.; Yuvacı, H.U.; Yazıcı, E. Detection of depression and anxiety in the perinatal period using Marine Predators Algorithm and kNN. Comput. Biol. Med. 2023, 161, 107003. [Google Scholar] [CrossRef]
  36. Wang, A.; An, N.; Chen, G.; Li, L.; Alterovitz, G. Accelerating wrapper-based feature selection with K-nearest-neighbor. Knowl.-Based Syst. 2015, 83, 81–91. [Google Scholar] [CrossRef]
  37. Van Thieu, N.; Nguyen, N.H.; Heidari, A.A. Feature selection using metaheuristics made easy: Open source MAFESE library in Python. Futur. Gener. Comput. Syst. 2024, 160, 340–358. [Google Scholar] [CrossRef]
Figure 1. V-shaped and S-shaped transfer functions.
Figure 1. V-shaped and S-shaped transfer functions.
Diagnostics 14 02244 g001
Figure 2. Fitness values of different transfer functions for each MHA FS method for Case 1.
Figure 2. Fitness values of different transfer functions for each MHA FS method for Case 1.
Diagnostics 14 02244 g002
Figure 3. Fitness values of different transfer functions for each MHA FS method for Case 2.
Figure 3. Fitness values of different transfer functions for each MHA FS method for Case 2.
Diagnostics 14 02244 g003
Figure 4. Comparison of MHA FS methods for individual transfer functions for Case 1.
Figure 4. Comparison of MHA FS methods for individual transfer functions for Case 1.
Diagnostics 14 02244 g004
Figure 5. Comparison of MHA FS methods for individual transfer functions for Case 2.
Figure 5. Comparison of MHA FS methods for individual transfer functions for Case 2.
Diagnostics 14 02244 g005
Figure 6. Selected feature sizes of MHA FS methods for S- and V-shaped transfer functions for Case 1.
Figure 6. Selected feature sizes of MHA FS methods for S- and V-shaped transfer functions for Case 1.
Diagnostics 14 02244 g006
Figure 7. Selected feature sizes of MHA methods for S- and V-shaped transfer functions for Case 2.
Figure 7. Selected feature sizes of MHA methods for S- and V-shaped transfer functions for Case 2.
Diagnostics 14 02244 g007
Table 1. Feature extraction methods employed in the study.
Table 1. Feature extraction methods employed in the study.
No.MethodAbbreviationExplanation
1Chroma Short Time Fourier TransformChro_stftIt calculates a chromatogram by analyzing either a waveform or a power spectrogram.
2Chroma Constant-Q chromagramChro_cqtIt calculates the Constant-Q chromagram.
3Chroma Energy NormalizedChro_censIt calculates the chroma variant known as Chroma Energy Normalized.
4Chroma Variable-Q chromagramChro_vqtIt calculates the Variable-Q chromagram.
5Mel-Frequency Cepstral CoefficientsMfccIt calculates the mel-frequency cepstral coefficients
6Root Mean SquareRmsIt calculates the root mean square value for each frame, using either the audio samples or the spectrogram.
7Spectral CentroidSpec_centIt calculates the spectral centroid. Every individual frame of a magnitude spectrogram is standardized and considered as a distribution across frequency bins. From this distribution, the average value is calculated for each frame.
8Spectral BandwidthSpec_bandwIt calculates the spectral bandwidth.
9Spectral ContrastSpec_contIt calculates the spectral content of each frame by dividing it into sub-bands. The energy contrast for each sub-band is determined by comparing the average energy in the highest quantile (peak energy) with that of the lowest quantile (valley energy).
10Spectral FlatnessSpec_flatIt calculates the spectral flatness. Spectral flatness, also known as tonality coefficient, is a metric used to quantify the degree to which a sound resembles noise rather than a distinct tone. A spectral flatness value closer to 1.0 indicates that the spectrum is more like white noise.
11Spectral Roll-off FrequencySpec_rofIt calculates the roll-off frequency. The roll-off frequency is determined for each frame as the central frequency of a spectrogram bin that contains at least roll percent of the energy of the spectrum in that frame, as well as the bins below it.
12Polynomial FeaturesPoly_feaIt obtains the coefficients for fitting a polynomial to the individual columns of a spectrogram.
13Tonal Centroid FeaturesTonnetzIt calculates the tonal centroid characteristics. This representation employs the technique of projecting chroma features onto a six-dimensional basis, where the perfect fifth, minor third, and major third are each represented by two-dimensional coordinates.
14Zero Crossing RateZero_crIt calculates the rate at which the audio time series crosses the zero axis.
Table 2. Distribution of data in Case 1 and Case 2.
Table 2. Distribution of data in Case 1 and Case 2.
CaseClassSample SizeNumber of Features
1Healthy3575
Diseased885
2Healthy35
Chronic diseased810
Non-chronic diseased75
Table 3. V-shaped and S-shaped transfer functions.
Table 3. V-shaped and S-shaped transfer functions.
V-ShapeS-Shape
NameFormulaNameFormula
V1 V 1 ( x ) = | e r f ( π 2 x ) | S1 S 1 ( x ) = 1 1 + e 2 x
V2 V 2 ( x ) = | t a n h ( x ) | S2 S 2 ( x ) = 1 1 + e x
V3 V 3 ( x ) = | x 1 + x 2 ) | S3 S 3 ( x ) = 1 1 + e ( x 2 )
V4 V 4 ( x ) = | 2 π a r c t a n ( π 2 x ) | S4 S 4 ( x ) = 1 1 + e ( x 3 )
Table 4. Comparison of Metaheuristic Algorithms for S- and V-shaped Transfer Function Families for Case 1.
Table 4. Comparison of Metaheuristic Algorithms for S- and V-shaped Transfer Function Families for Case 1.
ShapeMethodMean MCCStd. D. of MCC
SEO0.59090.1211
GA0.56270.1206
GWO0.58030.1136
PSO0.59560.1043
TLO0.61710.1055
WOA0.53920.1174
VEO0.49160.1335
GA0.41930.1704
GWO0.51910.1476
PSO0.53950.1621
TLO0.47670.1891
WOA0.42720.1626
Numbers given in bold are the best scores of S and V shaped functions.
Table 5. Comparison of Metaheuristic Algorithms for S- and V-shaped Transfer Function Families for Case 2.
Table 5. Comparison of Metaheuristic Algorithms for S- and V-shaped Transfer Function Families for Case 2.
ShapeMethodMean MCCStd. D. of MCC
SEO0.69270.0528
GA0.67300.0548
GWO0.69180.0523
PSO0.68820.0414
TLO0.69050.0466
WOA0.67680.0550
VEO0.65900.0410
GA0.62920.0604
GWO0.66570.0345
PSO0.67090.0445
TLO0.66880.0326
WOA0.62620.0667
Numbers given in bold represents the best scores for S and V shapes in terms of Mean MCC metric
Table 6. Mean and standard deviation of number of selected features of MHA FS methods for Case 1.
Table 6. Mean and standard deviation of number of selected features of MHA FS methods for Case 1.
MethodShapeMeanStd. D.ShapeMeanStd. D.
EOS30.904.49V3.210.94
GA31.023.483.031.00
GWO30.563.753.440.87
PSO27.133.538.473.44
TLO30.853.843.931.59
WOA19.735.882.961.00
Table 7. Mean and standard deviation of number of selected features of MHA FS methods for Case 2.
Table 7. Mean and standard deviation of number of selected features of MHA FS methods for Case 2.
MethodShapeMeanStd. D.ShapeMeanStd. D.
EOS39.575.29V5.830.59
GA42.567.965.721.33
GWO39.785.645.900.44
PSO30.713.0518.313.57
TLO36.993.696.171.00
WOA33.265.555.550.77
Table 8. Best performances of Classical FS methods for Case 1.
Table 8. Best performances of Classical FS methods for Case 1.
MethodACCF1 MacroMCCNumber of FeaturesComp. Time (s)
DR0.9780.7440.506171.21
Lasso0.9670.492−0.01231.24
MAD0.9730.4930.00072.77
MI0.9780.6610.442916.28
RELIEF0.9730.4930.000760.43
Recursive0.9780.7440.50631758.31
SPEARMAN0.9890.8720.770552.68
Sequential0.9890.8970.794472597.45
Tree0.9780.6610.442116.26
VAR0.9730.4930.00070.10
Numbers given in bold represents the best values achieved for ACC, F1 Macro, and MCC metrics.
Table 9. Best performances of Classical FS methods for Case 2.
Table 9. Best performances of Classical FS methods for Case 2.
MethodACCF1 MacroMCCNumber of Features Comp. Time
(s)
DR0.9180.6890.597181.46
Lasso0.8420.3520.05451.22
MAD0.8910.5110.528142.89
MI0.8910.5110.5285717.35
RELIEF0.8910.5110.5281458.73
Recursive0.9400.6770.71139941.72
SPEARMAN0.8910.5110.528722.63
Sequential0.9620.7970.829672590.82
Tree0.9080.5150.533275.87
VAR0.8910.5110.528150.15
Numbers given in bold represents the best values achieved for ACC, F1 Macro, and MCC metrics.
Table 10. Best results according to the average of 25 trials of MHA FS methods for Case 1.
Table 10. Best results according to the average of 25 trials of MHA FS methods for Case 1.
MethodTransfer FunctionMCCBalanced AccuracyACCF1 MacroAverage Number of FeaturesComp. Time (s)
EOS40.6280.6500.9180.63029.92150.98
V30.5350.6130.9050.6123.36108.22
GAS20.6060.6310.9140.61633.44152.08
V20.4480.5640.8910.5673.20102.35
GWOS40.6300.6510.9180.62930.52146.72
V20.5690.6390.9110.6373.40100.77
PSOS40.6160.6580.9160.64328.08150.78
V10.5640.6190.9080.6109.24140.44
TLOS30.6380.6590.9200.63629.04508.87
V10.5250.5970.9050.5904.20232.92
WOAS30.5570.6110.9030.59925.56285.29
V20.4360.5380.8880.5363.16124.85
Bold numbers represents the best value achieved for MM, Balanced Accuracy, ACC, and F1 Macro metrics.
Table 11. Best results according to average of 25 trials of MHA FS methods for Case 2.
Table 11. Best results according to average of 25 trials of MHA FS methods for Case 2.
MethodTransfer FunctionMCCBalanced AccuracyACCF1 MacroAverage Number of FeaturesComp. Time (s)
EOS20.6980.6950.9370.70440.04280.76
V20.6690.6890.9310.6985.92188.76
GAS20.6840.6750.9340.68343.44277.84
V20.6520.6760.9280.6855.92186.54
GWOS40.7080.7090.9380.71637.6151.40
V10.6740.6960.9310.7066.00128.04
PSOS40.6960.7020.9360.71331.88202.25
V20.6810.6980.9330.71217.76251.96
TLOS40.7070.7030.9380.71034.68342.42
V10.6750.6940.9310.7015.96423.54
WOAS40.6970.6890.9370.69734.88220.66
V40.6280.6570.9240.6565.56216.66
Bold numbers represents the best value achieved for MM, Balanced Accuracy, ACC, and F1 Macro metrics.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gürkan Kuntalp, D.; Özcan, N.; Düzyel, O.; Kababulut, F.Y.; Kuntalp, M. A Comparative Study of Metaheuristic Feature Selection Algorithms for Respiratory Disease Classification. Diagnostics 2024, 14, 2244. https://doi.org/10.3390/diagnostics14192244

AMA Style

Gürkan Kuntalp D, Özcan N, Düzyel O, Kababulut FY, Kuntalp M. A Comparative Study of Metaheuristic Feature Selection Algorithms for Respiratory Disease Classification. Diagnostics. 2024; 14(19):2244. https://doi.org/10.3390/diagnostics14192244

Chicago/Turabian Style

Gürkan Kuntalp, Damla, Nermin Özcan, Okan Düzyel, Fevzi Yasin Kababulut, and Mehmet Kuntalp. 2024. "A Comparative Study of Metaheuristic Feature Selection Algorithms for Respiratory Disease Classification" Diagnostics 14, no. 19: 2244. https://doi.org/10.3390/diagnostics14192244

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop