Next Article in Journal
Biomechanical Multipurpose Miniscrew Strategy for Simultaneous Distalization in Class II Patients—The BiGa System
Previous Article in Journal
Development of a New Aggregation Method to Remove Nanoplastics from the Ocean: Proof of Concept Using Mussel Exposure Tests
Previous Article in Special Issue
Biomineralization of Polyelectrolyte-Functionalized Electrospun Fibers: Optimization and In Vitro Validation for Bone Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Survival Prediction of Patients after Heart Attack and Breast Cancer Surgery with a Hybrid Model Built with Particle Swarm Optimization, Stacked AutoEncoders, and the Softmax Classifier

by
Mehmet Akif Bülbül
1,* and
Mehmet Fatih Işık
2
1
Department of Software Engineering, Kayseri University, Kayseri 38280, Turkey
2
Department of Electrical-Electronics Engineering, Hitit University, Çorum 19030, Turkey
*
Author to whom correspondence should be addressed.
Biomimetics 2024, 9(5), 304; https://doi.org/10.3390/biomimetics9050304
Submission received: 23 April 2024 / Revised: 15 May 2024 / Accepted: 17 May 2024 / Published: 19 May 2024
(This article belongs to the Special Issue Biomimetic Approaches in Healthcare—Innovations Inspired by Nature)

Abstract

:
The prediction of patient survival is crucial for guiding the treatment process in healthcare. Healthcare professionals rely on analyzing patients’ clinical characteristics and findings to determine treatment plans, making accurate predictions essential for efficient resource utilization and optimal patient support during recovery. In this study, a hybrid architecture combining Stacked AutoEncoders, Particle Swarm Optimization, and the Softmax Classifier was developed for predicting patient survival. The architecture was evaluated using the Haberman’s Survival dataset and the Echocardiogram dataset from UCI. The results were compared with several Machine Learning methods, including Decision Trees, K-Nearest Neighbors, Support Vector Machines, Neural Networks, Gradient Boosting, and Gradient Bagging applied to the same datasets. The findings indicate that the proposed architecture outperforms other Machine Learning methods in predicting patient survival for both datasets and surpasses the results reported in the literature for the Haberman’s Survival dataset. In the light of the findings obtained, the models obtained with the proposed architecture can be used as a decision support system in determining patient care and applied methods.

1. Introduction

Health is paramount in people’s lives, representing the most invaluable aspect. However, health-related adversities often necessitate extensive treatment processes. Particularly, diseases requiring prolonged treatment or exhibiting a low probability of recovery can impose significant challenges on individuals. Such challenges underscore the critical importance of early detection and effective management, particularly for diseases amenable to treatment. Surgery, in particular, marks a pivotal juncture in disease management [1]. Predicting survival is an important parameter for understanding patient treatment and recovery. Predictions are essential to help healthcare providers develop treatment plans and evaluate treatment outcomes. Accurate prediction is crucial for developing the most appropriate treatment plan and managing the lives of patients [2].
In this study, our primary objectives were twofold. Firstly, we aimed to enhance the accuracy of predicting survival duration for patients diagnosed with breast cancer and those who have experienced heart attacks. This would be achieved through the integration of Stacked AutoEncoders, the Particle Swarm Optimization algorithm, and the Softmax Classifier. Stacked AutoEncoders are one of the most popular models in the field of Deep Learning. The Particle Swarm Optimization algorithm is frequently used in the literature and is an effective optimization algorithm in solving multidimensional problems. The Softmax Classifier is a classification method that attracts attention with its high classification performance. By harnessing the strengths of these methodologies, our goals was to provide more precise prognostications, ultimately improving patient care and treatment outcomes. Secondly, we endeavored to develop a methodology capable of generating diverse and effective architectures tailored to specific problem domains. This would involve crafting adaptable frameworks that could accommodate various datasets and challenges inherent to different medical scenarios. By designing such a methodology, we aimed to contribute to the broader field of Machine Learning by providing a flexible and scalable approach to solving complex problems in healthcare and beyond.
In alignment with these aims and objectives, a Deep Learning Network was constructed, leveraging an optimization approach that integrated Stacked AutoEncoders, the Particle Swarm Optimization algorithm, and the Softmax Classifier. This amalgamation aimed to enhance the network’s performance in predicting survival duration for patients diagnosed with breast cancer and those affected by heart attacks. To tailor the network to the specific tasks at hand, both the architectural parameters and the hyperparameters were meticulously optimized using the Particle Swarm Optimization algorithm. This optimization process aimed to fine-tune the network’s configuration for optimal performance, ensuring its efficacy in addressing the targeted medical prognostication tasks. The developed architecture was then put to the test using real-world datasets obtained from the Haberman’s Survival and Echocardiogram datasets sourced from the UCI repository. Additionally, various conventional Machine Learning methods were employed on these datasets to provide a basis for comparative analysis. The results derived from the hybrid architecture were juxtaposed with outcomes from alternative Machine Learning methods and existing studies within the literature that utilized identical datasets. This comparison was conducted to provide a comprehensive evaluation, and the findings are presented in a comparative manner.
Innovations derived from the experimental studies: The proposed hybrid architecture demonstrated superior success rates compared to the Machine Learning methods employed in the research. Furthermore, it achieved notably higher success rates on the Haberman’s Survival dataset compared to those reported in the existing literature. This study introduces a novel hybrid Machine Learning approach that integrates Stacked AutoEncoders, the Softmax Classifier, and Particle Swarm Optimization. Additionally, an alternative decision support system is proposed for predicting the survival outcomes of patients following breast cancer surgery and heart attack incidents.

2. Literature Review

There are many studies in the literature for patient survival predictions. For example, Li and Wang [3] performed survival probability estimation with the semi-parametric model averaging method. In the study, a Brier Score-type criterion was used to select the most appropriate model, which was the average of the weights. In the study, successful results were obtained in predicting the survival probabilities of patients with heart failure.
Evangeline et al. [4] used three different models—namely, the Cox Proportional Hazards (CoxPH) model, the Random Survival Forests (RSF) model, and DeepHit—in their study for the survival prediction of breast cancer patients. As a result of these experimental studies, successful results were obtained in the RFS and DeepHit models.
Hussain et al. [5] proposed a model to predict the survival time of patients with brain tumors. This model was the first model to consider the volume of brain tumor in survival time calculation. In the study, where 4D-MRI, 3D CNN, and CoxPH models were used, it was emphasized that more successful results were obtained than the leading survival prediction models.
Huang et al. [6] used Machine Learning and Deep Learning techniques to predict the survival times of patients with lung cancer. Using 12 demographic and clinical characteristics, the study applied seven Machine Learning models, including Logistic Regression (LR), Bayesian Classifier (BayesNet), Lazy Classifier (LWL), Meta-Classifier (AttributeSelectedClassifier (ASC)), Rule Learner (OneR), Decision Tree (J48), and Deep Neural Network (DNN). Among the models used, DNN was more successful than the other methods.
Atallah et al. [7] proposed a hybrid model to predict survival in the next five years after kidney transplantation. In the hybridization of the pure Bayes and K-Nearest Neighbor Machine Learning methods, successful results were obtained in predicting survival after kidney transplantation.
Azam et al. [8] presented a novel approach aimed at enhancing accuracy in classification processes. This approach focused on generating Type-1 Fuzzy Triangular and Trapezoidal membership functions. The study employed the generation of c-means clustering, membership matrix, and cluster centers to produce Type-1 Fuzzy membership functions. It has been applied to various datasets, including the Haberman’s Survival dataset. The findings from the experimental study indicate the success of the approach.
Kurama [9] proposed a novel similarity-based classifier integrated with aggregator operators. Four distinct Dombi aggregation operators—conjunctive, discrete, weighted conjunctive, and product operators—were employed to aggregate similarities within the classifier. Utilizing these proposed methods, the accuracy rate for predicting patient survival after surgery increased by 0.29 compared to the existing literature.
Bataineh et al. [10] utilized the Clonal Selection Algorithm for multi-layer sensor training in their work. The study also employed Genetic Algorithms, Ant Colony Optimization, Particle Swarm Optimization, Harris Hawks Optimization, Moth Flame Optimization, Flower Pollination, and Backpropagation methods to compare the performance of the proposed method. Successful results were achieved with the proposed methods in the study, where all the methods were used to estimate patient survival after surgery.
Kaushik et al. [11] have proposed a new approach for predicting the survival of breast cancer patients. The new approach, recommended for lumpectomy and mastectomy, utilizes Support Vector Machine Communication-Efficient Distributed Double Coordinate Ascent (SVM-CoCoA). The originality of the new model was observed through an experiment conducted using the Simple Compilation Tool (SBT) and Apache Flink. Experimental studies were carried out using the Haberman’s Survival dataset. The findings revealed the success of the proposed method.
Aljawad et al. [12] used Bayes Network and Support Vector Machines to estimate the survival status of patients undergoing breast cancer surgery. The results from experimental studies conducted on the Weka software platform showed that the Support Vector Machines method produced more successful results than the Bayes Network.
Remya Ajai et al. [13] conducted an analysis of logistically mapped neurons in neuroacoustic learning architectures for data classification. The study proposed two different extensions of Neurochaos Learning, a chaos-based learning algorithm for classification. To further enhance the classification performance of 1D logistic map neurons and the Heterogeneous Neuroacoustic Learning Architecture, Chaos-based attributes were extracted from the Extruded and Support Vector Machines instead of the GLS Neurons. Successful results were obtained in the approaches used to predict the survival of patients undergoing breast cancer surgery.
The utilization of Machine Learning (ML) methods in health decision support systems is witnessing a steady rise. This trend holds significant importance for healthcare professionals in discerning the most appropriate treatment methods and assessing the efficacy of applied treatments. ML methods employed in the healthcare domain encompass numerous parameters, making it exceedingly challenging to fine-tune these parameters specifically for a given problem through trial-and-error methods alone. In this context, integrating various ML methods and optimizing the hyperparameters within established architectures play a pivotal role in achieving successful outcomes. In light of all this, in this study, a hybrid model is proposed that creates the most suitable model for the problems by optimizing its own architectural structure and the hyperparameters within the architecture.

3. Materials and Methods

In this section, the Particle Swarm Optimization Algorithm used in the study, Stacked AutoEncoders, the Softmax Classifier, the other ML methods used, and the metrics used to evaluate the performance of the proposed hybrid architecture will be explained in detail.

3.1. Particle Swarm Optimization Algorithm

The Particle Swarm Optimization (PSO) algorithm is a widely used optimization technique grounded in swarm intelligence principles [14]. It is widely applied in the literature for addressing various optimization problems. PSO operates by facilitating social information sharing within the swarm, enabling individuals’ positions to converge towards the best position within the swarm [15]. Figure 1 illustrates the flow diagram of the PSO algorithm:
In the flow diagram illustrated in Figure 1, the algorithm commences by randomly dispersing particles, each representing a potential solution for the problem, across the solution space. During this step, other parameters within the PSO are also established. Subsequently, the fitness value of all the particles in the swarm is computed, indicating the quality of solutions contained within each particle. Following this, the particle with the best position in the current iteration ( p b e s t ) is identified based on its fitness value. Then, the global best ( g b e s t ) is determined among the particles possessing the best fitness value in the current iteration. In the subsequent step, the velocities of the particles and their corresponding positions are updated based on this new velocity value, as outlined in Equations (1) and (2):
V ( i + 1 ) j = W V i j + c 1 r a n d 1 [ p b e s t i X i j ] + c 2 r a n d 2 [ g b e s t i X i j ]
X i j ( t + 1 ) = X i j ( t ) + V i j ( t + 1 )
where V i j represents the velocity of particle j at iteration i; r a n d 1 and r a n d 2 represent random values generated in the range [0–1]; c 1 represents the cognitive learning ability within the swarm, while c 2 represents the social learning ability within the swarm; X i j represents the position of particle j in iteration i; p b e s t i represents the best particle at iteration i; and g b e s t i represents the global best particle [16].
The algorithm iterates until the stopping criterion is satisfied. Upon meeting the stopping criterion, the p b e s t and g b e s t values are updated, and the best particle suitable for the problem is determined.

3.2. Stacked AutoEncoders

Stacked AutoEncoders (SAE) is a type of stacked unsupervised Neural Network comprising AutoEncoder (AE) Neural Networks. An AE Neural Network consists of two main components: an encoder and a decoder. The encoder is responsible for transforming input data into a compressed representation, known as f e a t u r e s , while the decoder reconstructs the original input data from these features.
Figure 2 illustrates the basic structure of an AE, which consists of three layers. In an AE, the output of each hidden layer serves as the input data for the subsequent hidden layer. An AE comprises two main components: the encoder and the decoder. AEs facilitate data size reduction by extracting features through the encoder component.
Multiple AutoEncoders (AEs) are combined to form a Stacked AutoEncoders (SAE) Neural Network. Figure 3 illustrates the structure of the SAE.

3.3. The Softmax Classifier

The Softmax Classifier is a Machine Learning method commonly employed in multi-layer Artificial Neural Network (ANN) architectures, particularly in classification tasks [17,18]. The mathematical representation of the softmax function is depicted in Equation (3) [19]:
σ ( z ) j = e z j k = 1 K e z k
Here, σ ( z ) j represents the softmax output calculated for class j; e represents the Euler number; z j represents the element of the input vector with index j; K represents the number of classes.

3.4. Evaluation Metrics

The Accuracy, Sensitivity, Precision, and F-score metrics used in the literature to measure the performance of classification problems were applied in estimating postoperative patient survival [20,21]. In order to calculate these performance measurement metrics, the confusion matrix of the classification process had first be determined [22,23]. A confusion matrix is a tabular layout that allows visualizing the performance of an algorithm [24]. An example confusion matrix is shown in Figure 4:
TP in Figure 4 represents the True Positive value. TN represents True Negative. FP represents False Positive. FN represents False Negative. The mathematical formula of Accuracy, Sensitivity, Precision, and F-score metrics over the complexity matrix is shown in Equations (4)–(7) [25]:
A c c u r a c y ( A ) = T P + T N T P + T N + F N + F P
S e n s i t i v i t y ( S ) = T P F N + T P
P r e c i s i o n ( P ) = T P T P + F P
F - S c o r e = 2 × P × S P + S
Sensitivity indicates the ratio of positively classified data to the actual positive data. Precision represents the ratio of correctly classified data. F-Score provides the harmonic mean of Precision and Sensitivity values, providing a complete picture of accuracy. The Accuracy value represents the accuracy itself.

3.5. Other Machine Learning Methods Used in the Study

Apart from the proposed architecture, the other Machine Learning methods applied on the same dataset are presented in Table 1.
The Machine Learning methods presented in Table 1 are used in many classification problems [32,33,34].

4. Datasets

In this section, detailed information about the various datasets utilized in the study will be provided.

4.1. Echocardiogram Dataset

One of the datasets used in the study is the Echocardiogram dataset. This dataset was obtained from UCI, which is publicly available [35]. The primary challenge addressed by the researchers is to predict whether a patient will survive for at least one year based on other variables. Nine meaningful features were considered in the dataset. The features included in the dataset are shown in Table 2.
There are a total of 132 data in the dataset whose characteristics are given in Table 2. Since no information could be obtained about 1 tansi among these data, this datum was excluded from the dataset. Experimental studies were conducted on these 131 data and the dataset was divided into two parts, 70% training and 30% testing. The missing information in the dataset made the predictability of the dataset difficult. The correlation of the dataset features with each other is shown in Figure 5:
In Figure 5, the correlation between each feature in the dataset and other features is depicted. For instance, a positive correlation is observed between the Age-at-heart-attack feature (var1) and the Pericardial-effusion (var2) feature, while a negative correlation exists between the Pericardial-effusion (var2) feature and Epss (var4).

4.2. Haberman’s Survival Dataset

In the study, the Haberman’s Survival dataset available at UCI was used for postoperative patient survival prediction [36]. The dataset characteristics are shown in Table 3.
In the dataset, whose characteristics are presented in Table 1, there are a total of 306 data, 225 of which are indicated as ’patient survived for 5 years or more’ and 81 of which are indicated as ’patient died within 5 years’. The dataset was divided into two different groups for the training and testing phases: 70% of the dataset was used for training and 30% for testing.
The correlation of the dataset features with each other is shown in Figure 6:
In Figure 6, the correlation between each feature in the dataset and other features is illustrated. For example, a positive correlation is observed between the ’Age of the patient at the time of operation’ feature (var1) and ’Patient’s year of operation’ (var2), while a negative correlation is evident between the ’Number of positive axillary nodes detected’ (var3) feature and ’Survival status’ (var4).

5. Proposed Model

An optimization-based hybrid architecture for predicting survival on datasets was created. The architecture is shown in Figure 7.
In the architecture presented in Figure 7, a Softmax classifier is connected to the Stacked AutoEncoders output for survival time estimation. Considering the resulting architecture and the layers used in the architecture, there were too many hyperparameters. An optimization algorithm was used to determine these hyperparameters in the hybrid architecture created with PSO–SAE–Softmax Classifier and the parameters of the architecture to produce the best result. PSO was used in this study, due to its minimum parameter requirement, success in solving multidimensional problems, and ease of implementation. The optimization-based hybrid architecture was implemented on the MATLAB platform for coding flexibility and ease of working with multidimensional matrices [37]. The figures and graphs in the following sections of the study were plotted in the MATLAB software (MATLAB R2021a) environment. The parameters to be optimized in the architecture are shown in Table 4.
The hyperparameters in Table 4 directly affect the success performance of the network. In such a problem, with a nine-dimensional solution space, the most successful architecture and hyperparameter values can only be determined by optimization algorithms [16,38]. One of the most significant features of the hybrid model is that each particle within the model possesses a unique architecture and distinct hyperparameters from the others. This hybrid model not only optimizes the architecture’s structure but also fine-tunes the hyperparameters specific to each particle’s architecture, thereby aiming to achieve optimal results.
As a first step in the proposed Deep Learning architecture, the PSO parameters that will optimize the structure and hyperparameters of the architecture were determined, as shown in Table 5, as a result of the experimental studies.
Given the expansive solution space and the expectation for the optimization algorithm to fine-tune both hyperparameters and architecture, a substantial number of iterations was maintained. Each member of the population, as outlined in Table 5, encompassed all the hyperparameters detailed in Table 4.
In the subsequent step, a number of particles equal to the population specified in Table 5 was randomly distributed within the solution space. The N S A E value contained within each particle, as outlined in Table 4, was randomly assigned in accordance with the constraint function delineated in Equation (8):
P i N S A E j ( n ) = 1 n < 1 n 1 n 6 6 n > 6
where P i N S A E j ( n ) represents the N S A E value of particle i at iteration j.
The N H E D values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (9):
P i N H E D j ( n ) = 10 n < 10 n 10 n 250 250 n > 250
where P i N H E D j ( n ) represents the N H E D value of particle i at iteration j.
The N E A F values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (10):
P i N E A F j ( n ) = 1 n < 1 n 1 n 2 2 n > 2
where P i N E A F j ( n ) represents the N E A F value of particle i at iteration j. The activation functions corresponding to the randomly determined N E A F value are listed in Table 6.
The N D A F values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (11):
P i N D A F j ( n ) = 1 n < 1 n 1 n 3 3 n > 3
where P i N D A F j ( n ) represents the N D A F value of particle i at iteration j. The activation functions corresponding to the randomly determined N D A F value are listed in Table 7.
The A E L 2 values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (12):
P i A E L 2 j ( n ) = 0.001 n < 0.001 n 0.001 n 0.01 0.01 n > 0.01
where P i A E L 2 j ( n ) represents the A E L 2 value of particle i at iteration j.
The A E S R values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (13):
P i A E S R j ( n ) = 1 n < 1 n 1 n 5 5 n > 5
where P i A E S R j ( n ) represents the A E S R value of particle i at iteration j.
The A E S P values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (14):
P i A E S P j ( n ) = 0 n < 0 n 0 n 1 1 n > 1
where P i A E S P j ( n ) represents the A E S P value of particle i at iteration j.
The A E L D values depicted in Table 4 were randomly assigned based on the constraint function described in Equation (15):
P i A E L D j ( n ) = 1 n < 1 n 1 n 2 2 n > 2
where P i A E L D j ( n ) represents the A E L D value of particle i at iteration j. The values corresponding to the randomly determined A E L D value are shown in Table 8.
After randomly distributing the particles across the solution space, the fitness values of each particle were computed according to Equation (16):
F P i j = T e s t ( A c c u r a c y ( S A E S o f t ) i j )
where F P i j denotes the fitness value of particle i at iteration j, and where ( S A E S o f t ) i j represents the architecture of particle i at iteration j. Equation (16) governed the calculation of the fitness value for each particle. Based on the computed fitness value, p b e s t (particle best) and g b e s t (global best) were determined.
In the subsequent step of the architecture, the velocity values of each particle were updated following Equation (1). In this process, the new velocities of each particle were determined based on the constraint functions utilized in the population formation. After adjusting the velocities, the position of each particle was updated according to the mathematical expression provided in Equation (2).
The hybrid architecture was separately executed for both datasets, following the number of iterations outlined in Table 4. In the final step, the fitness values of each particle within the hybrid architecture were computed separately for both datasets, using Equation (16), and the global best values were determined based on the experimental studies conducted with both datasets. The architectural parameters and hyperparameters associated with the particle exhibiting the best g b e s t value from the experimental studies for the Haberman’s Survival dataset are presented in Table 9.
The hybrid architecture obtained with the parameters shown in Table 9 as a result of the experimental studies for Haberman’s Survival dataset is shown in Figure 8.
As depicted in Figure 8, the input values traversed through a 98-layer Encoder section before being connected to the Softmax Classifier. Here, they underwent the classification process, resulting in the production of output.
The architectural parameters and hyperparameters associated with the particle exhibiting the best g b e s t value from the experimental studies for the Echocardiogram dataset are presented in Table 10.
The hybrid architecture obtained with the parameters shown in Table 10 as a result of the experimental studies for the Echocardiogram dataset is shown in Figure 9.
As depicted in Figure 9, the input values traversed through a 159-layer Encoder section before being connected to the Softmax Classifier. Here, they underwent the classification process, resulting in the production of output.
Acknowledging the potential for the PSO to become trapped in local optima, the architecture depicted in Figure 7 was executed 25 times for each dataset. Models were generated using the optimal parameters obtained from these runs. The performance of the architecture across each experimental run for the datasets is depicted in Figure 10.
The results from the experimental investigations illustrated in Figure 10 were analyzed by assessing the model’s minimum misclassification performance in each experimental trial.
Furthermore, the computational complexity of both the hybrid architecture proposed in this study and the other Machine Learning methods utilized are displayed in Table 11.
Upon reviewing Table 11, it becomes evident that the proposed method exhibits remarkable complexity performance.

6. The Experimental Results and Discussion

6.1. Survival Prediction Model after Heart Attack

In this study, in addition to the architecture outlined for the survival datasets, commonly utilized methods such as Decision Trees (DT), K-Nearest Neighbors (KNN), Support Vector Machines (SVMs), Neural Networks (NN), Gradient Boosting, and Gradient Bagging were also employed. All methodologies utilized in this study underwent training with allocated training datasets and subsequent testing with independent test datasets. The experimental findings, conducted with the Chocardiogram dataset, are presented through the confusion matrices of the models evaluated with the test dataset, as illustrated in Figure 11.
When examining the confusion matrices obtained from the methods illustrated in Figure 11, it becomes apparent that KNN, exhibiting the highest number of incorrect groupings among its predictions, is the least effective predictor of survival times after a heart attack. Conversely, the proposed hybrid model emerges as the most accurate estimator. Utilizing the confusion matrices provided in Figure 11 and the evaluation metrics summarized in Equations (4) to (7), the performance percentages of the models are calculated and presented in Table 12.
Based on the metric values summarized in Table 12, it is evident that the proposed hybrid architecture attained the highest performance metrics and accuracy rate. Upon evaluating these results, it becomes apparent that the hybrid model, optimized with the optimization-based hybrid architecture, outperforms other Machine Learning methods.

6.2. Survival Prediction Model after Breast Cancer Surgery

The confusion matrices resulting from experimental studies on the test dataset for survival prediction after breast cancer surgery, comparing the hybrid architecture with other methods, are depicted in Figure 12.
When the confusion matrices obtained from the methods shown in Figure 12 are analyzed, it is seen that Decision Trees (DT), which has the highest number of incorrect groupings among its predictions, is the weakest predictor of survival times after breast cancer surgery, while the proposed hybrid model emerges as the most accurate predictor. Based on the confusion matrices provided in Figure 12 and the evaluation metrics outlined in Equations (4)–(7), the performances of the models are computed and presented in Table 13.
Based on the metric values outlined in Table 13, although the SVM exhibits the highest precision value, the proposed hybrid architecture achieved the highest accuracy rate. Evaluating these results, it becomes evident that the hybrid model optimized with the optimization-based hybrid architecture attains more successful outcomes compared to other Machine Learning methods.
The comparative results of the outcomes obtained using the proposed optimization-based hybrid architecture with those from studies conducted in the literature employing the same dataset are presented in Table 14.
As depicted in Table 14, the proposed hybrid Deep Learning architecture achieved superior results compared to the other studies in the literature employing the same dataset.

7. Conclusions

Survival prediction models that achieve high accuracy play a critical role in patient treatments. By more accurately estimating patients’ survival probabilities, these models help optimize treatment processes, offering more effective and personalized treatment strategies to patients, thereby improving health outcomes. Additionally, they provide healthcare professionals with the opportunity to manage patient care better by distributing resources more efficiently and providing early warnings to identify high-risk situations. Therefore, survival prediction models that achieve high accuracy are essential tools in clinical decision-making processes, enhancing the quality of healthcare services. In this study, we present an optimization-based hybrid method that integrates SAE, PSO, and the Softmax Classifier to predict the survival of patients undergoing breast cancer surgery and those experiencing a heart attack. The architecture derived from our proposed hybrid method was applied to the Haberman’s Survival and Echocardiogram datasets from UCI. In addition to our proposed architecture, various methods, including DT, SVMs, KNN, ANN, Gradient Boosting, and Gradient Bagging were also applied on the same datasets. Our proposed architecture achieved 72% and 89% accuracy rates on the datasets used in the study, respectively, and provided superior results compared to the other methods. Furthermore, the outcomes obtained with our proposed architecture outperformed previous studies in the literature on the Haberman’s Survival dataset. With the proposed optimization-based hybrid architecture:
  • Various Deep Learning architectures based on optimization have been developed to predict survival after breast cancer surgery and the survival of patients following a heart attack.
  • Utilizing Particle Swarm Optimization (PSO), both architecture parameters and hyperparameters within the architecture are optimized.
  • Through leveraging features from the Haberman’s Survival dataset, a model exhibiting superior performance compared to the existing literature studies has been achieved.
  • The optimized architecture, facilitated by the developed Deep Learning framework, offers access to both features and hyperparameters.
  • The hybrid Deep Learning Network created is not only tailored to this particular problem but can also be seamlessly applied to various other problems.
  • The architecture obtained through the hybrid Deep Learning Network stands as a viable alternative method for integration into health decision support systems.
In future studies, the findings can pave the way for developing applications tailored for smart devices.

Author Contributions

Conceptualization, M.A.B. and M.F.I.; methodology, M.A.B. and M.F.I.; software, M.A.B.; validation, M.A.B. and M.F.I.; formal analysis, M.A.B. and M.F.I.; investigation, M.A.B. and M.F.I.; resources, M.A.B. and M.F.I.; data curation, M.A.B.; writing—original draft preparation, M.A.B. and M.F.I.; writing—review and editing, M.A.B. and M.F.I.; visualization, M.A.B. and M.F.I.; supervision, M.A.B.; project administration, M.A.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study used the Haberman’s Survival dataset and the Echocardiogram dataset, both of which are publicly available at the UC Irvine Machine Learning Repository. (https://archive.ics.uci.edu/dataset/43/haberman+s+survival, https://archive.ics.uci.edu/dataset/38/echocardiogram), accessed on 20 April 2024.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vinod, D.; Prakash, S.S.; AlSalman, H.; Muaad, A.Y.; Heyat, M.B.B. Ensemble Technique for Brain Tumour Patient Survival Prediction. IEEE Access 2024, 12, 19285–19298. [Google Scholar] [CrossRef]
  2. Joffe, E.; Coombes, K.R.; Qiu, Y.H.; Yoo, S.Y.; Zhang, N.; Bernstam, E.V.; Kornblau, S.M. Survival prediction in high dimensional datasets–Comparative evaluation of lasso regularization and random survival forests. Blood 2013, 122, 1728. [Google Scholar] [CrossRef]
  3. Li, M.; Wang, X. Semiparametric model averaging method for survival probability predictions of patients. Comput. Stat. Data Anal. 2023, 185, 107759. [Google Scholar] [CrossRef]
  4. Evangeline, I.K.; Kirubha, S.A.; Precious, J.G. Survival analysis of breast cancer patients using machine learning models. Multimed. Tools Appl. 2023, 82, 30909–30928. [Google Scholar] [CrossRef]
  5. Hussain, S.; Haider, S.; Maqsood, S.; Damaševičius, R.; Maskeliūnas, R.; Khan, M. ETISTP: An enhanced model for brain tumor identification and survival time prediction. Diagnostics 2023, 13, 1456. [Google Scholar] [CrossRef] [PubMed]
  6. Huang, S.; Arpaci, I.; Al-Emran, M.; Kılıçarslan, S.; Al-Sharafi, M.A. A comparative analysis of classical machine learning and deep learning techniques for predicting lung cancer survivability. Multimed. Tools Appl. 2023, 82, 34183–34198. [Google Scholar] [CrossRef]
  7. Atallah, D.M.; Badawy, M.; El-Sayed, A.; Ghoneim, M.A. Predicting kidney transplantation outcome based on hybrid feature selection and KNN classifier. Multimed. Tools Appl. 2019, 78, 20383–20407. [Google Scholar] [CrossRef]
  8. Azam, M.H.; Hasan, M.H.; Hassan, S.; Abdulkadir, S.J. A novel approach to generate type-1 fuzzy triangular and trapezoidal membership functions to improve the classification accuracy. Symmetry 2021, 13, 1932. [Google Scholar] [CrossRef]
  9. Kurama, O. A new similarity-based classifier with Dombi aggregative operators. Pattern Recognit. Lett. 2021, 151, 229–235. [Google Scholar] [CrossRef]
  10. Al Bataineh, A.; Kaur, D.; Jalali, S.M.J. Multi-layer perceptron training optimization using nature inspired computing. IEEE Access 2022, 10, 36963–36977. [Google Scholar] [CrossRef]
  11. Kaushik, D.; Prasad, B.R.; Sonbhadra, S.K.; Agarwal, S. Post-surgical survival forecasting of breast cancer patient: A novel approach. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 37–41. [Google Scholar]
  12. Aljawad, D.A.; Alqahtani, E.; Ghaidaa, A.K.; Qamhan, N.; Alghamdi, N.; Alrashed, S.; Alhiyafi, J.; Olatunji, S.O. Breast cancer surgery survivability prediction using bayesian network and support vector machines. In Proceedings of the 2017 International Conference on Informatics, Health & Technology (ICIHT), Riyadh, Saudi Arabia, 21–23 February 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  13. AS, R.A.; Harikrishnan, N.; Nagaraj, N. Analysis of logistic map based neurons in neurochaos learning architectures for data classification. Chaos Solitons Fractals 2023, 170, 113347. [Google Scholar]
  14. Işık, M.F.; Avcil, F.; Harirchian, E.; Bülbül, M.A.; Hadzima-Nyarko, M.; Işık, E.; İzol, R.; Radu, D. A Hybrid Artificial Neural Network—Particle Swarm Optimization Algorithm Model for the Determination of Target Displacements in Mid-Rise Regular Reinforced-Concrete Buildings. Sustainability 2023, 15, 9715. [Google Scholar] [CrossRef]
  15. Farhat, Y.; Atig, A.; Zribi, A.; Ben Abdennour, R. Neural emulator for nonlinear systems based on PSO algorithm: Real-time validation. Clust. Comput. 2024, 1–13. [Google Scholar] [CrossRef]
  16. Bülbül, M.A.; Öztürk, C.; Işık, M.F. Optimization of climatic conditions affecting determination of the amount of water needed by plants in relation to their life cycle with particle swarm optimization, and determining the optimum irrigation schedule. Comput. J. 2022, 65, 2654–2663. [Google Scholar] [CrossRef]
  17. Zahra, A.; Amin, M.; El-Samie, F.E.A.; Emam, M. Efficient utilization of deep learning for the detection of fabric defects. Neural Comput. Appl. 2024, 36, 6037–6050. [Google Scholar] [CrossRef]
  18. Gasmi, K.; Kharrat, A.; Ammar, L.B.; Ltaifa, I.B.; Krichen, M.; Mrabet, M.; Alshammari, H.; Yahyaoui, S.; Khaldi, K.; Hrizi, O. Classification of MRI brain tumors based on registration preprocessing and deep belief networks. AIMS Math. 2024, 9, 4604–4631. [Google Scholar] [CrossRef]
  19. Fan, B.; Ma, H.; Liu, Y.; Yuan, X. BWLM: A Balanced Weight Learning Mechanism for Long-Tailed Image Recognition. Appl. Sci. 2024, 14, 454. [Google Scholar] [CrossRef]
  20. Bülbül, M.A. Optimization of Artificial Neural Network structure and hyperparameters in hybrid model by genetic algorithm: iOS–android application for breast cancer diagnosis/prediction. J. Supercomput. 2024, 80, 4533–4553. [Google Scholar] [CrossRef]
  21. Xu, Z.; Selvaraj, V.; Min, S. Intelligent G-code-based power prediction of ultra-precision CNC machine tools through 1DCNN-LSTM-Attention model. J. Intell. Manuf. 2024, 1–24. [Google Scholar] [CrossRef]
  22. Praveena, S.; Pavithra, S.; Kumar, A.D.V.; Veeresha, P. CNN-based Indian medicinal leaf type identification and medical use recommendation. Neural Comput. Appl. 2024, 36, 5399–5412. [Google Scholar] [CrossRef]
  23. Bülbül, M.A. Kuru fasulye tohumlarının çok sınıflı sınıflandırılması için hibrit bir yaklaşım. J. Inst. Sci. Technol. 2023, 13, 33–43. [Google Scholar]
  24. Carlson, K.; Buttenfield, B.P.; Qiang, Y. Wetland Classification, Attribute Accuracy, and Scale. ISPRS Int. J. Geo-Inf. 2024, 13, 103. [Google Scholar] [CrossRef]
  25. Rajalakshmi, A.; Sridhar, S. Classification of yoga, meditation, combined yoga–meditation EEG signals using L-SVM, KNN, and MLP classifiers. Soft Comput. 2024, 28, 4607–4619. [Google Scholar] [CrossRef]
  26. Lin, B.; Liu, C.; Miao, D. An improved decision tree algorithm based on boundary mixed attribute dependency. Appl. Intell. 2024, 54, 2136–2153. [Google Scholar] [CrossRef]
  27. Emimal, M.; Hans, W.J.; Inbamalar, T.; Lindsay, N.M. Classification of EMG signals with CNN features and voting ensemble classifier. Comput. Methods Biomech. Biomed. Eng. 2024. online ahead of print. [Google Scholar] [CrossRef] [PubMed]
  28. Xiao, Y.; Pan, G.; Liu, B.; Zhao, L.; Kong, X.; Hao, Z. Privileged multi-view one-class support vector machine. Neurocomputing 2024, 572, 127186. [Google Scholar] [CrossRef]
  29. Yang, J.; Zhao, J. A novel parallel merge neural network with streams of spiking neural network and artificial neural network. Inf. Sci. 2023, 642, 119034. [Google Scholar] [CrossRef]
  30. Gunasekara, N.; Pfahringer, B.; Gomes, H.; Bifet, A. Gradient boosted trees for evolving data streams. Mach. Learn. 2024, 113, 3325–3352. [Google Scholar] [CrossRef]
  31. Aldayel, M.; Al-Nafjan, A. A comprehensive exploration of machine learning techniques for EEG-based anxiety detection. PeerJ Comput. Sci. 2024, 10, e1829. [Google Scholar] [CrossRef]
  32. Mang, L.D.; González Martínez, F.D.; Martinez Muñoz, D.; García Galán, S.; Cortina, R. Classification of Adventitious Sounds Combining Cochleogram and Vision Transformers. Sensors 2024, 24, 682. [Google Scholar] [CrossRef]
  33. Glenis, A.; Vouros, G.A. SCALE-BOSS-MR: Scalable Time Series Classification Using Multiple Symbolic Representations. Appl. Sci. 2024, 14, 689. [Google Scholar] [CrossRef]
  34. Bülbül, M.A. Urinary Bladder Inflammation Prediction with the Gray Wolf Optimization Algorithm and Multi-Layer Perceptron-Based Hybrid Architecture. Bitlis Eren Üniv. Fen Bilim. Derg. 2023, 12, 1185–1194. [Google Scholar] [CrossRef]
  35. UC Irvine Machine Learning Repository. 2024. Available online: https://archive.ics.uci.edu/dataset/38/echocardiogram (accessed on 20 April 2024).
  36. UCI Machine Learning Repository. 2024. Available online: https://archive.ics.uci.edu/dataset/43/haberman+s+survival (accessed on 20 April 2024).
  37. Konak, F.; Bülbül, M.A.; Türkoǧlu, D. Feature Selection and Hyperparameters Optimization Employing a Hybrid Model Based on Genetic Algorithm and Artificial Neural Network: Forecasting Dividend Payout Ratio. Comput. Econ. 2024, 63, 1673–1693. [Google Scholar] [CrossRef]
  38. Bülbül, M.A.; Harirchian, E.; Işık, M.F.; Aghakouchaki Hosseini, S.E.; Işık, E. A hybrid ANN-GA model for an automated rapid vulnerability assessment of existing RC buildings. Appl. Sci. 2022, 12, 5138. [Google Scholar] [CrossRef]
  39. Sethi, D.; Nagaraj, N.; Harikrishnan, N.B. Neurochaos feature transformation for Machine Learning. Integration 2023, 90, 157–162. [Google Scholar] [CrossRef]
  40. Melin, P.; Sánchez, D.; Castillo, O. Interval Type-3 Fuzzy Inference System Design for Medical Classification Using Genetic Algorithms. Axioms 2023, 13, 5. [Google Scholar] [CrossRef]
Figure 1. PSO flow diagram.
Figure 1. PSO flow diagram.
Biomimetics 09 00304 g001
Figure 2. AutoEncoder structure.
Figure 2. AutoEncoder structure.
Biomimetics 09 00304 g002
Figure 3. Stacked AutoEncoders structure.
Figure 3. Stacked AutoEncoders structure.
Biomimetics 09 00304 g003
Figure 4. Confusion matrix for a two-class classifier.
Figure 4. Confusion matrix for a two-class classifier.
Biomimetics 09 00304 g004
Figure 5. Echocardiogram dataset correlation chart.
Figure 5. Echocardiogram dataset correlation chart.
Biomimetics 09 00304 g005
Figure 6. Haberman’s Survival dataset correlation chart.
Figure 6. Haberman’s Survival dataset correlation chart.
Biomimetics 09 00304 g006
Figure 7. The proposed SAE–Softmax Classifier–PSO hybrid architecture.
Figure 7. The proposed SAE–Softmax Classifier–PSO hybrid architecture.
Biomimetics 09 00304 g007
Figure 8. Survival prediction model after breast cancer surgery.
Figure 8. Survival prediction model after breast cancer surgery.
Biomimetics 09 00304 g008
Figure 9. Survival prediction model after heart attack.
Figure 9. Survival prediction model after heart attack.
Biomimetics 09 00304 g009
Figure 10. Hybrid architecture performance in experimental studies.
Figure 10. Hybrid architecture performance in experimental studies.
Biomimetics 09 00304 g010
Figure 11. Confusion matrices of models for predicting survival after heart attack.
Figure 11. Confusion matrices of models for predicting survival after heart attack.
Biomimetics 09 00304 g011
Figure 12. Confusion matrices of models for predicting survival after breast cancer surgery.
Figure 12. Confusion matrices of models for predicting survival after breast cancer surgery.
Biomimetics 09 00304 g012
Table 1. Machine Learning methods used in the study.
Table 1. Machine Learning methods used in the study.
MethodsInformation
Decision Trees (DT)It is a tree-like model that obtains results by
dividing and classifying the dataset with
successive decisions [26].
K-Nearest Neighbors (KNN)It is a method that predicts the class or value
of a data point based on the class of the
K data points closest to it [27].
Support Vector Machines (SVMs)It is a method that aims to find an optimal
hyperplane to best classify or predict data
points [28].
Artificial Neural Networks (ANNs)Artificial Neural Networks are a mathematical
representation of nerve cells found in humans [29].
Gradient BoostingGradient Boosting is a widely used Machine Learning
technique that has proven highly effective in batch
learning [30].
Gradient BaggingGradient Bagging is a Machine Learning method
that shows high performance in classification problems [31].
Table 2. Echocardiogram dataset properties and descriptions.
Table 2. Echocardiogram dataset properties and descriptions.
FeaturesVariable InformationData Type
Age-at-heart-attackNumericalInput
Pericardial-effusionBinaryInput
Fractional-shorteningNumericalInput
EpssNumericalInput
LvddNumericalInput
Wall-motion-scoreNumericalInput
Wall-motion-indexNumericalInput
MultiNumericalInput
Still-aliveClass attribute
0—the patient is dead at end of survival period
1—the patient is still alive
Output
Table 3. Dataset properties and descriptions.
Table 3. Dataset properties and descriptions.
FeaturesVariable InformationData Type
Age of patient at time of operationNumericalInput
Patient’s year of operationNumericalInput
Number of positive axillary nodes detectedNumericalInput
Survival statusClass attribute
0—the patient survived 5 years or longer
1—the patient died within 5 years Output
Output
Table 4. Parameters used in the architecture and to be optimized.
Table 4. Parameters used in the architecture and to be optimized.
Parameters to OptimizeParameter FeatureRepresentation in
Architecture
Number of AutoEncoders in the
Stacked AutoEncoder architecture
Hyperparameter in Hybrid Architecture structureNSAE
Number of hidden layers in the
encoder layer in each AutoEncoder in
the architecture
Hyperparameter in the Stacked AutoEncoders structureNHED
Number of hidden layers in the decoder
layer in each AutoEncoder in the architecture
Hyperparameter in the Stacked AutoEncoders structureNHED
Activation functions used in the hidden
layers in the encoder section of each
AutoEncoder structure
Hyperparameter in the Stacked AutoEncoders structureNEAF
Activation functions used in the hidden
layers in the decoder section of each
AutoEncoder structure
Hyperparameter in the Stacked AutoEncoders structureNDAF
L2WeightRegularization coefficient used
in the AutoEncoder structure
Hyperparameter in the Stacked AutoEncoders structureAEL2
SparsityRegularization coefficient used
in the AutoEncoder structure
Hyperparameter in the Stacked AutoEncoders structureAESR
SparsityProportion coefficient used in
the AutoEncoder structure
Hyperparameter in the Stacked AutoEncoders structureAESP
Lscaledata coefficient used in the
AutoEncoder structure
Hyperparameter in the Stacked AutoEncoders structureAELD
Table 5. Parameters used in PSO.
Table 5. Parameters used in PSO.
PSO ParametersValues
Number of particles40
Solution Space9
Iteration Number500
Table 6. Activation functions corresponding to the NEAF value.
Table 6. Activation functions corresponding to the NEAF value.
NEAF ValuesActivation Function
1logsig
2satlin
Table 7. Activation functions corresponding to the NDAF value.
Table 7. Activation functions corresponding to the NDAF value.
NDAF ValuesActivation Function
1logsig
2satlin
3purelin
Table 8. Values corresponding to the AELD value.
Table 8. Values corresponding to the AELD value.
AELD ValuesValues
1true
2false
Table 9. Parameters and their values in the gbest particle for the Haberman’s Survival dataset.
Table 9. Parameters and their values in the gbest particle for the Haberman’s Survival dataset.
ParametersValues
NSAE1
NHED98
NEAF2 (satlin)
NDAF2 (satlin)
AEL20.0063
AESR3.96
AESP0.16
AELD2 (false)
Table 10. Parameters and their values in the gbest particle for the Echocardiogram dataset.
Table 10. Parameters and their values in the gbest particle for the Echocardiogram dataset.
ParametersValues
NSAE1
NHED159
NEAF1 (logsig)
NDAF2 (satlin)
AEL20.0038
AESR4.64
AESP0.56
AELD1 (true)
Table 11. Computational complexities of the methods used in the study.
Table 11. Computational complexities of the methods used in the study.
MethodsComputational ComplexityParameters
DTO(H)H: height of tree
KNNO(MLog(k)NLog(N))M: number of features
SVMO(N2)k: number of neighbors
ANNO(hNM)N: number of observations
Gradient BoostingO(d N logn d)d: number of trees
Gradient BaggingO(MNlog(N))h: number of hidden neurons
Proposed ModelO(iNK+iMhN)K: sum number of hidden neurons of AE
i: number of iterations
Table 12. Performance of methods used in survival after heart attack.
Table 12. Performance of methods used in survival after heart attack.
ModelsPrecisionRecallF-ScoreAccuracy
DT62.288873.3359
KNN63.337669.0956
SVM63.899275.4062
ANN64.708874.5862
Gradient Boosting69.566466.6759
Gradient Bagging63.899275.462
Proposed architecture70.599681.3572
Table 13. Performance of methods used in survival after breast cancer surgery.
Table 13. Performance of methods used in survival after breast cancer surgery.
ModelsPrecisionRecallF-ScoreAccuracy
DT31.104536.7366
KNN27.594032.6564
SVM25509.3376
ANN30303069
Gradient Boostng44.444042.1076
Gradient Bagging47.064043.2477
Proposed architecture81.256572.2289
Table 14. Literature comparison with the proposed method for the Haberman’s Survival dataset.
Table 14. Literature comparison with the proposed method for the Haberman’s Survival dataset.
AuthorsYearMethodAccuracy (%)F-Score
Bataineh et al. [10]2022Clonal Selection Algorithm + MLP76.10
Remya Ajai et al. [13]2023Heterogeneous Neurochaos Learning Architecture77.420.72
Sethi et al. [39]2023Neurochaos Learning features+Random Forest 0.61
Melin at al. [40]2024IT3FIS-GA77.05
Proposed method2024SAE-SOFTMAX-PSO890.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bülbül, M.A.; Işık, M.F. Survival Prediction of Patients after Heart Attack and Breast Cancer Surgery with a Hybrid Model Built with Particle Swarm Optimization, Stacked AutoEncoders, and the Softmax Classifier. Biomimetics 2024, 9, 304. https://doi.org/10.3390/biomimetics9050304

AMA Style

Bülbül MA, Işık MF. Survival Prediction of Patients after Heart Attack and Breast Cancer Surgery with a Hybrid Model Built with Particle Swarm Optimization, Stacked AutoEncoders, and the Softmax Classifier. Biomimetics. 2024; 9(5):304. https://doi.org/10.3390/biomimetics9050304

Chicago/Turabian Style

Bülbül, Mehmet Akif, and Mehmet Fatih Işık. 2024. "Survival Prediction of Patients after Heart Attack and Breast Cancer Surgery with a Hybrid Model Built with Particle Swarm Optimization, Stacked AutoEncoders, and the Softmax Classifier" Biomimetics 9, no. 5: 304. https://doi.org/10.3390/biomimetics9050304

Article Metrics

Back to TopTop