Next Article in Journal
Spatio-Temporal Variations and Influencing Factors of Country-Level Carbon Emissions for Northeast China Based on VIIRS Nighttime Lighting Data
Next Article in Special Issue
An Integrated System of Braden Scale and Random Forest Using Real-Time Diagnoses to Predict When Hospital-Acquired Pressure Injuries (Bedsores) Occur
Previous Article in Journal
Gender and Sex in Medical Practice: An Exploratory Study on Knowledge, Behavior, and Attitude among Sicilian Physicians
Previous Article in Special Issue
Machine Learning Techniques, Applications, and Potential Future Opportunities in Pressure Injuries (Bedsores) Management: A Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Integrated System of Multifaceted Machine Learning Models to Predict If and When Hospital-Acquired Pressure Injuries (Bedsores) Occur

1
Department of Systems Science and Industrial Engineering, Binghamton University, Binghamton, NY 13902, USA
2
Wound Ostomy Continence Nursing, ChristianaCare Health System, Newark, DE 19718, USA
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2023, 20(1), 828; https://doi.org/10.3390/ijerph20010828
Submission received: 13 November 2022 / Revised: 21 December 2022 / Accepted: 27 December 2022 / Published: 1 January 2023
(This article belongs to the Special Issue New Insights from Big Data and Advanced Analytics in Health Care)

Abstract

:
Hospital-Acquired Pressure Injury (HAPI), known as bedsore or decubitus ulcer, is one of the most common health conditions in the United States. Machine learning has been used to predict HAPI. This is insufficient information for the clinical team because knowing who would develop HAPI in the future does not help differentiate the severity of those predicted cases. This research develops an integrated system of multifaceted machine learning models to predict if and when HAPI occurs. Phase 1 integrates Genetic Algorithm with Cost-Sensitive Support Vector Machine (GA-CS-SVM) to handle the high imbalance HAPI dataset to predict if patients will develop HAPI. Phase 2 adopts Grid Search with SVM (GS-SVM) to predict when HAPI will occur for at-risk patients. This helps to prioritize who is at the highest risk and when that risk will be highest. The performance of the developed models is compared with state-of-the-art models in the literature. GA-CS-SVM achieved the best Area Under the Curve (AUC) (75.79 ± 0.58) and G-mean (75.73 ± 0.59), while GS-SVM achieved the best AUC (75.06) and G-mean (75.06). The research outcomes will help prioritize at-risk patients, allocate targeted resources and aid with better medical staff planning to provide intervention to those patients.

1. Introduction

Hospital-Acquired Pressure Injuries (HAPIs) is one of the most common health conditions in the US, which costs more than $ 26.8 billion annually [1]. Known by many names, such as pressure injury (PI), bedsore, or decubitus ulcer, the injury develops due to pressure or pressure in combination with shear that results in tissue deformation or tissue ischemia. HAPI refers to these injuries that occur while admitted to the healthcare system [2].
HAPI can happen almost everywhere on the body; however, they seem to happen more frequently over bony prominences or behind medical equipment, as shown in Figure 1. Pressure injuries are staged according to the level of exposed tissue [2]. Stage 1 pressure injuries present as intact skin with a localized area of non-blanchable erythema. Stage 2 is partial thickness skin loss with exposed dermis or a serum-filled blister. Stage 3 is full-thickness skin loss in which adipose tissue is visible. Stage 4 is full-thickness skin loss and tissue loss which exposes underlying structures such as fascia, muscle, tendon, or bone [2]. Unstageable injuries include injuries where the extent of skin and tissue loss is obscured by slough or eschar. In contrast, a Deep Tissue Pressure Injury (DTPI) is a localized area of non-blanchable deep red, maroon, or purple discoloration, which may evolve rapidly as the extent of the injury is revealed [2].
When HAPI develops in the hospital, the patient’s length-of-stay increases, and the patient requires additional resources. From the perspective of the individual patient and family, there is a decreased trust in the health care system, which can lead to difficulty managing the patient’s care. The staff who care for the patient as well as the hospital system may face litigation and quality of care concerns from local, state, and national agencies [3].
Standardized risk assessment and targeted prevention treatments can prevent most HAPI cases [2]. However, most acute or long-term care patients are at risk based on the standardized risk assessment. Therefore, the expense of prevention in terms of labor and preventive goods can be high. The Braden scale is an example of a risk assessment tool that nurses use to identify patients at risk for early HAPI development and implement individualized preventative measures against pressure injuries [4,5,6,7].
In the last decade, researchers have utilized Machine Learning (ML) approaches to predict if patients would develop HAPI before it occurs by utilizing patients’ Electronic Health Record (EHR) and therefore reduce the HAPI rate. Until now, no studies have answered the research question of when HAPI occurs for at-risk patents. This research is the first to have an integrated system of multifaceted ML models to predict if and when HAPI occurs by collecting a new piece of information not introduced in the literature, i.e., time for HAPI. Furthermore, it is the first research that combines Genetic Algorithm (GA), Cost-Sensitive (CS) learning, and Grid Search (GS) with ML algorithms to provide an indication as to not only who will develop HAPI but also when HAPI is likely to occur by training and testing two integrated models for highly unbalanced problems.
This paper is structured as follows: Section 2 summarizes and criticizes the related literature in predicting HAPI. Section 3 describes the variables, data source, participants, and model development. Section 4 summarizes the results of the developed method and compares it with the most used classification algorithms. Section 5 discusses the output of the suggested results, their implications for the new approach in the medical field, and limitations of this research. Lastly, Section 6 provides the conclusion and future direction of this research.

2. Related Literature

A systematic literature review is applied to the field of applying ML in predicting HAPI. The database used in this criterion were PubMed, Web of Science, Scopus, and Science Direct. The timeline is between 2007 through July 2022. The database search keywords were pressure injury, pressure ulcer, machine learning, deep learning, data mining, hospital-acquired pressure injury, HAPI, early detection, predictive modeling of pressure injury, bedsores, decubitus ulcer, and others. In the end, 26 studies met the criteria. [9] is excluded because it was a survey on predicting HAPI using ML. The remaining 25 studies used ML methods to predict HAPI early [1,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33].
HAPIs rate is considered a rare event in hospitals because it occurs among a minor volume of the total population [2]. Two studies out of 25 had a highly unbalanced dataset of less than 3% [20,26]. However, most of the other studies designed their samples to have a high portion of HAPI. For example, HAPI rate was 52.31% [10], 50.00% [16], 31.92% [18], 28.78% [19], 28.10% [22], 20.10% [25], and 20.00% [34].
Random Oversampling (RO) techniques were applied to most studies (i.e., in 76% of the studies) [1,10,11,14,15,16,18,19,20,21,22,23,25,26,28,30,31,32]. On the other hand, four studies used Synthetic Minority Oversampling Technique (SMOTE) to deal with the unbalanced dataset [12,13,24,27]. Oversampling techniques can be a solution by replicating the patients with HAPI (minor class); good results might be generated for training purposes. However, in implementation, overfitting might occur. Therefore, the algorithm will misclassify any HAPI cases and consider them as non-HAPI because it is already trained on a minor sample. Therefore, there is a need to use other advanced/hybrid models to overcome the challenge of the unbalanced dataset.
All studies used traditional ML methods to predict HAPI. Most of the researchers tried multiple approaches to the same study. The most used approaches are Logistic Regression (LR), which was used 17 times in the above studies during model development and experimentation [1,10,12,13,15,17,18,19,20,21,22,24,26,28,30,32,33], Random Forest (RF) was used 12 times [1,11,12,13,17,19,24,26,27,28,33,34], Decision Tree (DT) was used 10 times [1,14,15,18,19,20,24,25,30,33], Support Vector Machine (SVM) was used nine times [1,15,17,24,25,26,28,29,30], Multilayer Perceptron (MLP) was used eight times [1,12,17,24,25,27,28,29], and k-nearest neighbor (kNN) was used three times [19,27,28]. Other algorithms were used in the experimentations: Linear Discriminant Analysis (LDA) [27] and Adaptive Boosting (AdaBoost) [12].
Nevertheless, metaheuristic optimization was not used till now in this field to optimize the ML hyperparameters. Only two researchers used GS optimization [17,26]. Besides, CS learning was used only once to deal with the unbalanced HAPI dataset [17] and was applied to a benchmark dataset.
All researchers utilized ML algorithms to predict only which patients will develop HAPI before it occurs by utilizing the patient’s historical data in the EHR [1,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33]. However, none of the studies predict when HAPI might occur for the predicted HAPI cases. Furthermore, the status of patients with HAPI changes during their stay in the hospital. All the studies used a snapshot of static data (i.e., the status of patients at admission or the most recent diagnosis of patients) [1,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33]. However, no studies use variables at multiple points (i.e., multiple tests per patient, which change over time). However, having one record per patient does not count the changes in the status of patients during their stay.
In summary, no more than 25 studies have been conducted in this field that answer the question who will develop HAPI among the patients. This is insufficient and incomplete information for the clinical team because knowing who would develop HAPI in the future (classification tasks) does not help differentiate the severity and urgency of those predicted cases. Further, conducting prevention actions for all predicted patients would require more resources, time, and costs. On the other hand, none of the studies used metaheuristic and CS learning to deal with the highly unbalanced problem of predicting HAPI. Most of the studies used traditional ML approaches and oversampling techniques. Finally, all studies used one snapshot of the patient records (i.e., features) to predict HAPI. None of the above research considered the changes in patients’ status during their hospital stay.
This research will fill the above gaps in the literature by developing an integrated system of multifaceted ML models to predict if and when HAPI occurs. This model considers the changes in the status of patients at multiple points in time. It integrates GA with CS learning to enhance the performance of SVM when dealing with highly unbalanced HAPI dataset. This research is the first to fill the above gaps and address HAPI time in the literature to prioritize who is at the highest risk and when the highest risk will occur.

3. Research Methodology

3.1. Methods

The scope of this study is patients who were admitted to ChristianaCare hospital located in Delaware, USA and discharged between May 2020 and February 2022 (n = 15,889). Patients under 18 years old were excluded from the scope of this study. Additionally, labor and delivery patients, and emergency visits were excluded. Patients with HAPIs (n = 485) were identified through nurse documentation of validated HAPIs.

3.2. Data Source

The variables were extracted from a SQL database that pulls information from patients’ EHR. Wound, Ostomy, Continence (WOC) nurses kept track of patient records with validated HAPIs by manually documenting their notes in tracking documents; that was the source for identifying HAPI patients (n = 485). These documents included admission date and time, HAP occurrence date and time, and multiple variables across different points of time during the patient visit. The final dataset included the unique record per patient visit represented in Figure S1 in Supplementary Materials.

3.3. Variables

The goal of this study is to predict patients’ risk for developing HAPI and when they are expected to develop HAPI based on multiple risk factors. Ninety-eight risk factors that include the Braden risk assessment subscales were used as inputs for a ML model as summarized in Table S1 in supplementary material. These variables were selected based on the previous literature survey and clinicians’ feedback. Changing variables were measured based on three points of time across the patient stay; on admission (First), before discharge (Last), and the average value of all measured values of variables during stay.
This model had only 485 patients who had HAPI; therefore, predicting HAPI time did not provide accurate results. Therefore, the problem was converted from predicting continuous time values to a classification problem by labeling the time to develop HAPI into two categories; high-risk patients who are expected to develop HAPI in fewer than seven days, and medium-risk patients who are expected to develop HAPI in more than seven days. The medical team confirmed this categorization of risk where seven days is an adequate period to provide earlier preventive actions. Furthermore, the clinical team will use the output of predicting HAPI time to stratify at-risk patients. Therefore, even if a continuous prediction of the time is found, the predicted time will be converted into risk satisfaction: at high risk to provide an immediate intervention and at medium risk to provide later interventions.

3.4. Model Development

The integrated system of multifaceted ML models has been developed in two separate phases. Phase 1 predicts if the patient will develop HAPI, and Phase 2 predicts when HAPI is expected to occur for at-risk patients as shown in Figure 2. Having two different models allows for flexibility in selecting the most important features for each phase, one target per phase, which uses different optimized parameters and validation processes. However, the data preprocessing step was the same for both phases. The patients’ distribution in each phase is summarized in Table 1.
Preprocessing data before developing a ML model is an important step to ‘maximize’ information and knowledge extraction from variables. It also reshapes the data into an appropriate format that is readable by ML algorithms. Imputation, normalization, and categorization are all examples of preprocessing applied to the dataset to remove errors and missing values.
Feature selection was applied to select the top important variables for predicting HAPIs. Recursive Feature Elimination (RFE) was the algorithm used to perform that; this algorithm fits a predictive model to predict class through iterations of removing variables from weakest to strongest until an optimal set of variables is selected [35,36].

3.4.1. Phase 1: Predict “If” Patient Will Develop HAPI

Phase 1 integrates GA with CS-SVM (GA-CS-SVM) to predict if a patient will develop HAPI before occurrence. This model considers 485 HAPI and 15,404 non-HAPI patients (n = 15,889); each patient has 98 risk factors/features. RFE was used to select the best features that affect HAPI development. The top 63 features were selected and used to develop the SVM model. The dataset was divided into 80% for training the model with 10-fold cross-validation, and the remaining 20% was used to test the model’s performance. GA was used to fine-tune the CS learning parameters to deal with the highly unbalanced training dataset and to enhance the performance. In this case, GA was used to select the best combination of class weights for False Negative (FN) and False Positive (FP) as explained below. Two oversampling techniques (SMOTE and RO) [37] were applied and compared for the same dataset to validate the suggested approach in terms of sensitivity, AUC, Geometric Mean (G-mean), and False Positive Rate (FPR). Furthermore, the performance of the optimized SVM (i.e., GA-CS-SVM) was compared to the non-optimized SVM to measure the optimization effect. Moreover, the seven commonly used algorithms in predicting HAPI in the literature were applied and tested to validate the suggested SVM model. A t-test compared the performance between GA-CS-SVM vs. non-optimized SVM. In contrast, Analysis of variance (ANOVA) was used to determine if there are any statistical differences between the means of all other validation methods, which includes GA-CS-SVM. Lastly, a t-test was used to measure the difference between the optimized SVM and the best method among other methods. The confidence level is kept at a 0.05 error margin for 50 experiments.
SVM is a supervised ML algorithm that is heavily used in the medical field. Research shows that it provides decent classification results in predicting HAPI. Nine out of 25 studies that predict HAPI adopted SVM [1,15,17,24,25,26,28,29,30]. SVM can deal with simple and complex classifications and is robust to high-dimensional data [38]. This research deals with 63 features selected by FRE and adopts a linear SVM.
The objective of SVM in binary classification problems is to find a margin hyperplane that maximizes the minimum distance to the hyperplane (i.e., find the optimal separation line with the maximum margin w between data points of two classes) [38], as shown in Figure 3. Having the optimal separation line means solving the mathematical optimization model below [39,40].
m i n   1 2 | | w | |   2 + C   1 N ξ i  
y i ( w T x + b ) 1 ξ i      i = 1 ,   ,   N
ξ i 0      i = 1 ,   ,   N
whereas x is the HAPI dataset with N patients,   x = ( x 1 ,   x 2 ,   ,   X N ), x i , i = 1, …, N represents a patient with m features, and y i ∈ {0,1}, 0 denotes non-HAPI patients, and 1 denotes patients with HAPI. w represents the separating margin, C is the penalty parameter or regularization parameter that balances the margin w and the training error (i.e., loss). ξ i is the slack variable penalty during training, i = 1, ..., N ,   ξ   R + N , and b is the bias or scalar offset [39,40].
The drawback of the above SVM structure is that it is inefficient to deal with highly unbalanced datasets. It is because C is fixed, and all data points (patients) are treated as equally likely during the training process. Therefore, the Cost-Sensitive SVM (CS-SVM) or Biased Penalties SVM (BP-SVM) [38,39] introduces different penalty coefficients C 1 and C 0 for HAPI and non-HAPI SVM slack variables during the training process [39,40]. The formulation for CS-SVM is provided below:
  w , b , ξ argmin 1 2   | | w | |   2 + C [ C 1 { i | y i = 1 } ξ i + C 0 { i | y i = 0 } ξ i ]
y i ( w T x + b ) 1 ξ i      i = 1 ,   ,   N
ξ i 0      i = 1 ,   ,   N
where C 1   represents the penalty/cost of a false negative (FN) that classifies HAPI as non-HAPI (i.e., the penalty for misclassification of HAPI cases or cost of minority class HAPI Type 2 error), and C 0 represents the penalty/cost of a false positive (FP) that classifies non-HAPI as HAPI (i.e., cost of majority class non-HAPI); therefore, CS-SVM assigns different costs/weights for FN, and FP, in most cases, C 1 > C 0 . As a result, FN will be minimized because its cost will be high (more penalty for misclassification). Therefore, the algorithm will learn to avoid misclassifying HAPI records during the training process, which increases the model sensitivity.
The size of the search space has an infinite number of scenarios because C 1 and C 0 can be any positive real numbers (i.e., C 1 R + N , and C 0 R + N ). To reduce the search space C 1 and C 0 are bounded between 1 and 100, which is deemed sufficient penalty to impose on the CS-SVM. Furthermore, it is assumed C 1 > C 0 as in the literature [39,40] . However, the search space is still infinite 0 < C 1 ≤ 100, 0 < C 0 ≤ 100, and C 1 > C 0 , C 1 , C 0 can be any real value within the boundaries. Therefore, there is a need for a heuristic method to find the semi-optimal solutions rather than trial and error in this infinite search space.
There are several methods used in the literature to find the values of C 1   and   C 0 in CS-SVM, such as Kernel-based Possibilistic c-Means (KPCM) algorithm, random search, intuition, Fuzzy SVM (FSVM), Bayes consistent classier, 2ν approach, incremental CS learning, self-adaptive cost weights-based CS Large margin Distribution Machine (CS-LDM) [39,40,41,42,43,44,45,46,47,48,49].
GA is an evolutionary algorithm that uses a stochastic approach for global search. It has been used heavily to optimize hyperparameters for general ML methods and SVM [40,50,51,52,53,54,55,56,57,58]. GA is used in this research as a robust heuristic tool to identify the semi-optimal distribution of C 1 and C 0 that satisfies Equations (4)–(6). CS-SVM is used as an emulator that represents the GA’s objective function, which maximizes the AUC. The AUC of CS-SVM is calculated based on a 10-fold cross-validation of the training set.
The pseudocode of the hybrid GA-CS-SVM is shown in Algorithm 1. This algorithm optimizes the CS-SVM parameters, which are encoded as real values in GA; initially a set of random solutions is selected, and each solution gets included in the training process. Once training is complete, each solution is validated with the 10-fold cross-validation, which represents the fitness value of each solution.
Algorithm 1. Combining GA-CS-SVM for optimizing the CS learning parameters
1: Set GA parameters (Pc, Pm, n, gmax)
2 :   Encode   solutions   ( CS   learning   parameters : C 1 , C 0 ) using real value encoding
3: Randomly generate n solutions
4: Calculate the fitness value (AUC) of each solution by the trained CS-SVMs
5: for i = 1 to gmax do
6:       for j = 1 to n/2 do
7:             Select two parents
8:             Crossover to create two children with Pc
9:              Mutate children with Pm
10:       end for
11: Replace parents with children
12: end for
13: Return the best solution
For every GA generation, a tournament selection process runs over k solutions to select two parents for breeding. The selected pair are combined using the crossover operator and then mutated to create two mutated children (solutions). The process of selection, crossover, and mutation is iterated until a certain number of solutions/children is generated. After that, the selected children replace an entire generation for the next iteration. The solutions keep evolving until the predefined maximum number of generations (gmax) is reached. Then, the best solution is delivered through the hybrid GA-CS-SVM that has the highest AUC value.
For the GA, the following parameters were used: tournament selection to select the parents (k = 2), population size = 50, 100 generations as a stopping criterion, crossover probability (Pc) and mutation probability (Pm) were 1.00 and 0.01, respectively. A weighted average of 60% of parent 1 and 40% of parent 2 was used to combine the two parent solutions. A Gaussian distribution was used with mean 0, and standard deviation of 0.01 to find the mutation value. Finally, the optimal values of C 1 and   C 0 to achieve the best AUC were 43.32 and 6.26, respectively.

3.4.2. Phase 2: Predict “When” a Patient Is Likely to Develop HAPI

Phase 2 combines GS with SVM (GS-SVM) to predict the second target, which is the timing of HAPI for at-risk patients. This model considers only the 485 patients with HAPI, with 98 risk factors. One hundred thirty-five patients developed HAPIs within the first seven days (at high risk), and the remaining developed HAPI after seven days (at medium risk). The distribution for the high-risk vs. medium-risk patients is 28.00% vs. 72.00%.
RFE was utilized to select the top features that impact the HAPI timing. Therefore, the 39 best features were selected and used as inputs in Phase 2. Leave-one-out Cross-Validation (LOOCV) was adopted in Phase 2 to deal with the small HAPI dataset [59]. Therefore, the learning algorithm was applied once for each record, which uses all other records as a training set and the selected patient as a single-patient test set. Because the dataset is not highly unbalanced as in Phase 1, there was no need to use CS learning. Instead, a GS with a 10-fold cross-validation was used to tune and select the best hyperparameters of the SVM.
Alternatively, SMOTE and RO were used with LOOCV. The performance of the optimized SVM (i.e., GS-SVM) was compared to the non-optimized SVM. In addition, the seven algorithms used in Phase 1 were used to validate the suggested model in Phase 2. Unlike the 80% training vs. 20% testing, LOOCV was applied once for each patient. Therefore, there is no variability in such a method. Therefore, confidence levels are not available for the results of Phase 2.
In GS, the algorithm searches exhaustively through a predefined manual subset of the hyperparameter space of SVM [60]. The hyperparameters used for the GS are the following: kernel type to be used in the SVM [Linear, Polynomial, Radial Basis Function, Sigmoid], the regularization parameter C [1, 10, 100, 1000], gamma for Radial Basis Function [0.0001, 0.001, 0.1, 1], and polynomial degree [2, 3, 4]. The algorithm provides the best combination of hyperparameters with the best performance when the regularization parameter C is 1, the kernel is linear, and gamma is 1.

3.5. Performance Metrics

Check for overfitting was conducted through comparison of the model’s performance metrics on training and testing sets; this included sensitivity, G-mean, AUC, and False Positive Rate (FPR). Refer to Table 2 for detailed explanation of the confusion matrix metrics that include False Negative (FN), False Positive (FP), True Negative (TN), and True Positive (TP). FN represents HAPI patients who were missed by the model. FP represents healthy patient who were predicted by model to have HAPI. TN represents healthy patients predicted correctly by the model. TP represents HAPI patients who were correctly predicted as at risk by the model.
Sensitivity is the ratio of TP to all actual cases with HAPIs. FPR measures the probability of non-HAPI cases predicted as HAPI cases. G-Mean measures the balance between classification performances on both the majority non-HAPI and minority HAPI cases. Lastly, AUC measures the ability of the model to distinguish between patients with and without HAPI [61].
Sensitivity = TP TP + FN × 100 %
FPR = FP FP + TN × 100 %
G - mean = ( TP TP + FN ) × ( TN TN + FP ) × 100 %

4. Results

This research developed an integrated system of multifaceted ML models to predict if and when HAPI occurs for at-risk patients. This study collected data for 15,889 patients with a 3% HAPI rate. Ninety-eight risk factors with two targets for each patient were collected and preprocessed. The two targets are HAPI occurrence and time to develop HAPI from admission to the occurrence. Risk factors that indicate the patient’s status were collected at three points. These factors change due to the status of the patients during the length-of-stay. Comprehensive features and diagnoses were collected to represent patients’ environment, such as demographical factors, lab factors, medical device factors, medications, diagnosis factors, and medical factors. Two different types of predictive models were investigated in two phases. Phase 1 integrated GA as a heuristic method with CS-SVM (GA-CS-SVM) to handle the high imbalance HAPI dataset to predict the occurrence of HAPI. RFE was used in Phase 1 to extract the best features. The dataset was separated into training with 10-fold cross-validation and testing. Phase 2 adopted LOOCV to train a different model to predict when HAPI will occur for HAPI patients. In this phase, the 39 best features were selected by RFE. GS was used to optimize the hyperparameters of the SVM in Phase 2. Both phases were compared to the seven algorithms used to predict HAPIs. Moreover, two oversampling methods were compared to the proposed approach. Statistical tests were performed to highlight the statistical significance of the proposed approach. Lastly, for each phase, Random Forest was used to measure each feature’s influence on prediction.
Figure 4 shows the top 20 features for each phase; the red-labeled features are common top features such as Count of Glasgow Score (GCS) comment, Feeding Tube, Number of Surgeries, and some Braden subscales such as Sensory Perception Status (Average), and Mobility Status (Average). Having two phases with a feature selection for each phase’s target allows flexibility in selecting the most important features.
In Phase 1 (predicting if HAPI will happen), GA-CS-SVM achieved the best sensitivity (74.29), AUC (75.79), and G-mean (75.73) compared to the most common algorithms used in predicting HAPIs (SVM, LR, AdaBoost, LDA, KNN, DT, RF, MLP) and when compared to other oversampling methods (RO and SMOTE). The results are summarized in Table 3. Training with 10-fold cross-validation and testing results were presented to reflect no overfitting for the proposed method. It is worth mentioning that overfitting happened when adopting SMOTE. A 95% Confidence Level (CL) of 50 experiments was implemented for each scenario.
t-test shows that there is a statistical significance between (un-optimized) SVM and an optimized one (GA-CS-SVM) in terms of sensitivity, AUC, G-mean, and FPR (p-value < 0.05) as shown in Table 4 and presented in Figure 5.
ANOVA is performed to determine a significant difference between all algorithms by testing for differences in means using variance, as shown in Table 4. p-value is less than 0.05 for sensitivity, AUC, G-mean, and FPR, which indicates a statistical difference among the methods. Figure 6 presents the mean and confidence interval for GA-CS-SVM vs. other methods in terms of sensitivity, AUC, G-mean, and FPR. It is observed that GA-CS-SVM performs the best. However, oversampling technique has acceptable results compared to the remaining techniques. Therefore, the t-test is performed between GA-CS-SVM and balancing using oversampling. Table 4 shows a statistical significance between RO and GA-CS-SVM in terms of sensitivity, AUC, G-mean, and FPR (p-value < 0.05), as presented in Figure 7.
In Phase 2 (predicting when HAPI happens), the optimized SVM (i.e., GS-SVM) achieved the best sensitivity (75.56), AUC (75.06), and G-mean (75.06) compared to the most common algorithms, as shown in Figure 8. However, balancing methods with LOOCV (RO and SMOTE) perform better than the GS-SVM because they had similar samples from the training set, which increases the chances of overfitting. In contrast, GS-SVM adopted LOOCV without oversampling. Therefore, the results are less than the oversampled one. Table 5 summarizes the performance metrics for GS-SVM and all other techniques. As discussed in the methodology section, the LOOCV is applied once for each patient, using all other patients as a training set and the selected patient as a single-patient test set. Therefore, there is no variability and statistical tests as applied in Phase 1. However, it is observed that the effect of optimization for SVM that uses the GS technique on SVM (GS-SVM) increases the sensitivity from 59.26 to 75.56 for instance, whereas AUC increased from 72.49 to 75.06.

5. Discussion

Phase 1 adopted 63 features using RFE; Phase 2 identified 39 features using RFE; 30 are common for both phases, as shown in bold in Table S1 in Supplementary Materials. Most of the Braden Scale subfactors are common factors, which further validates that the Braden Scale is a critical assessment. The proposed approach takes the daily assessment to a new level, one that would not feasibly be able to be performed by the nurse who cares for the patient; it can take a combination of the first score, average score, and most recent score when determining HAPI and the timing of HAPI. In addition to the Braden Score provided by nurses who care for the patient throughout the hospitalization, the proposed model takes all other historical factors into account, such as the prior year’s inpatient visit count, the number of surgeries, specific comorbidities; all factors that help determine the level of risk but are things that would require significant additional resources if performed by the nurse who cares for the patient at the bedside.
At the basis of the Orlando theory of nursing process is the holistic assessment of the patient, formulation of nursing diagnosis, planning, implementation, and evaluation of plan of care [62]. Assessment tools have been developed to assist with the identification of at-risk patients for some of the most common nursing diagnoses, that include patients at risk for development of HAPIs. The currently available risk assessment scores have adequate specificity and sensitivity to identify patients at risk but have an elevated false positive rate to identify a large number of at-risk patients who do not develop pressure injuries. Phase 1 of the ML project predicted which patients will develop HAPIs, the second phase is aimed at prioritizing at-risk patients by predicting the timing of HAPI. The clinical implications of not only knowing who is at risk of developing HAPI, but also when during their hospitalization this is likely to occur are many. The first and easiest to measure is the reduction in HAPIs, which results in reduced length-of-stay and reduced financial penalties. The second is a reduction of patient harm, which results in reduced financial penalties and increased external hospital ratings/confidence of care. For the bedside staff, the ability to target and focus on those patients most at risk for development of HAPIs will result in better patient care and appropriate allocation of resources. Using ML in combination with clinical assessment tools, there is a reduction in the number of patients identified as being at risk; the addition of the likely time frame of development will further reduce the number of patients who require advanced prevention techniques. Targeting highest-risk patients will allow for nursing and support staff to individualize the care plan and allocate the greatest resources to those most at risk, while continuing to provide appropriate nursing care to all at-risk patients as shown in Figure 9.
Phase 1 helped to narrow down which patients are at most risk; the second phase helps determine the time frame the patient is most likely to develop a pressure injury during the hospital stay. According to the Agency for Healthcare Research and Quality (AHRQ), the average hospital length-of-stay was 4.5 days in 2021; however, the length-of-stay for vulnerable populations and those with chronic conditions tends to be longer. The average length-of stay for a person who experiences homelessness is around 6.5 days [63], whereas those in recent years who experience COVID-19 complications can have an average length-of-stay > 25 days [64]. Socioeconomic factors can also increase the risk of prolonged length-of-stay in the hospital; a small percentage of hospitalized patients have a length-of-stay measured in months [65]. The second phase of the project further stratifies level of risk over time, which helps to prioritize not only who is at risk, but when they will be at highest risk. Phase 2 gives patients’ level of risk per week. The time frame of seven days was chosen secondary to guidance from Kottner et al. (2019) [2] on the development of a DTPI. Development of a DTPI takes place up to 72 h prior to visibility at skin level [2]. Minimal intervention in a DTI would be 72 h, but an earlier intervention would better maximize interventions and skin condition; however, beginning intense interventions too early can be burdensome for the patient and caregiver. Seven days of intervention will help maximize tissue tolerance and reduce the risk of prolonged pressure over bony prominences, which is the goal in pressure injury prevention [2].
All of the proposed models in the literature were offline models that did not run on active patients’ records and update their inputs continuously, and because these models where built and tested on static dataset that did not change across patient stays, static data represent a summary snapshot of the data for the patient stay regardless of how many days were spent in the hospital [1,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33], In contrast, this research considers the changes of patients’ records at three points of time [66].
The study methods used to gather the information for the HAPI risk score uses a combination of subjective and objective information gathered from the patient’s chart. The Braden Scale performed by the bedside nurse has an inter-rater reliability of 87.1% for overall score but tends to be lower at the level of the Braden Risk subscale level, for example, of moisture inter rater reliability of only 13% [67]. Differences among nurses who perform the Braden Scale may provide inaccurate or an incomplete picture of the patient’s current level of functioning. Furthermore, the suggested approach does not consider all the changes in patients’ status from admission to discharge; it considers the first, last, and average status during the length-of-stay.

6. Conclusions and Future Work

HAPIs is one of the most prevalent health disorders in the United States, with yearly expenses exceeding $ 26.8 billion. It is often referred to as a pressure injury, bedsore, or decubitus ulcer. Sustained skin pressure can cause injuries to the skin’s underlying tissue. Most of the HAPI cases can be prevented through prevention interventions. Prevention is the most effective method for managing HAPIs. Most patients in acute care or long-term care settings are considered at risk, and the cost of interventions and prevention for all patients in terms of nurses and preventive products can be significant. In the literature, researchers utilized ML approaches to predict if patients would develop HAPI before it occurs by utilizing EHR to provide prevention actions. However, no more than 25 studies have been conducted in this field that answer the question who will develop HAPI among the patients. This is insufficient and incomplete information for the clinical team because knowing who would develop HAPI in the future (classification tasks) does not help differentiate the severity and urgency of those predicted cases. Further, conducting prevention actions for all predicted patients would require more resources, time, and costs. Moreover, patients predicted as at risk, most likely will remain at risk until discharge. Therefore, this research introduces for the first time a robust integrated ML approach to answer the question of not only who will develop HAPI, but also the timing of HAPI when it occurs for at-risk patients. The performance of the developed models is compared with state-of-the-art models in the literature, which achieved higher sensitivity, AUC, G-mean, and FPR than the eight most common algorithms used to predict HAPI and other balancing methods.
This research outcome will help the medical team prioritize the at-risk patients and allocate additional targeted resources to the patients who will likely have HAPIs during a specific time period (i.e., highest risk patients). Furthermore, this work will reduce patient harm (HAPI rate), which potentially reduces the length-of-stay. Lastly, it will help with better planning of medical staff to provide intervention to predicted HAPI patients and save costs, time, and resources. Future work will investigate the feasibility to automate the categories of the Braden Score to use a multidisciplinary approach to determine level of risk, which includes data pulled from Occupational or Physical Therapies, Registered Dietitian, as well as attending or consulting providers. Moreover, studies will utilize the online learning concept to capture all patients’ records during their stay (i.e., dynamic model). Lastly, a multi-task learning model can be developed for future work by training one dataset for both targets rather than training two different models separately.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/ijerph20010828/s1, Table S1: Models variables (features/risk factors for predicting phase 1 and phase 2), Figure S1. Database connection diagram.

Author Contributions

Conceptualization, O.Y.D. and S.S.L.; Methodology, O.Y.D. and S.S.L.; Software, O.Y.D.; Validation, O.Y.D. and S.S.L.; Formal analysis, O.Y.D.; Investigation, O.Y.D. and S.S.L.; Resources, O.Y.D.; Data curation, O.Y.D.; Writing—original draft, O.Y.D. and L.M.; Writing—review & editing, O.Y.D., S.S.L. and L.M.; Visualization, O.Y.D.; Supervision, S.S.L.; Project administration, O.Y.D. and S.S.L.; Funding acquisition, O.Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Ethical review and approval were waived for this study because this research is considered a quality improvement project that does not meet the federal definition of research in accordance with 45 CFR 46.102(l) and therefore does not require review by the ChristianaCare Institutional Review Board (IRB).

Informed Consent Statement

This research was based on a secondary analysis of EHRs at ChristianaCare Hospital in Delaware, United States. The ChristianaCare Hospital Institutional Review Board approved the study without written informed consent from participants. The dataset was de-identified where patient identifiers were removed before processing and modeling.

Data Availability Statement

These data were extracted from ChristianaCare Health Systems’ databases in Delaware, United States. It was de-identified for the purpose of this research. The data is not available for the public as it is owned by ChristianaCare.

Acknowledgments

The authors would like to extend their gratitude to Wei Liu (Lead Data Scientist), Raghad Alkhawaldeh (Principal Organizational Excellence Consultant), Erica Aiken (Wound Ostomy Continence Clinical Leader), Susan Mascioli (Vice President of Nursing Quality and Safety), Vernon L. Alders (Vice President of Organizational Excellence), and Edward F. Ewen (Chief Data and Analytics Officer) in ChristianaCare Health System, and Mohammad T. Khasawneh (SUNY Distinguished Professor and Chair of the Systems Science and Industrial Engineering Department) at Binghamton University, for providing valuable feedback during the development of this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Song, W.; Kang, M.J.; Zhang, L.; Jung, W.; Song, J.; Bates, D.W.; Dykes, P.C. Predicting Pressure Injury Using Nursing Assessment Phenotypes and Machine Learning Methods. J. Am. Med. Inform. Assoc. 2021, 28, 759–765. [Google Scholar] [CrossRef] [PubMed]
  2. Kottner, J.; Cuddigan, J.; Carville, K.; Balzer, K.; Berlowitz, D.; Law, S.; Litchford, M.; Mitchell, P.; Moore, Z.; Pittman, J.; et al. Prevention and Treatment of Pressure Ulcers/Injuries: The Protocol for the Second Update of the International Clinical Practice Guideline 2019. J. Tissue Viability 2019, 28, 51–58. [Google Scholar] [CrossRef] [PubMed]
  3. Hartman, M.; Martin, A.B.; Washington, B.; Catlin, A. The National Health Expenditure Accounts Team National Health Care Spending In 2020: Growth Driven by Federal Spending In Response To The COVID-19 Pandemic. Health Aff. (Millwood) 2022, 41, 13–25. [Google Scholar] [CrossRef] [PubMed]
  4. Gaspar, S.; Peralta, M.; Budri, A.; Ferreira, C.; Gaspar de Matos, M. Pressure Ulcer Risk Profiles of Hospitalized Patients Based on the Braden Scale: A Cluster Analysis. Int. J. Nurs. Pract. 2022, 28, e13038. [Google Scholar] [CrossRef] [PubMed]
  5. Cheng, F.M.; Jin, Y.J.; Chien, C.W.; Chuang, Y.C.; Tung, T.H. The Application of Braden Scale and Rough Set Theory for Pressure Injury Risk in Elderly Male Population. J. Mens. Health 2021, 17, 156–165. [Google Scholar] [CrossRef]
  6. Huang, C.; Ma, Y.; Wang, C.; Jiang, M.; Yuet Foon, L.; Lv, L.; Han, L. Predictive Validity of the Braden Scale for Pressure Injury Risk Assessment in Adults: A Systematic Review and Meta-Analysis. Nurs. Open 2021, 8, 2194–2207. [Google Scholar] [CrossRef]
  7. Jansen, R.C.S.; Silva, K.B.d.A.; Moura, M.E.S. Braden Scale in Pressure Ulcer Risk Assessment. Rev. Bras. Enferm. 2020, 73, e20190413. [Google Scholar] [CrossRef]
  8. Zahia, S.; Sierra-Sosa, D.; Garcia-Zapirain, B.; Elmaghraby, A. Tissue Classification and Segmentation of Pressure Injuries Using Convolutional Neural Networks. Comput. Methods Programs Biomed. 2018, 159, 51–58. [Google Scholar] [CrossRef]
  9. Ribeiro, F.; Fidalgo, F.; Silva, A.; Metrôlho, J.; Santos, O.; Dionisio, R. Literature Review of Machine-Learning Algorithms for Pressure Ulcer Prevention: Challenges and Opportunities. Informatics 2021, 8, 76. [Google Scholar] [CrossRef]
  10. Ahmad, M.A.; Larson, B.; Overman, S.; Kumar, V.; Xie, J.; Rossington, A.; Patel, A.; Teredesai, A. Machine Learning Approaches for Pressure Injury Prediction. In Proceedings of the 2021 IEEE 9th International Conference on Healthcare Informatics, ISCHI 2021, Victoria, BC, Canada, 9–12 August 2021; Institute of Electrical and Electronics Engineers Inc.: New York City, NY, USA, 2021; pp. 427–431. [Google Scholar]
  11. Alderden, J.; Pepper, G.A.; Wilson, A.; Whitney, J.D.; Richardson, S.; Butcher, R.; Jo, Y.; Cummins, M.R. Predicting Pressure Injury in Critical Care Patients: A Machine Learning Model. Am. J. Crit. Care 2018, 27, 461–468. [Google Scholar] [CrossRef]
  12. Alderden, J.; Drake, K.P.; Wilson, A.; Dimas, J.; Cummins, M.R.; Yap, T.L. Hospital Acquired Pressure Injury Prediction in Surgical Critical Care Patients. BMC Med. Inform. Decis. Mak. 2021, 21, 1–11. [Google Scholar] [CrossRef] [PubMed]
  13. Anderson, C.; Bekele, Z.; Qiu, Y.; Tschannen, D.; Dinov, I.D. Modeling and Prediction of Pressure Injury in Hospitalized Patients Using Artificial Intelligence. BMC Med. Inform. Decis. Mak. 2021, 21, 1–13. [Google Scholar] [CrossRef] [PubMed]
  14. Borlawsky, T.; Hripcsak, G. Evaluation of an Automated Pressure Ulcer Risk Assessment Model. Home Health Care Manag. Pract. 2007, 19, 272–284. [Google Scholar] [CrossRef]
  15. Chen, Y.-C.; Wang, P.-C.; Su, C.-T. Pressure Ulcers Prediction Using Support Vector Machines. In Proceedings of the 2008 4th International Conference on Wireless Communications, Networking and Mobile Computing, Dalian, China, 12–14 October 2008; IEEE: New York, NY, USA; pp. 1–4. [Google Scholar]
  16. Cichosz, S.L.; Voelsang, A.B.; Tarnow, L.; Hasenkam, J.M.; Fleischer, J. Prediction of In-Hospital Pressure Ulcer Development. Adv. Wound. Care (New Rochelle) 2019, 8, 1–6. [Google Scholar] [CrossRef] [Green Version]
  17. Cramer, E.M.; Seneviratne, M.G.; Sharifi, H.; Ozturk, A.; Hernandez-Boussard, T. Predicting the Incidence of Pressure Ulcers in the Intensive Care Unit Using Machine Learning. eGEMs (Gener. Evid. Methods Improv. Patient Outcomes) 2019, 7, 49. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Deng, X.; Yu, T.; Hu, A. Predicting the Risk for Hospital-Acquired Pressure Ulcers in Critical Care Patients. Crit Care Nurse 2017, 37, e1–e11. [Google Scholar] [CrossRef] [Green Version]
  19. Do, Q.; Lipatov, K.; Ramar, K.; Rasmusson, J.; Pickering, B.W.; Herasevich, V. Pressure Injury Prediction Model Using Advanced Analytics for At-Risk Hospitalized Patients. Patients. J. Patient Saf. 2022, 18, e1083–e1089. [Google Scholar] [CrossRef]
  20. Gao, L.; Yang, L.; Li, X.; Chen, J.; Du, J.; Bai, X.; Yang, X. The Use of a Logistic Regression Model to Develop a Risk Assessment of Intraoperatively Acquired Pressure Ulcer. J. Clin. Nurs. 2018, 27, 2984–2992. [Google Scholar] [CrossRef]
  21. Hyun, S.; Moffatt-Bruce, S.; Cooper, C.; Hixon, B.; Kaewprag, P. Prediction Model for Hospital-Acquired Pressure Ulcer Development: Retrospective Cohort Study. JMIR Med. Inform. 2019, 7, e13785. [Google Scholar] [CrossRef]
  22. Jin, Y.; Jin, T.; Lee, S.M. Automated Pressure Injury Risk Assessment System Incorporated into an Electronic Health Record System. Nurs. Res. 2017, 66, 462–472. [Google Scholar] [CrossRef]
  23. Kaewprag, P.; Newton, C.; Vermillion, B.; Hyun, S.; Huang, K.; Machiraju, R. Predictive Models for Pressure Ulcers from Intensive Care Unit Electronic Health Records Using Bayesian Networks. BMC Med. Inform. Decis. Mak. 2017, 17, 81–91. [Google Scholar] [CrossRef] [PubMed]
  24. Ladios-Martin, M.; Fernández-De-maya, J.; Ballesta-López, F.J.; Belso-Garzas, A.; Mas-Asencio, M.; Cabañero-Martínez, M.J. Predictive Modeling of Pressure Injury Risk in Patients Admitted to an Intensive Care Unit. Am. J. Crit. Care 2020, 29, e70–e80. [Google Scholar] [CrossRef]
  25. Li, H.L.; Lin, S.W.; Hwang, Y.T. Using Nursing Information and Data Mining to Explore the Factors That Predict Pressure Injuries for Patients at the End of Life. CIN—Comput. Inform. Nurs. 2019, 37, 133–140. [Google Scholar] [CrossRef] [PubMed]
  26. Nakagami, G.; Yokota, S.; Kitamura, A.; Takahashi, T.; Morita, K.; Noguchi, H.; Ohe, K.; Sanada, H. Supervised Machine Learning-Based Prediction for in-Hospital Pressure Injury Development Using Electronic Health Records: A Retrospective Observational Cohort Study in a University Hospital in Japan. Int. J. Nurs. Stud. 2021, 119, 103932. [Google Scholar] [CrossRef]
  27. Ossai, C.I.; O’Connor, L.; Wickramasighe, N. Real-Time Inpatients Risk Profiling in Acute Care: A Comparative Study of Falls and Pressure Injuries Vulnerabilities; University of Maribor: Maribor, Slovenia, 2021; pp. 35–50. [Google Scholar]
  28. Šín, P.; Hokynková, A.; Marie, N.; Andrea, P.; Krč, R.; Podroužek, J. Machine Learning-Based Pressure Ulcer Prediction in Modular Critical Care Data. Diagnostics 2022, 12, 850. [Google Scholar] [CrossRef] [PubMed]
  29. Song, J.; Gao, Y.; Yin, P.; Li, Y.; Li, Y.; Zhang, J.; Su, Q.; Fu, X.; Pi, H. The Random Forest Model Has the Best Accuracy among the Four Pressure Ulcer Prediction Models Using Machine Learning Algorithms. Risk Manag. Healthc. Policy 2021, 14, 1175–1187. [Google Scholar] [CrossRef] [PubMed]
  30. Su, C.T.; Wang, P.C.; Chen, Y.C.; Chen, L.F. Data Mining Techniques for Assisting the Diagnosis of Pressure Ulcer Development in Surgical Patients. J. Med. Syst. 2012, 36, 2387–2399. [Google Scholar] [CrossRef]
  31. Vyas, K.; Samadani, A.; Milosevic, M.; Ostadabbas, S.; Parvaneh, S. Additional Value of Augmenting Current Subscales in Braden Scale with Advanced Machine Learning Technique for Pressure Injury Risk Assessment. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine, BIBM 2020, Seoul, Korea, 16–19 December 2020; Institute of Electrical and Electronics Engineers Inc.: New York City, NY, USA, 2020; pp. 2993–2995. [Google Scholar]
  32. Walther, F.; Heinrich, L.; Dresden, T.U.; Schmitt, J.; Roessler, M. Prediction of Inpatient Pressure Ulcers Based on Routine Healthcare Data Using Machine Learning Methodology. Sci. Rep. 2021, 12, 1–10. [Google Scholar] [CrossRef]
  33. Xu, J.; Chen, D.; Deng, X.; Pan, X.; Chen, Y.; Zhuang, X.; Sun, C. Development and Validation of a Machine Learning Algorithm–Based Risk Prediction Model of Pressure Injury in the Intensive Care Unit. Int. Wound J. 2022, 19, 1637–1649. [Google Scholar] [CrossRef]
  34. Lu, J.; Song, E.; Ghoneim, A.; Alrashoud, M. Machine Learning for Assisting Cervical Cancer Diagnosis: An Ensemble Approach. Future Generation Computer Systems. Future Gener. Comput. Syst. 2020, 106, 199–205. [Google Scholar] [CrossRef]
  35. Isabelle, G.; Elisseeff, A. An Introduction to Variable and Feature Selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  36. Dweekat, O.Y.; Lam, S.S.; Alders, V.; Alkhawaldeh, R.; Lu, W.; Wadhawa, T.; Jarrold, K. Addressing Cancer Readmission Prediction Model Drift: A Case Study. In Proceedings of the IISE Annual Conference & Expo 2022, Seattle, WA, USA, 21–24 May 2022; Institute of Industrial & Systems Engineers (IISE): Seattle, WA, USA, 2022; pp. 43–48. [Google Scholar]
  37. Blagus, R.; Lusa, L. SMOTE for High-Dimensional Class-Imbalanced Data. BMC Bioinform. 2013, 14, 1–16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  39. Iranmehr, A.; Masnadi-Shirazi, H.; Vasconcelos, N. Cost-Sensitive Support Vector Machines. Neurocomputing 2019, 343, 50–64. [Google Scholar] [CrossRef] [Green Version]
  40. Guido, R.; Groccia, M.C.; Conforti, D. Hyper-Parameter Optimization in Support Vector Machine on Unbalanced Datasets Using Genetic Algorithms; AIRO Springer Series; Springer: Cham, Switzerland, 2022; Volume 8, pp. 37–47. [Google Scholar] [CrossRef]
  41. Bach, F.; Heckerman, D.; Horvitz, E. Considering Cost Asymmetry in Learning Classifiers. Mach. Learn. Res. 2005, 7, 1713–1741. [Google Scholar]
  42. Davenport, M.A.; Baraniuk, R.G.; Scott, C.D. Controlling False Alarms with Support Vector Machines. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 5. [Google Scholar] [CrossRef] [Green Version]
  43. Lin, Y.; Lee, Y.; Wahba, G. Support Vector Machines for Classification in Nonstandard Situations. Mach. Learn. 2002, 46, 191–202. [Google Scholar] [CrossRef] [Green Version]
  44. Kim, K.H.; Sohn, S.Y. Hybrid Neural Network with Cost-Sensitive Support Vector Machine for Class-Imbalanced Multimodal Data. Neural Netw. 2020, 130, 176–184. [Google Scholar] [CrossRef] [PubMed]
  45. Ma, Y.; Zhao, K.; Wang, Q.; Tian, Y. Incremental Cost-Sensitive Support Vector Machine with Linear-Exponential Loss. IEEE Access 2020, 8, 149899–149914. [Google Scholar] [CrossRef]
  46. Cheng, F.; Zhang, J.; Wen, C. Cost-Sensitive Large Margin Distribution Machine for Classification of Imbalanced Data. Pattern Recognit. Lett. 2016, 80, 107–112. [Google Scholar] [CrossRef]
  47. Tao, X.; Li, Q.; Guo, W.; Ren, C.; Li, C.; Liu, R.; Zou, J. Self-Adaptive Cost Weights-Based Support Vector Machine Cost-Sensitive Ensemble for Imbalanced Data Classification. Inf. Sci. 2019, 487, 31–56. [Google Scholar] [CrossRef]
  48. Akbani, R.; Kwek, S.; Japkowicz, N. Applying Support Vector Machines to Imbalanced Datasets; Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science); Springer: Berlin/Heidelberg, Germany, 2004; Volume 3201, pp. 39–50. [Google Scholar] [CrossRef]
  49. Yang, X.; Song, Q.; Cao, A. Weighted Support Vector Machine for Data Classification. In Proceedings of the International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005; Volume 2, pp. 859–864. [Google Scholar] [CrossRef]
  50. Laref, R.; Losson, E.; Sava, A.; Siadat, M. On the Optimization of the Support Vector Machine Regression Hyperparameters Setting for Gas Sensors Array Applications. Chemom. Intell. Lab. Syst. 2019, 184, 22–27. [Google Scholar] [CrossRef]
  51. Xu, L.; Hou, L.; Zhu, Z.; Li, Y.; Liu, J.; Lei, T.; Wu, X. Mid-Term Prediction of Electrical Energy Consumption for Crude Oil Pipelines Using a Hybrid Algorithm of Support Vector Machine and Genetic Algorithm. Energy 2021, 222, 119955. [Google Scholar] [CrossRef]
  52. Hunter, R.; Anis, H. Genetic Support Vector Machines as Powerful Tools for the Analysis of Biomedical Raman Spectra. J. Raman Spectrosc. 2018, 49, 1435–1444. [Google Scholar] [CrossRef]
  53. Ji, B.; Xie, F.; Wang, X.; He, S.; Song, D. Investigate Contribution of Multi-Microseismic Data to Rockburst Risk Prediction Using Support Vector Machine with Genetic Algorithm. IEEE Access 2020, 8, 58817–58828. [Google Scholar] [CrossRef]
  54. Ali, L.; Wajahat, I.; Amiri Golilarz, N.; Keshtkar, F.; Bukhari, S.A.C. LDA–GA–SVM: Improved Hepatocellular Carcinoma Prediction through Dimensionality Reduction and Genetically Optimized Support Vector Machine. Neural Comput. Appl. 2020, 33, 2783–2792. [Google Scholar] [CrossRef]
  55. Wicaksono, A.S.; Supianto, A.A. Hyper Parameter Optimization Using Genetic Algorithm on Machine Learning Methods for Online News Popularity Prediction. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 263–267. [Google Scholar] [CrossRef] [Green Version]
  56. Villalobos-Arias, L.; Quesada-López, C.; Guevara-Coto, J.; Martínez, A.; Jenkins, M. Evaluating Hyper-Parameter Tuning Using Random Search in Support Vector Machines for Software Effort Estimation. In Proceedings of the PROMISE 2020—16th ACM International Conference on Predictive Models and Data Analytics in Software Engineering, Co-Located with ESEC/FSE, Virtual, 8–9 November 2020; pp. 31–40. [Google Scholar] [CrossRef]
  57. Al-Zoubi, A.M.; Heidari, A.A.; Habib, M.; Faris, H.; Aljarah, I.; Hassonah, M.A. Salp Chain-Based Optimization of Support Vector Machines and Feature Weighting for Medical Diagnostic Information Systems; Springer: Singapore, 2020; pp. 11–34. [Google Scholar] [CrossRef]
  58. Dweekat, O.Y.; Lam, S.S. Cervical Cancer Diagnosis Using an Integrated System of Principal Component Analysis, Genetic Algorithm, and Multilayer Perceptron. Healthcare 2022, 10, 2002. [Google Scholar] [CrossRef]
  59. Cawley, G.C. Leave-One-out Cross-Validation Based Model Selection Criteria for Weighted LS-SVMs. In Proceedings of the 2006 IEEE International Joint Conference on Neural Network Proceedings, Vancouver, BC, Canada, 16–21 July 2006; pp. 1661–1668. [Google Scholar] [CrossRef]
  60. Syarif, I.; Prugel-Bennett, A.; Wills, G. SVM Parameter Optimization Using Grid Search and Genetic Algorithm to Improve Classification Performance. TELKOMNIKA (Telecommun. Comput. Electron. Control.) 2016, 14, 1502–1509. [Google Scholar] [CrossRef]
  61. Canbek, G.; Temizel, T.T.; Sagiroglu, S.; Baykal, N. Binary Classification Performance Measures/Metrics: A Comprehensive Visualized Roadmap to Gain New Insights. In Proceedings of the 2nd International Conference on Computer Science and Engineering (UBMK), Antalya, Turkey, 5–8 October 2017; pp. 821–826. [Google Scholar] [CrossRef]
  62. Toney-Butler, T.J.; Thayer, J.M. Nursing Process; StatPearls Publishing: Treasure Island, FL, USA, 2022; ISBN 9781496309945. [Google Scholar]
  63. Wadhera, R.K.; Choi, E.; Shen, C.; Yeh, R.W.; Joynt Maddox, K.E. Trends, Causes, and Outcomes of Hospitalizations for Homeless Individuals. Med. Care 2019, 57, 21–27. [Google Scholar] [CrossRef]
  64. Lavery, A.M.; Preston, L.E.; Ko, J.Y.; Chevinsky, J.R.; DeSisto, C.L.; Pennington, A.F.; Kompaniyets, L.; Datta, S.D.; Click, E.S.; Golden, T.; et al. Characteristics of Hospitalized COVID-19 Patients Discharged and Experiencing Same-Hospital Readmission—United States, March–August 2020. Morb. Mortal. Wkly. Rep. 2022, 69, 1695–1699. [Google Scholar] [CrossRef]
  65. Ghosh, A.K.; Geisler, B.P.; Ibrahim, S. Racial/Ethnic and Socioeconomic Variations in Hospital Length of Stay: A State-Based Analysis. Medicine 2021, 100, e25976. [Google Scholar] [CrossRef] [PubMed]
  66. Dweekat, O.Y.; Lam, S.S.; McGrath, L. A Hybrid System of Braden Scale and Machine Learning to Predict Hospital-Acquired Pressure Injuries (Bedsores): A Retrospective Observational Cohort Study. Diagnostics 2023, 13, 31. [Google Scholar] [CrossRef]
  67. Gawade, S.M.; Lakhani, R.; Patil, Y. A Descriptive Study to Measure the Reliability of Braden Scale Score Calculated by Clinical Nurses and Evaluate Its Predictive Value for Pressure Ulcer Risk among ICU Patients. J. Pharm. Res. Int. 2021, 33, 57–63. [Google Scholar] [CrossRef]
Figure 1. Different locations of HAPI [8].
Figure 1. Different locations of HAPI [8].
Ijerph 20 00828 g001
Figure 2. The framework of the proposed approach.
Figure 2. The framework of the proposed approach.
Ijerph 20 00828 g002
Figure 3. SVM algorithm.
Figure 3. SVM algorithm.
Ijerph 20 00828 g003
Figure 4. Top 20 features for phase 1 and phase 2.
Figure 4. Top 20 features for phase 1 and phase 2.
Ijerph 20 00828 g004
Figure 5. Comparison between GA-CS-SVM and SVM.
Figure 5. Comparison between GA-CS-SVM and SVM.
Ijerph 20 00828 g005
Figure 6. Results for GA-CS-SVM vs. other methods in terms of performance metrics (phase 1).
Figure 6. Results for GA-CS-SVM vs. other methods in terms of performance metrics (phase 1).
Ijerph 20 00828 g006
Figure 7. Comparison between GA-CS-SVM and balancing using RO.
Figure 7. Comparison between GA-CS-SVM and balancing using RO.
Ijerph 20 00828 g007
Figure 8. Results for GS-SVM vs. other methods in terms of performance metrics (phase 2).
Figure 8. Results for GS-SVM vs. other methods in terms of performance metrics (phase 2).
Ijerph 20 00828 g008
Figure 9. Implications of the proposed research on the medical team.
Figure 9. Implications of the proposed research on the medical team.
Ijerph 20 00828 g009
Table 1. Dataset distribution in phases 1 and 2.
Table 1. Dataset distribution in phases 1 and 2.
Phase 1
(n = 15,889)
TargetIf HAPI Developed or Not
Distribution of Phase 1Non-HAPIHAPI
Number of Patients15,404 (97%)485 (3%)
Phase 2
(n = 485)
TargetWhen HAPI developed for patients with HAPI
Distribution of Phase 20–7 days (High-Risk)>7 days (Medium Risk)
Number of Patients136 patients (28%)349 patients (72%)
Table 2. Confusion matrix.
Table 2. Confusion matrix.
Predicted HAPI
Non-HAPI (0)HAPI (1)
Actual HAPINon-HAPI (0)TNFP
HAPI (1)FNTP
Table 3. Phase 1 results (training with 10-fold cross-validation and testing).
Table 3. Phase 1 results (training with 10-fold cross-validation and testing).
Models80% Training (10-Fold Cross-Validation)20% Testing
SensitivityAUCG-meanFPRSensitivityAUCG-MeanFPR
Proposed
Approach
GA-CS-SVMMean74.0675.6775.6522.7174.2975.7975.7322.71
* CL0.450.200.210.151.230.580.590.37
Other
Algorithms
SVMMean0.0150.000.100.010.0150.000.100.01
CL0.010.000.200.000.010.000.200.00
LRMean6.5553.1925.450.165.8752.8623.830.15
CL0.330.160.670.010.560.281.170.02
AdaBoostMean8.4054.0528.880.297.4153.5526.880.30
CL0.310.150.540.010.600.301.100.03
LDAMean20.6759.5745.111.5321.0159.7445.301.54
CL0.250.130.270.021.020.511.120.06
KNNMean1.4950.6811.870.131.6650.7711.120.13
CL0.170.080.780.010.410.211.800.02
DTMean16.3356.3839.623.5716.2556.3339.383.60
CL0.490.250.600.050.890.451.110.12
RFMean1.6850.7912.710.091.6350.7811.300.08
CL0.180.090.710.010.350.181.650.01
MLPMean54.7254.7273.411.5010.8354.6732.191.48
CL0.200.200.130.040.950.451.520.11
Balancing MethodsRO Mean93.4389.0088.8715.4361.5773.0168.5615.55
CL0.360.780.791.275.142.145.631.84
SMOTE Mean96.7698.3898.360.010.1050.051.020.01
CL0.020.010.010.000.920.040.840.00
* CL = confidence level at 95%.
Table 4. Statistical tests performed on phase 1.
Table 4. Statistical tests performed on phase 1.
PerformanceANOVAt-test
All Methods (Figure 6)GA-CS-SVM vs. SVM (Figure 5)GA-CS-SVM vs. RO (Figure 7)
F-Valuep-ValueT-Statisticp-ValueT-Statisticp-Value
Sensitivity811.490.00−117.410.004.650.00
AUC628.320.00−85.720.002.630.01
G-mean486.170.00−234.730.007.170.02
FPR1302.700.00−119.260.0010.380.00
Table 5. Phase 2 results (training and LOOCV).
Table 5. Phase 2 results (training and LOOCV).
ModelsTrainingLOOCV
SensitivityAUCG-MeanFPRSensitivityAUCG-MeanFPR
Proposed ApproachGS-SVM80.0079.2979.2821.4375.5675.0675.0625.43
Other
Algorithms
SVM53.3369.1067.2715.1459.2672.4971.2714.29
LR54.8173.5571.127.7144.4466.7962.9410.86
AdaBoost54.0772.6170.208.8642.2264.5460.5613.14
LDA56.3072.7270.8410.8649.6368.3965.7612.86
KNN51.1173.4169.944.2935.5663.2156.849.14
DT100.00100.00100.000.0045.1963.4560.7618.3
RF100.00100.00100.000.0045.9369.3965.397.10
MLP82.9689.7789.5117.0444.4465.5162.0313.43
Balancing MethodsRO100.0095.1495.059.7194.2986.0085.6022.29
SMOTE92.0093.1493.145.7186.0084.7984.5616.86
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dweekat, O.Y.; Lam, S.S.; McGrath, L. An Integrated System of Multifaceted Machine Learning Models to Predict If and When Hospital-Acquired Pressure Injuries (Bedsores) Occur. Int. J. Environ. Res. Public Health 2023, 20, 828. https://doi.org/10.3390/ijerph20010828

AMA Style

Dweekat OY, Lam SS, McGrath L. An Integrated System of Multifaceted Machine Learning Models to Predict If and When Hospital-Acquired Pressure Injuries (Bedsores) Occur. International Journal of Environmental Research and Public Health. 2023; 20(1):828. https://doi.org/10.3390/ijerph20010828

Chicago/Turabian Style

Dweekat, Odai Y., Sarah S. Lam, and Lindsay McGrath. 2023. "An Integrated System of Multifaceted Machine Learning Models to Predict If and When Hospital-Acquired Pressure Injuries (Bedsores) Occur" International Journal of Environmental Research and Public Health 20, no. 1: 828. https://doi.org/10.3390/ijerph20010828

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop