Author Contributions
Conceptualization, H.L. and H.G.; methodology, Y.K. and M.C.; software, Y.K. and M.C.; validation, N.C., H.G. and H.L.; formal analysis, Y.K.; resources, H.G. and N.C.; data curation, N.C.; writing—original draft preparation, Y.K. and M.C.; writing—review and editing, H.L. and H.G.; visualization, Y.K. and M.C.; supervision, H.L.; project administration, H.L. and H.G.; funding acquisition, H.L. and H.G. All authors have read and agreed to the published version of the manuscript.
Figure 1.
Patient selection, data preprocessing, and dataset. The training dataset number was 7291, and the validation set number was 1695, and the holdout fold number was 2421, and the test dataset number was 3458.
Figure 1.
Patient selection, data preprocessing, and dataset. The training dataset number was 7291, and the validation set number was 1695, and the holdout fold number was 2421, and the test dataset number was 3458.
Figure 2.
Processing by the recent data imputer (RDI). The black nodes indicate measured values and the white nodes indicate missing values.
Figure 2.
Processing by the recent data imputer (RDI). The black nodes indicate measured values and the white nodes indicate missing values.
Figure 3.
Processing by the MissForest. (1) Interpolate missing values into the median. (2) Train via dataset. (3) Interpolate missing values into the prediction results.
Figure 3.
Processing by the MissForest. (1) Interpolate missing values into the median. (2) Train via dataset. (3) Interpolate missing values into the prediction results.
Figure 4.
Processing by sliding window.
Figure 4.
Processing by sliding window.
Figure 5.
SMOTE processing. Blue indicates the majority dataset (non-respiratory failure). Orange indicates the minority dataset (respiratory failure). Green indicates the minority dataset (respiratory failure), which is generated by SMOTE.
Figure 5.
SMOTE processing. Blue indicates the majority dataset (non-respiratory failure). Orange indicates the minority dataset (respiratory failure). Green indicates the minority dataset (respiratory failure), which is generated by SMOTE.
Figure 6.
Borderline-SMOTE processing. Blue indicates the majority dataset (non-respiratory failure). Orange indicates the minority dataset (respiratory failure). Green indicates the minority dataset (respiratory failure) generated by Borderline-SMOTE.
Figure 6.
Borderline-SMOTE processing. Blue indicates the majority dataset (non-respiratory failure). Orange indicates the minority dataset (respiratory failure). Green indicates the minority dataset (respiratory failure) generated by Borderline-SMOTE.
Figure 7.
ADASYN processing. Blue indicates the majority dataset (non-respiratory failure). Orange indicates the minority dataset (respiratory failure). Green indicates the minority dataset (respiratory failure) generated by ADASYN. When there are more majority data surrounding minority data, more minority data are produced.
Figure 7.
ADASYN processing. Blue indicates the majority dataset (non-respiratory failure). Orange indicates the minority dataset (respiratory failure). Green indicates the minority dataset (respiratory failure) generated by ADASYN. When there are more majority data surrounding minority data, more minority data are produced.
Figure 8.
Decision tree. Orange indicates the prediction of respiratory failure in each node. Blue indicates non-respiratory failure. Darker nodes contain more data.
Figure 8.
Decision tree. Orange indicates the prediction of respiratory failure in each node. Blue indicates non-respiratory failure. Darker nodes contain more data.
Figure 9.
Processing of random forests. The red box in the decision tree indicates the predicted outcome. The random forest classifier is based on decision tree outcomes. Random Forest performs bootstrap aggregation on many decision trees and uses voting to determine the prediction results of many decision trees.
Figure 9.
Processing of random forests. The red box in the decision tree indicates the predicted outcome. The random forest classifier is based on decision tree outcomes. Random Forest performs bootstrap aggregation on many decision trees and uses voting to determine the prediction results of many decision trees.
Figure 10.
It shows the classification of SVM.
Figure 10.
It shows the classification of SVM.
Figure 11.
Processing of adaptive boost. Each estimator transfers the weight of an incorrect prediction after training via the dataset.
Figure 11.
Processing of adaptive boost. Each estimator transfers the weight of an incorrect prediction after training via the dataset.
Figure 12.
Structure of MLP in this paper. The number of units in hidden layer is 256.
Figure 12.
Structure of MLP in this paper. The number of units in hidden layer is 256.
Figure 13.
Structure of the RNN model in this paper. The number of units in the RNN layer is 128.
Figure 13.
Structure of the RNN model in this paper. The number of units in the RNN layer is 128.
Figure 14.
Structure of the LSTM model in this paper. The number of units in the LSTM layer is 128.
Figure 14.
Structure of the LSTM model in this paper. The number of units in the LSTM layer is 128.
Figure 15.
Structure of the GRU model in this paper. The number of units in the GRU layer is 128.
Figure 15.
Structure of the GRU model in this paper. The number of units in the GRU layer is 128.
Figure 16.
Stratified k-fold processing (in this case, k is 4).
Figure 16.
Stratified k-fold processing (in this case, k is 4).
Table 1.
The respiratory failure was correlated to p-values of features.
Table 1.
The respiratory failure was correlated to p-values of features.
Features | p-Value | Features | p-Value |
---|
Pesticide dose | 0.000 | WBC | 0.000 |
Sex | 0.000 | PLT | 0.000 |
Age | 0.000 | Albumin | 0.000 |
BMI | 0.023 | Glucose | 0.070 |
Smoking | 0.326 | BUN | 0.622 |
Alcohol | 0.313 | Creatinine | 0.100 |
Diabetes disease | 0.000 | Total CO2 | 0.000 |
Respiratory disease | 0.000 | C-reactive protein 1 | 0.000 |
Cardiovascular disease | 0.995 | pH | 0.000 |
GCS | 0.000 | pCO2 | 0.199 |
SBP max | 0.088 | pO2 | 0.000 |
DBP max | 0.897 | O2 saturation | 0.000 |
HR max | 0.000 | HCO3 standard | 0.356 |
RR max | 0.174 | BE | 0.120 |
BT max | 0.000 | Troponin | 0.555 |
Hb | 0.221 | Lactate | 0.000 |
Table 2.
Performance result of tree-based feature selection method.
Table 2.
Performance result of tree-based feature selection method.
Features | Tree-Based Feature Selection | Features | Tree-Based Feature Selection |
---|
RF | GB | RF | GB |
---|
Pesticide category | 0.056 | 0.073 | PLT | 0.032 | 0.013 |
Pesticide dose | 0.032 | 0.017 | Albumin | 0.034 | 0.008 |
Sex | 0.001 | 0.0003 | total_CO2 | 0.072 | 0.025 |
Age | 0.039 | 0.019 | C-reactive_protein_1 | 0.053 | 0.057 |
GCS | 0.053 | 0.068 | pH | 0.175 | 0.247 |
SBP_max | 0.010 | 0.003 | pO2 | 0.018 | 0.003 |
HR_max | 0.087 | 0.097 | O2_saturation | 0.062 | 0.017 |
BT_max | 0.033 | 0.029 | Lactate | 0.017 | 0.006 |
WBC | 0.224 | 0.316 | | | |
Table 3.
Performance result of recursive feature elimination of each algorithm. Low-rank features are albumin, platelet, and sex. High-rank feature is pH.
Table 3.
Performance result of recursive feature elimination of each algorithm. Low-rank features are albumin, platelet, and sex. High-rank feature is pH.
Machine Learning Algorithm | Low-Rank Feature | High-Rank Feature |
---|
SVM with linear | Albumin | pH |
LR | PLT | pH |
RF | Sex | pH |
DT | Albumin | pH |
GB | Sex | pH |
Table 4.
Performance result of feature selection based on RF, GB, and MLP.
Table 4.
Performance result of feature selection based on RF, GB, and MLP.
Feature | Algorithm | PPV | Sensitivity | F1 Score | AUC |
---|
Reference | RF | 98.92% | 96.34% | 0.9761 | 0.9812 |
GB | 96.81% | 95.29% | 0.9604 | 0.9751 |
MLP | 96.81% | 95.29% | 0.9604 | 0.9751 |
Reference exclude sex | RF | 98.92% | 95.81% | 0.9734 | 0.9786 |
GB | 92.78% | 94.24% | 0.9351 | 0.9681 |
MLP | 92.78% | 94.24% | 0.9351 | 0.9681 |
Reference exclude PLT | RF | 99.46% | 95.81% | 0.9760 | 0.9788 |
GB | 98.89% | 93.19% | 0.9596 | 0.9655 |
MLP | 92.78% | 94.24% | 0.9351 | 0.9681 |
Reference exclude albumin | RF | 98.39% | 95.81% | 0.9708 | 0.9784 |
GB | 96.70% | 92.15% | 0.9437 | 0.9594 |
MLP | 96.70% | 92.15% | 0.9437 | 0.9594 |
Table 5.
Confusion matrix.
Table 5.
Confusion matrix.
| | Actual |
| | Respiratory failure | Non-respiratory failure |
Precision | Respiratory failure | True positive (TP) | False positive (FP) |
Non-respiratory failure | False negative (FN) | True negative (TN) |
Table 6.
The characteristics of our dataset.
Table 6.
The characteristics of our dataset.
| Training Data (n = 8068) | Test Data (n = 3458) |
---|
Pesticide dose | 171.87 ± 138.07 | 177.18 ± 144.12 |
Sex, male | 4994, 61.90% | 2210, 63.91% |
Age | 61.07 ± 16.37 | 61.40 ± 16.48 |
GCS | 13.90 ± 2.10 | 13.85 ± 2.19 |
1h_SBP_max | 124.00 ± 18.03 | 123.89 ± 18.16 |
1h_HR_max | 76.40 ± 14.37 | 76.30 ± 14.36 |
1h_BT_max | 36.55 ± 0.36 | 36.55 ± 0.36 |
1h_WBC | 7.53 ± 3.17 | 7.54 ± 3.15 |
1h_PLT | 144.52 ± 64.19 | 142.90 ± 65.09 |
1h_albumin | 3.59 ± 0.48 | 3.57 ± 0.49 |
1h_total_CO2 | 24.14 ± 3.15 | 24.23 ± 3.15 |
1h_C-reactive_protein_1 | 26.97 ± 50.49 | 28.73 ± 51.61 |
1h_pH | 7.43 ± 0.06 | 7.43 ± 0.06 |
1h_pO2 | 93.29 ± 26.62 | 93.02 ± 24.74 |
1h_O2_saturation | 96.02 ± 3.67 | 96.03 ± 3.62 |
1h_lactate | 2.43 ± 2.02 | 2.48 ± 2.12 |
2h_SBP_max | 123.92 ± 17.90 | 123.93 ± 18.32 |
2h_HR_max | 76.40 ± 14.38 | 76.38 ± 14.43 |
2h_BT_max | 36.55 ± 0.35 | 36.56 ± 0.37 |
2h_WBC | 7.53 ± 3.17 | 7.54 ± 3.15 |
2h_PLT | 144.12 ± 63.80 | 142.47 ± 64.86 |
2h_albumin | 3.58 ± 0.48 | 3.57 ± 0.49 |
2h_total_CO2 | 24.14 ± 3.15 | 24.23 ± 3.15 |
2h_C-reactive_protein_1 | 26.98 ± 50.49 | 28.74 ±51.61 |
2h_pH | 7.43 ± 0.06 | 7.43 ± 0.07 |
2h_pO2 | 93.41 ± 26.20 | 92.98 ± 24.72 |
2h_O2_saturation | 95.99 ± 3.79 | 96.00 ± 3.72 |
2h_lactate | 2.44 ± 2.03 | 2.48 ± 2.13 |
3h_SBP_max | 123.84 ± 18.01 | 123.79 ± 17.99 |
3h_HR_max | 76.48 ± 14.52 | 76.34 ± 14.36 |
3h_BT_max | 36.56 ± 0.35 | 36.56 ± 0.36 |
3h_WBC | 7.53 ± 3.17 | 7.53 ± 3.15 |
3h_PLT | 143.94 ± 63.64 | 142.42 ± 64.84 |
3h_albumin | 3.58 ± 0.48 | 3.57 ± 0.49 |
3h_total_CO2 | 24.14 ± 3.16 | 24.24 ± 3.15 |
3h_C-reactive_protein_1 | 26.99 ± 50.48 | 28.74 ± 51.61 |
3h_pH | 7.43 ± 0.07 | 7.43 ± 0.07 |
3h_pO2 | 93.44 ± 25.88 | 92.98 ± 24.74 |
3h_O2_saturation | 95.99 ± 3.81 | 95.97 ± 3.88 |
3h_lactate | 2.44 ± 2.04 | 2.49 ± 2.15 |
Table 7.
Performance comparison of KNN and missForest based on RF, GB, and MLP.
Table 7.
Performance comparison of KNN and missForest based on RF, GB, and MLP.
Imputation | Algorithm | PPV | Sensitivity | F1 Score | AUC |
---|
KNN imputer | RF | 98.92% | 96.34% | 0.9761 | 0.9812 |
GB | 97.80% | 93.19% | 0.9544 | 0.9651 |
MLP | 96.81% | 95.29% | 0.9604 | 0.9751 |
MissForest imputer | RF | 98.92% | 96.34% | 0.9761 | 0.9812 |
GB | 97.74% | 90.58% | 0.9402 | 0.9520 |
MLP | 96.17% | 92.15% | 0.9412 | 0.9592 |
Table 8.
Performance comparison of hyperparameter turning.
Table 8.
Performance comparison of hyperparameter turning.
Algorithm | Hyperparameter | PPV | Sensitivity | F1 Score | AUC |
---|
DT | Function of computational complexity | Gini | Max depth | 8 | 95.76% | 82.72% | 0.8876 | 0.9120 |
9 | 96.49% | 86.39% | 0.9116 | 0.9306 |
10 | 98.17% | 84.29% | 0.9070 | 0.9208 |
Entropy | 8 | 97.27% | 93.19% | 0.9519 | 0.9648 |
9 | 98.35% | 93.72% | 0.9598 | 0.9679 |
10 | 98.31% | 91.62% | 0.9485 | 0.9574 |
RF | Function of computational complexity | Gini | Max depth | 8 | 98.66% | 76.96% | 0.8647 | 0.8844 |
9 | 98.76% | 83.25% | 0.9034 | 0.9158 |
10 | 98.74% | 82.20% | 0.8971 | 0.9105 |
Entropy | 8 | 98.86% | 91.10% | 0.9482 | 0.9550 |
9 | 98.92% | 96.34% | 0.9761 | 0.9812 |
10 | 98.91% | 95.29% | 0.9707 | 0.9760 |
SVM | Regularization parameter | 5.5 | kernel | RBF | 99.32% | 76.96% | 0.8673 | 0.8846 |
Linear | 77.36% | 42.92% | 0.5522 | 0.7093 |
Poly | 97.43% | 79.58% | 0.8761 | 0.8970 |
6.0 | RBF | 99.33% | 77.49% | 0.8706 | 0.8872 |
Linear | 76.58% | 44.50% | 0.5629 | 0.7167 |
Poly | 96.84% | 80.10% | 0.8768 | 0.8994 |
6.5 | RBF | 99.34% | 79.06% | 0.8805 | 0.8950 |
Linear | 75.68% | 43.98% | 0.5563 | 0.7138 |
Poly | 86.84% | 80.10% | 0.8768 | 0.8994 |
AB | Number of estimators | 140 | Learning rate | 0.5 | 94.08% | 74.87% | 0.8338 | 0.8723 |
1.0 | 95.43% | 87.43% | 0.9126 | 0.9354 |
2.0 | 6.91% | 84.82% | 0.1277 | 0.4344 |
150 | 0.5 | 94.34% | 78.53% | 0.8571 | 0.8907 |
1.0 | 96.07% | 89.53% | 0.9268 | 0.9461 |
2.0 | 6.91% | 84.82% | 0.1277 | 0.4344 |
160 | 0.5 | 94.97% | 79.06% | 0.8629 | 0.8935 |
1.0 | 94.97% | 89.01% | 0.9189 | 0.9430 |
2.0 | 6.91% | 84.82% | 0.1277 | 0.4344 |
GB | Number of estimators | 110 | Max depth | 3 | 95.73% | 82.99% | 0.8845 | 0.9094 |
4 | 97.77% | 91.62% | 0.9459 | 0.9572 |
5 | 97.80% | 93.19% | 0.9544 | 0.9651 |
120 | 3 | 96.45% | 85.34% | 0.9056 | 0.9254 |
4 | 97.80% | 93.19% | 0.9544 | 0.9651 |
5 | 97.80% | 93.19% | 0.9544 | 0.9651 |
130 | 3 | 96.45% | 85.34% | 0.9056 | 0.9254 |
4 | 97.78% | 92.15% | 0.9488 | 0.9598 |
5 | 96.55% | 87.96% | 0.9205 | 0.9384 |
MLP | Unit size | 64 | Dropout rate | 20% | 97.13% | 88.48% | 0.9260 | 0.9413 |
30% | 96.63% | 90.05% | 0.9322 | 0.9489 |
40% | 95.71% | 81.68% | 0.8814 | 0.9068 |
128 | 20% | 97.71% | 89.53% | 0.9344 | 0.9467 |
30% | 98.24% | 87.43% | 0.9252 | 0.9365 |
40% | 94.97% | 79.06% | 0.8629 | 0.8935 |
256 | 20% | 96.81% | 95.29% | 0.9604 | 0.9751 |
30% | 97.80% | 93.19% | 0.9544 | 0.9651 |
40% | 92.73% | 80.10% | 0.8596 | 0.8798 |
RNN | Unit size | 64 | Dropout rate | 20% | 76.28 | 62.30% | 0.6859 | 0.8032 |
30% | 86.26% | 82.20% | 0.8418 | 0.9054 |
40% | 71.34% | 66.49% | 0.6883 | 0.8210 |
128 | 20% | 98.32% | 92.15% | 0.9514 | 0.9601 |
30% | 97.85% | 95.29% | 0.9655 | 0.9755 |
40% | 78.08% | 59.69% | 0.6766 | 0.7913 |
256 | 20% | 96.65% | 90.58% | 0.9351 | 0.9515 |
30% | 64.88% | 69.63% | 0.6717 | 0.8320 |
40% | 98.22% | 86.91% | 0.9222 | 0.9339 |
LSTM | Unit size | 64 | Dropout rate | 20% | 95.72% | 93.72$ | 0.9471 | 0.9668 |
30% | 98.29% | 90.05% | 0.9399 | 0.9496 |
40% | 72.15% | 59.69% | 0.6533 | 0.7886 |
128 | 20% | 96.15% | 91.62% | 0.9383 | 0.9565 |
30% | 98.29% | 90.05% | 0.9399 | 0.9459 |
40% | 98.88% | 92.67% | 0.9568 | 0.9629 |
256 | 20% | 98.87% | 91.62% | 0.9511 | 0.9577 |
30% | 97.77% | 91.62% | 0.9459 | 0.9572 |
40% | 98.86% | 91.10% | 0.9482 | 0.9550 |
GRU | Unit size | 64 | Dropout rate | 20% | 97.75% | 91.10% | 0.9431 | 0.9546 |
30% | 98.82% | 87.43% | 0.9278 | 0.9367 |
40% | 98.32% | 92.15% | 0.9514 | 0.9601 |
128 | 20% | 98.88% | 92.15% | 0.9539 | 0.9603 |
30% | 97.16% | 89.53% | 0.9319 | 0.9465 |
40% | 97.30% | 94.24% | 0.9574 | 0.9701 |
256 | 20% | 97.78% | 92.15% | 0.9488 | 0.9598 |
30% | 98.31% | 91.10% | 0.9457 | 0.9548 |
40% | 98.34% | 93.19% | 0.9570 | 0.9653 |
Table 9.
Prediction performance. GRU with ADASYN demonstrated the highest performance.
Table 9.
Prediction performance. GRU with ADASYN demonstrated the highest performance.
Machine Learning Algorithm | Oversampling | PPV | Sensitivity | F1 Score | AUC |
---|
LR | SMOTE | 42.42% | 86.39% | 0.5690 | 0.8817 |
Borderline SMOTE | 43.24% | 85.34% | 0.5739 | 0.8787 |
ADASYN | 44.24% | 86.39% | 0.5851 | 0.8853 |
DT | SMOTE | 86.27% | 92.15% | 0.8911 | 0.9545 |
Borderline SMOTE | 92.22% | 93.19% | 0.9271 | 0.9626 |
ADASYN | 86.12% | 94.24% | 0.9000 | 0.9647 |
RF | SMOTE | 96.43% | 98.95% | 0.9767 | 0.9932 |
Borderline SMOTE | 95.43% | 98.43% | 0.9691 | 0.9901 |
ADASYN | 96.43% | 98.90% | 0.9765 | 0.9929 |
SVM | SMOTE | 91.09% | 96.34% | 0.9364 | 0.9776 |
Borderline SMOTE | 90.53% | 90.05% | 0.9029 | 0.9462 |
ADASYN | 90.40% | 93.72% | 0.9203 | 0.9643 |
AB | SMOTE | 83.11% | 95.29% | 0.8878 | 0.9681 |
Borderline SMOTE | 85.57% | 90.05% | 0.8776 | 0.9438 |
ADASYN | 83.49% | 92.67% | 0.8784 | 0.9555 |
GB | SMOTE | 95.96% | 99.48% | 0.9769 | 0.9956 |
Borderline SMOTE | 95.90% | 97.91% | 0.9689 | 0.9877 |
ADASYN | 98.45% | 99.48% | 0.9896 | 0.9967 |
MLP | SMOTE | 98.94% | 97.91% | 0.9842 | 0.9891 |
Borderline SMOTE | 96.84% | 96.34% | 0.9659 | 0.9803 |
ADASYN | 98.38% | 95.29% | 0.9681 | 0.9758 |
RNN | SMOTE | 97.33% | 95.29% | 0.9630 | 0.9753 |
Borderline SMOTE | 98.35% | 93.72% | 0.9598 | 0.9679 |
ADASYN | 96.35% | 96.86% | 0.9661 | 0.9827 |
LSTM | SMOTE | 97.85% | 95.29% | 0.9655 | 0.9755 |
Borderline SMOTE | 98.38% | 95.29% | 0.9681 | 0.9758 |
ADASYN | 98.93% | 96.86% | 0.9788 | 0.9838 |
GRU | SMOTE | 98.3–8% | 95.29% | 0.9681 | 0.9758 |
Borderline SMOTE | 98.38% | 95.29% | 0.9681 | 0.9758 |
ADASYN | 98.42% | 97.91% | 0.9816 | 0.9889 |
Table 10.
Highest prediction performance of each machine learning algorithm.
Table 10.
Highest prediction performance of each machine learning algorithm.
Machine Learning Algorithms | Oversampling | PPV | Sensitivity | F1 Score | AUC |
---|
LR | ADSYN | 44.47% | 83.88% | 0.5812 | 0.8745 |
DT | Borderline SMOTE | 92.11% | 94.14% | 0.9312 | 0.9672 |
RF | SMOTE | 96.09% | 98.90% | 0.9747 | 0.9928 |
SVM | SMOTE | 89.49% | 96.70% | 0.9296 | 0.9786 |
AB | SMOTE | 86.33% | 94.87% | 0.9040 | 0.9679 |
GB | ADASYN | 96.10% | 99.27% | 0.9766 | 0.9946 |
MLP | SMOTE | 97.48% | 99.27% | 0.9837 | 0.9952 |
RNN | ADASYN | 95.67% | 97.07% | 0.9636 | 0.9835 |
LSTM | ADASYN | 98.18% | 98.90% | 0.9854 | 0.9937 |
GRU | ADASYN | 98.53% | 98.17% | 0.9835 | 0.9902 |
Table 11.
Performance result of each scenario.
Table 11.
Performance result of each scenario.
Scenario | Algorithm | PPV | Sensitivity | F1 Score | AUC |
---|
Reference | LR | 44.47% | 83.88% | 0.5812 | 0.8745 |
DT | 92.11% | 94.14% | 0.9312 | 0.9672 |
RF | 96.09% | 98.90% | 0.9747 | 0.9928 |
SVM | 89.49% | 96.70% | 0.9296 | 0.9786 |
AB | 86.33% | 94.87% | 0.9040 | 0.9679 |
GB | 96.10% | 99.27% | 0.9766 | 0.9946 |
MLP | 97.48% | 99.27% | 0.9837 | 0.9952 |
RNN | 95.67% | 97.07% | 0.9636 | 0.9835 |
LSTM | 98.18% | 98.90% | 0.9854 | 0.9937 |
GRU | 98.53% | 98.17% | 0.9835 | 0.9902 |
Replace missing values via average | LR | 22.26% | 70.70% | 0.3386 | 0.7477 |
DT | 63.06% | 83.15% | 0.7172 | 0.8949 |
RF | 53.24% | 87.18% | 0.6611 | 0.9031 |
SVM | 41.05% | 83.15% | 0.5496 | 0.8646 |
AB | 58.21% | 73.99% | 0.6516 | 0.8472 |
GB | 75.40% | 86.45% | 0.8055 | 0.9201 |
MLP | 65.49% | 81.32% | 0.7255 | 0.8882 |
RNN | 62.10% | 78.02% | 0.6916 | 0.8697 |
LSTM | 41.63% | 80.22% | 0.5482 | 0.8529 |
GRU | 65.73% | 85.71% | 0.7440 | 0.9094 |
Does not perform oversampling | LR | 81.46% | 45.05% | 0.5802 | 0.7209 |
DT | 98.05% | 92.31% | 0.9509 | 0.9608 |
RF | 98.85% | 94.14% | 0.9644 | 0.9702 |
SVM | 99.06% | 77.29% | 0.8683 | 0.8861 |
AB | 94.80% | 86.81% | 0.9063 | 0.9320 |
GB | 98.38% | 89.01% | 0.9346 | 0.9444 |
MLP | 95.67% | 80.95% | 0.8770 | 0.9032 |
RNN | 90.87% | 83.88% | 0.8724 | 0.9158 |
LSTM | 77.73% | 62.64% | 0.6937 | 0.8055 |
GRU | 91.09% | 86.08% | 0.8851 | 0.9268 |
Table 12.
Comparison of the performance between the proposed algorithm and algorithms in other studies.
Table 12.
Comparison of the performance between the proposed algorithm and algorithms in other studies.
Algorithms | Features | Patient Data Range | Sensitivity | PPV | AUC |
---|
[4] | Semi-supervised learning | 25 | 32 h | 0.78 | 0.023 | 0.78 |
[5] | XGBoost | 24 | - | 0.71 | - | - |
[6] | LR | 26 | - | 0.72 | 0.74 | 0.89 |
[7] | LightGBM | 25 | - | 0.80 | - | 0.746 |
[8] | GradientBoosting | 106 | 6 h | 0.534 | 0.643 | 0.769 |
[9] | LR | 7 | - | 0.606 | 0.833 | 0.912 |
[10] | LSTM | 8 | 2 h | 0.881 | 0.226 | 0.886 |
Our algorithm | LSTM | 17 | 3 h | 0.9817 | 0.9890 | 0.9937 |