Author Contributions
Conceptualization, S.L. and J.J.; methodology, S.L.; software, S.L.; validation, S.L. and J.J.; formal analysis, S.L.; investigation, S.L.; resources, J.J.; data curation, S.L.; writing—original draft preparation, S.L.; writing—review and editing, J.J.; visualization, S.L.; supervision, J.J.; project administration, J.J.; funding acquisition, J.J. All authors have read and agreed to the published version of the manuscript.
Figure 1.
The structure of SSA algorithm, (a) shows embedding process block, (b) shows SVD process block, (c) shows grouping process block and (d) shows averaging process block.
Figure 1.
The structure of SSA algorithm, (a) shows embedding process block, (b) shows SVD process block, (c) shows grouping process block and (d) shows averaging process block.
Figure 2.
The architecture of the vanilla transformer.
Figure 2.
The architecture of the vanilla transformer.
Figure 3.
The framework of SSA-Sl transformer.
Figure 3.
The framework of SSA-Sl transformer.
Figure 4.
The architecture of the Sl transformer.
Figure 4.
The architecture of the Sl transformer.
Figure 5.
The block diagram of the proposed research process.
Figure 5.
The block diagram of the proposed research process.
Figure 6.
(a) The bearing simulator of CWRU and (b) its cross-section view.
Figure 6.
(a) The bearing simulator of CWRU and (b) its cross-section view.
Figure 7.
Components of rolling bearing.
Figure 7.
Components of rolling bearing.
Figure 8.
Schematic of an electric discharge machining (EDM) machine tool.
Figure 8.
Schematic of an electric discharge machining (EDM) machine tool.
Figure 9.
Data comparison diagram, original data (inner race), noise, noise with original data, and reconstructed data from above.
Figure 9.
Data comparison diagram, original data (inner race), noise, noise with original data, and reconstructed data from above.
Figure 10.
Figure shows how SSA decomposes noise-mixed data. The top right, it shows the first 11 elements, the bottom left—maintaining 339 components, and the bottom right—the original time series (TS), and the top left shows mixture of components.
Figure 10.
Figure shows how SSA decomposes noise-mixed data. The top right, it shows the first 11 elements, the bottom left—maintaining 339 components, and the bottom right—the original time series (TS), and the top left shows mixture of components.
Figure 11.
The result of SSA when L is 20, the middle shows first 11 components, the right shows remaining 339 components, the left shows mixture of components.
Figure 11.
The result of SSA when L is 20, the middle shows first 11 components, the right shows remaining 339 components, the left shows mixture of components.
Figure 12.
The result of SSA when L is 100, the middle shows first 11 components, the right shows remaining 339 components, the left shows mixture of components.
Figure 12.
The result of SSA when L is 100, the middle shows first 11 components, the right shows remaining 339 components, the left shows mixture of components.
Figure 13.
Comparison of GeLU, ReLU, and swish.
Figure 13.
Comparison of GeLU, ReLU, and swish.
Figure 14.
The loss graph of Sl transformer. The x-axis describes epochs and the y-axis describes loss.
Figure 14.
The loss graph of Sl transformer. The x-axis describes epochs and the y-axis describes loss.
Figure 15.
This accuracy graph of Sl transformer. The x-axis describes epochs and the y-axis describes accuracy.
Figure 15.
This accuracy graph of Sl transformer. The x-axis describes epochs and the y-axis describes accuracy.
Figure 16.
The accuracy graph of non-applying SSA algorithm for noise data.
Figure 16.
The accuracy graph of non-applying SSA algorithm for noise data.
Figure 17.
The accuracy graph of applying SSA algorithm for noise data.
Figure 17.
The accuracy graph of applying SSA algorithm for noise data.
Table 1.
The optimization structure and hyperparameters for the proposed Sl transformer.
Table 1.
The optimization structure and hyperparameters for the proposed Sl transformer.
Hyperparameter | Value |
---|
Input size | [224, 224, 3] |
Batch size | 128 |
Max epochs | 500 |
Learning rate | 5 × 10 |
Label smoothing rate | 0.1 |
Number of encoder layers N | 6 |
Hidden dimension | 256 |
Optimizer | Adam |
Dropout rate | 0.2 |
Position encoding | 1D |
Table 2.
The table of accuracy of the Sl transformer and compared model.
Table 2.
The table of accuracy of the Sl transformer and compared model.
Model | SVM | CNN | LSTM | Transformer | Sl Transformer |
---|
Accuracy | 64.32 | 85.69 | 90.15 | 93.47 | 95.54 |
F1-score | 57.44 | 84.44 | 89.72 | 92.31 | 94.47 |
Table 3.
The accuracy table of non-applying SSA algorithm for noise data.
Table 3.
The accuracy table of non-applying SSA algorithm for noise data.
Model | −4 dB | −2 dB | 0 dB | 2 dB | 4 dB |
---|
SVM | 55.33 | 57.43 | 59.62 | 59.63 | 59.44 |
CNN | 60.32 | 62.33 | 64.45 | 65.43 | 67.32 |
LSTM | 63.79 | 65.44 | 67.59 | 68.32 | 70.10 |
Transformer | 65.72 | 70.50 | 73.44 | 75.62 | 77.66 |
Sl transformer | 69.89 | 75.55 | 77.44 | 77.48 | 78.72 |
Table 4.
The F1-score table of non-applying SSA algorithm for noise data.
Table 4.
The F1-score table of non-applying SSA algorithm for noise data.
Model | −4 dB | −2 dB | 0 dB | 2 dB | 4 dB |
---|
SVM | 54.79 | 55.44 | 58.41 | 58.01 | 58.88 |
CNN | 59.44 | 61.15 | 61.90 | 63.57 | 64.23 |
LSTM | 61.55 | 64.31 | 65.11 | 66.15 | 69.77 |
Transformer | 64.16 | 68.51 | 72.45 | 74.10 | 76.09 |
Sl transformer | 68.13 | 74.10 | 76.36 | 75.66 | 77.77 |
Table 5.
The accuracy table of applying SSA algorithm.
Table 5.
The accuracy table of applying SSA algorithm.
Model | −4 dB | −2 dB | 0 dB | 2 dB | 4 dB |
---|
SVM | 65.30 | 75.65 | 78.79 | 80.75 | 82.33 |
CNN | 79.25 | 80.95 | 82.55 | 85.88 | 86.75 |
LSTM | 81.10 | 85.97 | 90.15 | 91.25 | 92.33 |
Transformer | 83.55 | 87.42 | 92.35 | 93.75 | 94.21 |
Sl transformer | 85.55 | 89.95 | 92.44 | 95.76 | 96.44 |
Table 6.
The F1-score table of applying SSA algorithm.
Table 6.
The F1-score table of applying SSA algorithm.
Model | −4 dB | −2 dB | 0 dB | 2 dB | 4 dB |
---|
SVM | 64.24 | 74.10 | 76.42 | 77.10 | 79.17 |
CNN | 78.15 | 79.95 | 81.34 | 83.11 | 85.15 |
LSTM | 80.15 | 82.44 | 85.57 | 90.15 | 91.99 |
Transformer | 82.11 | 85.17 | 88.44 | 92.28 | 93.12 |
Sl transformer | 84.00 | 86.37 | 92.15 | 94.35 | 95.04 |
Table 7.
Before and after comparison with SSA-Sl transformer introduction scenario.
Table 7.
Before and after comparison with SSA-Sl transformer introduction scenario.
Before | After |
---|
- Bearing data are retrieved from the sensor for bearing fault detection, but noise occurs in the data. | - Accept situations in which noise is generated rather than attempts to block the noise at the source. |
- Remove the generated noise, and use denoise autoencoder as a representative method. As the neural network for classify and the neural network for denoise work together, real-time inspection of defects becomes difficult. | - By performing SSA preprocessing on the generated noise, the data mixed with noise are decomposed and reconstructed, so that a large number of noises are separated from the data. |
- If noise is not removed, the performance of the artificial intelligence model will drop sharply due to noise generation, and deep learning projects will fail. | - Solves the classify problem by using a robust Sl transformer model on time series data on noise-free data. |