Research on the Prediction Method for the Wear State of Aerospace Titanium Alloy Cutting Tools Based on Knowledge Distillation
Abstract
:1. Introduction
2. Tool Wear State Prediction Encoder–Decoder Structure
2.1. Tool Wear State Prediction Knowledge Distillation Framework
2.2. Knowledge Distillation Model Construction
- (1)
- Construction of the Knowledge Distillation Teacher Model
- (2)
- Construction of the Knowledge Distillation Student Model
2.3. Definition of the Knowledge Distillation Loss Function
- (1)
- Euclidean Distance Loss Function
- (2)
- KL Divergence Loss Function
- (3)
- Cross-Entropy Loss Function
- (4)
- Combined Loss Function
3. Experiments
3.1. TC18 Titanium Alloy Machining Tool Wear Experiment
- (1)
- Tool Cutting Experiment and Signal Acquisition
- (2)
- Data Feature Extraction and Partitioning Strategy
3.2. Evaluation Method
3.3. Knowledge Distillation Experiment and Analysis
- (1)
- Model Hyperparameters and Experimental Environment
- (2)
- Knowledge Distillation Experiment
4. Comparative Experiments
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Li, S.; Li, M.; Gao, Y. Deep Learning Tool Wear State Identification Method Based on Cutting Force Signal. Sensors 2025, 25, 662. [Google Scholar] [CrossRef] [PubMed]
- Wang, K.; Wang, A.; Wu, L. Research on Tool Wear Monitoring Technology Based on Variational Mode Decomposition and Back Propagation Neural Network Model. Sensors 2024, 24, 8107. [Google Scholar] [CrossRef] [PubMed]
- Wang, C.; Wang, G.; Wang, T.; Xiong, X.; Ouyang, Z.; Gong, T. Exploring the Processing Paradigm of Input Data for End-to-End Deep Learning in Tool Condition Monitoring. Sensors 2024, 24, 5300. [Google Scholar] [CrossRef] [PubMed]
- Zhao, R.; Yan, R.; Wang, J.; Mao, K. Learning to monitor machine health with convolutional Bi-Directional LSTM networks. Sensors 2017, 17, 273. [Google Scholar] [CrossRef] [PubMed]
- Li, Y.; Wang, J.; Huang, Z.; Gao, R.X. Physics-informed meta learning for machining tool wear prediction. J. Manuf. Syst. 2022, 62, 17–27. [Google Scholar] [CrossRef]
- Qiao, H.; Wang, T.; Wang, P. A tool wear monitoring and prediction system based on multiscale deep learning models and fog computing. Int. J. Adv. Manuf. Technol. 2020, 108, 2367–2384. [Google Scholar] [CrossRef]
- Xu, X.; Wang, J.; Zhong, B.; Ming, W.; Chen, M. Deep learning-based tool wear prediction and its application for machining process using multi-scale feature fusion and channel attention mechanism. Measurement 2021, 177, 109254. [Google Scholar] [CrossRef]
- Liu, X.; Liu, S.; Li, X.; Zhang, B.; Yue, C.; Liang, S.Y. Intelligent tool wear monitoring based on parallel residual and stacked bidirectional long short-term memory network. J. Manuf. Syst. 2021, 60, 608–619. [Google Scholar] [CrossRef]
- Cai, W.; Zhang, W.; Hu, X.; Liu, Y. A hybrid information model based on long short-term memory network for tool condition monitoring. J. Intell. Manuf. 2020, 31, 1497–1510. [Google Scholar] [CrossRef]
- Karandikar, J.; Schmitz, T.; Smith, S. Physics-guided logistic classification for tool life modeling and process parameter optimization in machining. J. Manuf. Syst. 2021, 59, 522–534. [Google Scholar] [CrossRef]
- Sun, S.; Cheng, Y.; Gan, Z. Patient Knowledge Distillation for BERT Model Compression. In Proceedings of the Conference on Empirical Methods in Natural Language Processing, EMNLP-IJCNLP, Hong Kong, China, 3–7 November 2019; pp. 4323–4332. [Google Scholar]
- Gumma, V.; Dabre, R.; Kumar, P. An Empirical Study of Leveraging Knowledge Distillation for Compressing Multilingual Neural Machine Translation Models. In Proceedings of the 24th Annual Conference of the European Association for Machine Translation, Tampere, Finland, 12–15 June 2023; European Association for Machine Translation: Tampere, Finland, 2023; pp. 103–114. [Google Scholar]
- Zhu, F.; Chen, Y. A Knowledge Distillation Network Combining Adversarial Training and Intermediate Feature Extraction for Lane Line Detection. In Proceedings of the 2024 Australian & New Zealand Control Conference (ANZCC), Gold Coast, Australia, 1–2 February 2024; pp. 92–97. [Google Scholar]
- Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
- Yim, J.; Joo, D.; Bae, J.; Kim, J. A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4133–4141. [Google Scholar]
- Liu, J.; Wen, D.; Gao, H.; Tao, W.; Chen, T.W.; Osa, K.; Kato, M. Knowledge representing: Efficient, sparse representation of prior knowledge for knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Ahn, S.; Hu, S.X.; Damianou, A.; Lawrence, N.D.; Dai, Z. Variational information distillation for knowledge transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9163–9171. [Google Scholar]
- Heo, B.; Lee, M.; Yun, S.; Choi, J.Y. Knowledge transfer via distillation of activation boundaries formed by hidden neurons. Proc. AAAI Conf. Artif. Intell. 2019, 33, 3779–3787. [Google Scholar] [CrossRef]
- Yu, L.; Yazici, V.O.; Liu, X.; Weijer, J.V.D.; Cheng, Y.; Ramisa, A. Learning metrics from teachers: Compact networks for image embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2907–2916. [Google Scholar]
- Chen, G.; Choi, W.; Yu, X.; Han, T.; Chandraker, M. Learning efficient object detection models with knowledge distillation. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 742–751. [Google Scholar]
- Kim, J.; Park, S.U.; Kwak, N. Paraphrasing complex network: Network compression via factor transfer. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 2765–2774. [Google Scholar]
- Mirzadeh, S.I.; Farajtabar, M.; Li, A.; Levine, N.; Matsukawa, A.; Ghasemzadeh, H. Improved knowledge distillation via teacher assistant. Proc. AAAI Conf. Artif. Intell. 2020, 34, 5191–5198. [Google Scholar] [CrossRef]
- Romero, A.; Ballas, N.; Kahou, S.E.; Chassang, A.; Gatta, C.; Bengio, Y. Fitnets: Hints for thin deep nets. arXiv 2014, arXiv:1412.6550. [Google Scholar]
- Zhao, B.; Cui, Q.; Song, R.; Qiu, Y.; Liang, J. Decoupled knowledge distillation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11953–11962. [Google Scholar]
- Xie, Y.; Gao, S.; Zhang, C.; Liu, J. Tool wear state recognition and prediction method based on laplacian eigenmap with ensemble learning model. Adv. Eng. Inform. 2024, 60, 102382. [Google Scholar] [CrossRef]
- Li, R.; Wei, P.; Liu, X.; Li, C.; Ni, J.; Zhao, W.; Zhao, L.; Hou, K. Cutting tool wear state recognition based on a channel-space attention mechanism. J. Manuf. Syst. 2023, 69, 135–149. [Google Scholar] [CrossRef]
- Cheng, M.; Jiao, L.; Yan, P.; Jiang, H.; Wang, R.; Qiu, T.; Wang, X. Intelligent tool wear monitoring and multi-step prediction based on deep learning model. J. Manuf. Syst. 2022, 62, 286–300. [Google Scholar] [CrossRef]
Number of Encoder Wins | Number of Decoder Wins | ||
---|---|---|---|
Bi-GRU | 465 | GRU | 359 |
GRU | 426 | Bi-GRU | 314 |
RNN | 384 | Bi-LSTM | 302 |
Bi-RNN | 372 | FC | 299 |
Bi-LSTM | 328 | RNN | 290 |
LSTM | 279 | LSTM | 286 |
CNN | 214 | Bi-RNN | 281 |
Transformer | 35 | CNN | 198 |
Transformer | 174 |
Serial Number | Spindle Speed (r/min) | Feed Speed (mm/min) | Cutting Depth (mm) | Cutting Width (mm) |
---|---|---|---|---|
1 | 1000 | 240 | 0.6 | 4 |
2 | 1000 | 360 | 0.8 | 5 |
3 | 1000 | 480 | 1.0 | 6 |
4 | 1500 | 240 | 0.8 | 5 |
5 | 1500 | 360 | 0.6 | 6 |
6 | 1500 | 480 | 1.0 | 4 |
7 | 2000 | 240 | 1.0 | 5 |
8 | 2000 | 360 | 0.6 | 6 |
9 | 2000 | 480 | 0.8 | 4 |
Category | Multi-Domain Features | ||||
---|---|---|---|---|---|
Time domain | Absolute mean | Peak value | Root mean square | Root amplitude | Skewness |
Kurtosis | Waveform factor | Pulse factor | Kurtosis factor | Peak factor | |
Margin factor | |||||
Frequency domain | Centroid frequency | Mean square frequency | Root mean square frequency | Frequency variance | |
Time–frequency domain | Number of bases in wavelet packet decomposition: 8 |
Name | Configuration |
---|---|
CPU | Intel(R) Xeon(R) Gold 6230 |
GPU | NVIDIA GeForce RTX 3090Ti |
Operating system | Ubuntu 18.04 |
Python | 3.8 |
Platform | Pytorch 1.6 |
CUDA | 10.1 |
Teacher | Student | P | R | F1 |
---|---|---|---|---|
BiGRU-GRU (T) | Transformer-GRU (S) | 55.46% ↓ | 54.99% ↓ | 54.77% ↓ |
- | Transformer-GRU (S) | 55.71% | 56.01% | 55.43% |
BiGRU-GRU (T) | Transformer-BiGRU(S) | 57.12% ↑ | 57.18% ↑ | 56.75% ↑ |
- | Transformer-BiGRU(S) | 54.53% | 55.00% | 54.18% |
BiGRU-GRU (T) | Transformer-FC(S) | 52.86% ↓ | 52.80% ↓ | 52.19% ↓ |
- | Transformer-FC(S) | 53.87% | 53.81% | 53.28% |
Teacher | Student | P | R | F1 |
---|---|---|---|---|
BiGRU-GRU (T) | Transformer-GRU (S) | 72.55% ↑ | 72.67% ↑ | 72.36% ↑ |
- | Transformer-GRU (S) | 69.24% | 69.31% | 69.09% |
BiGRU-GRU (T) | Transformer-BiGRU(S) | 70.66% ↑ | 70.27% ↑ | 70.12% ↑ |
- | Transformer-BiGRU(S) | 69.74% | 69.82% | 69.61% |
BiGRU-GRU (T) | Transformer-FC(S) | 69.98% ↑ | 69.66% ↑ | 69.59% ↑ |
- | Transformer-FC(S) | 68.50% | 68.24% | 68.10% |
Data Splitting Method | Model | P | R | F1 |
---|---|---|---|---|
Tool grouping | Transformer-FC | 52.86% | 52.80% | 52.19% |
Transformer-GRU | 55.46% | 54.99% | 54.77% | |
Transformer-BiGRU | 57.12% | 57.18% | 56.75% | |
GS-XGBoost | 48.11% | 52.67% | 50.70% | |
Attention-CNN | 46.18% | 45.64% | 45.20% | |
PCNN-BiLSTM | 52.31% | 51.88% | 51.74% | |
Mixed grouping | Transformer-FC | 69.98% | 69.66% | 69.59% |
Transformer-GRU | 72.55% | 72.67% | 72.36% | |
Transformer-BiGRU | 70.66% | 70.27% | 70.12% | |
GS-XGBoost | 58.88% | 57.34% | 58.01% | |
Attention-CNN | 58.01% | 58.09% | 57.65% | |
PCNN-BiLSTM | 60.19% | 60.03% | 59.82% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, B.; Li, B.; Xue, B.; Dong, Z.; Wu, W. Research on the Prediction Method for the Wear State of Aerospace Titanium Alloy Cutting Tools Based on Knowledge Distillation. Processes 2025, 13, 1300. https://doi.org/10.3390/pr13051300
Liu B, Li B, Xue B, Dong Z, Wu W. Research on the Prediction Method for the Wear State of Aerospace Titanium Alloy Cutting Tools Based on Knowledge Distillation. Processes. 2025; 13(5):1300. https://doi.org/10.3390/pr13051300
Chicago/Turabian StyleLiu, Bengang, Baode Li, Bo Xue, Zeguang Dong, and Wenjiang Wu. 2025. "Research on the Prediction Method for the Wear State of Aerospace Titanium Alloy Cutting Tools Based on Knowledge Distillation" Processes 13, no. 5: 1300. https://doi.org/10.3390/pr13051300
APA StyleLiu, B., Li, B., Xue, B., Dong, Z., & Wu, W. (2025). Research on the Prediction Method for the Wear State of Aerospace Titanium Alloy Cutting Tools Based on Knowledge Distillation. Processes, 13(5), 1300. https://doi.org/10.3390/pr13051300