Model and Method for Providing Resilience to Resource-Constrained AI-System
Abstract
:1. Introduction
1.1. Motivation
1.2. State-of-the-Art
1.3. Objectives and Contributions
- -
- To develop a new resource-efficient model and a training method, which simultaneously implement components of resilience such as robustness, fast recovery, and improvement;
- -
- To test the model’s and training method’s ability to provide robustness, fast recovery, and improvement.
2. Architecture of Resource-Efficient and Resilient AI-Model
2.1. Principles
- -
- Combining ReLU6 and LeakyReLU activation functions for efficient training and inference under conditions of data and neural network weight perturbations;
- -
- Dividing the model into blocks (e.g., several convolutional layers, multi-head self-attention, feed-forward network), and adding skip-connections that bypass each block and gate modules that can disable blocks depending on the input context;
- -
- Gate modules should be significantly computationally simpler than the model blocks they switch.
2.2. Selected Architectures
- -
- Low computational complexity compared to the building block that is activated or deactivated;
- -
- Stochasticity to prevent the mode from decaying into trivial decisions, such as always or never executing a block;
- -
- The ability to generate discrete solutions and calculate gradients to optimize the parameters of the gate unit.
3. Training Method Design
3.1. Principles
- -
- Simultaneous training of the main network weights and the weights of the gate units;
- -
- Training is performed first on the main training set under normal conditions, and then performed as episodic few-shot learning tasks with adaptation to each type of synthetic perturbation;
- -
- Generation of synthetic perturbations of data or weights according to the white box scenario;
- -
- Implementing generation of perturbations, such as novelty or drift of concepts, by changing tasks;
- -
- Generalization of experience during the preparatory phase should be based on meta-updating using the MAML, REPTIL algorithm or meta-free weight averaging;
- -
- The loss function of the neural network model should contain a component that characterizes the complexity of calculations or the degree of deviation from the desired level of network compression.
- -
- Only those blocks that were activated during the forward pass are corrected, so you need to save the output of each gate after forward passing;
- -
- Unlabeled images on which the model has a significant predictive uncertainty can be used for test-time adaptation.
3.2. Methods of Ensuring the Resilience of Resource-Constrained AI-System
4. Experiments and Results
4.1. Experimental Setup
4.2. Results
5. Discussion
6. Conclusions
6.1. Summary
6.2. Limitations
6.3. Future Research Directions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Li, Z.; Li, H.; Meng, L. Model Compression for Deep Neural Networks: A Survey. Computers 2023, 12, 60. [Google Scholar] [CrossRef]
- Marinó, G.C.; Petrini, A.; Malchiodi, D.; Frasca, M. Deep neural networks compression: A comparative survey and choice recommendations. Neurocomputing 2023, 520, 152–170. [Google Scholar] [CrossRef]
- Moskalenko, V.; Kharchenko, V.; Moskalenko, A.; Kuzikov, B. Resilience and Resilient Systems of Artificial Intelligence: Taxonomy, Models and Methods. Algorithms 2023, 16, 165. [Google Scholar] [CrossRef]
- Olowononi, F.O.; Rawat, D.B.; Liu, C. Resilient Machine Learning for Networked Cyber Physical Systems: A Survey for Machine Learning Security to Securing Machine Learning for CPS. IEEE Commun. Surv. Tutor. 2021, 23, 524–552. [Google Scholar] [CrossRef]
- Samangouei, P.; Kabkab, M.; Chellappa, R. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models (Version 2). arXiv 2018. [Google Scholar] [CrossRef]
- Hussain, M.; Hong, J.-E. Reconstruction-Based Adversarial Attack Detection in Vision-Based Autonomous Driving Systems. Mach. Learn. Knowl. Extr. 2023, 5, 1589–1611. [Google Scholar] [CrossRef]
- Ho, J.; Jain, A.; Abbeel, P. Denoising Diffusion Probabilistic Models (Version 2). arXiv 2020. [Google Scholar] [CrossRef]
- Sooksatra, K.; Hamerly, G.; Rivas, P. Is ReLU Adversarially Robust? [Poster Presentation]. In Proceedings of the Computer Vision and Pattern Recognition Conference: LatinX in AI (LXAI) Research Workshop 2023, Honolulu, Hawaii, 23–29 July 2023. [Google Scholar] [CrossRef]
- Hou, X.; Breier, J.; Jap, D.; Ma, L.; Bhasin, S.; Liu, Y. Security Evaluation of Deep Neural Network Resistance Against Laser Fault Injection. In Proceedings of the 2020 IEEE International Symposium on the Physical and Failure Analysis of Integrated Circuits (IPFA), Singapore, 20–23 July 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar] [CrossRef]
- Guo, Y.; Li, S.; Lerman, G. The effect of Leaky ReLUs on the training and generalization of overparameterized networks. Proc. Mach. Learn. Res. 2024, 238, 4393–4401. [Google Scholar] [CrossRef]
- Cavagnero, N.; Santos, F.D.; Ciccone, M.; Averta, G.; Tommasi, T.; Rech, P. Transient-Fault-Aware Design and Training to Enhance DNNs Reliability with Zero-Overhead. In Proceedings of the 2022 IEEE 28th International Symposium on On-Line Testing and Robust System Design (IOLTS), Torino, Italy, 12–14 September 2022; pp. 1–7. [Google Scholar] [CrossRef]
- Niu, Z.; Chen, Z.; Li, L.; Yang, Y.; Li, B.; Yi, J. On the Limitations of Denoising Strategies as Adversarial Defenses. arXiv 2020. [Google Scholar] [CrossRef]
- Eleftheriadis, C.; Symeonidis, A.; Katsaros, P. Adversarial robustness improvement for deep neural networks. Mach. Vis. Appl. 2024, 35, 3. [Google Scholar] [CrossRef]
- Sum, J.; Leung, C.-S. Regularization Effect of Random Node Fault/Noise on Gradient Descent Learning Algorithm. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 2619–2632. [Google Scholar] [CrossRef] [PubMed]
- Zhang, M.; Levine, S.; Finn, C. MEMO: Test Time Robustness via Adaptation and Augmentation. arXiv 2022. [Google Scholar] [CrossRef]
- Son, X.; Yang, Y.; Choromanski, K.; Caluwaerts, K.; Gao, W.; Finn, C.; Tan, J. Rapidly adaptable legged robots via evolutionary meta-learning. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 3769–3776. [Google Scholar] [CrossRef]
- Wang, R.; Xu, K.; Liu, S.; Chen, P.-Y.; Weng, T.W.; Gan, C.; Wang, M. On Fast Adversarial Robustness Adaptation in Model-Agnostic Meta-Learning. arXiv 2021. [Google Scholar] [CrossRef]
- Ding, N.; Qin, Y.; Yang, G.; Wei, F.; Yang, Z.; Su, Y.; Hu, S.; Chen, Y.; Chan, C.M.; Chen, W.; et al. Parameter-efficient fine-tuning of large-scale pre-trained language models. Nat. Mach. Intell. 2023, 5, 220–235. [Google Scholar] [CrossRef]
- Asadi, N.; Beitollahi, M.; Khalil, Y.; Li, Y.; Zhang, G.; Chen, X. Does Combining Parameter-efficient Modules Improve Few-shot Transfer Accuracy? (Version 1). arXiv 2024. [Google Scholar] [CrossRef]
- Wang, M.; Mo, J.; Lin, J.; Wang, Z.; Du, L. DynExit: A Dynamic Early-Exit Strategy for Deep Residual Networks. In Proceedings of the 2019 IEEE International Workshop on Signal Processing Systems (SiPS), Nanjing, China, 20–23 October 2019; pp. 178–183. [Google Scholar] [CrossRef]
- Veit, A.; Belongie, S. Convolutional Networks with Adaptive Inference Graphs. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8-14 September 2018; pp. 3–18. [Google Scholar] [CrossRef]
- Haque, M.; Yang, W. Dynamic Neural Network is All You Need: Understanding the Robustness of Dynamic Mechanisms in Neural Networks. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), Paris, France, 2–6 October 2023; pp. 1497–1506. [Google Scholar] [CrossRef]
- Moskalenko, V.; Moskalenko, A. Neural network based image classifier resilient to destructive perturbation influences—Architecture and training method. Radioelectron. Comput. Syst. 2022, 3, 95–109. [Google Scholar] [CrossRef]
- Wang, J.; Zhang, Z.; Wang, M.; Qiu, H.; Zhang, T.; Li, Q.; Li, Z.; Wei, T.; Zhang, C. Aegis: Mitigating Targeted Bit-flip Attacks against Deep Neural Networks (Version 1). arXiv 2023. [Google Scholar] [CrossRef]
- Peng, Y.; Lee, J.; Watanabe, S. I3D: Transformer Architectures with Input-Dependent Dynamic Depth for Speech Recognition. In Proceedings of the ICASSP 2023—2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
- Meng, L.; Li, H.; Chen, B.-C.; Lan, S.; Wu, Z.; Jiang, Y.-G.; Lim, S.-N. AdaViT: Adaptive Vision Transformers for Efficient Image Recognition. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 12309–12318. [Google Scholar] [CrossRef]
- Shen, T.; Lee, C.; Narayanan, V. Multi-Exit Vision Transformer with Custom Fine-Tuning for Fine-Grained Image Recognition. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023; pp. 2830–2834. [Google Scholar] [CrossRef]
- Moskalenko, V.; Kharchenko, V. Resilience-aware MLOps for AI-based medical diagnostic system. Front. Public Health 2024, 12, 1342937. [Google Scholar] [CrossRef]
- Gharoun, H.; Momenifar, F.; Chen, F.; Gandomi, A.H. Meta-learning approaches for few-shot learning: A survey of recent advances (Version 1). arXiv 2023. [Google Scholar] [CrossRef]
- Bortsova, G.; González-Gonzalo, C.; Wetstein, S.C.; Dubost, F.; Katramados, I.; Hogeweg, L. Adversarial attack vulnerability of medical image analysis systems: Unexplored factors. Med. Image Anal. 2021, 73, 102141. [Google Scholar] [CrossRef]
- Kotyan, S.; Vargas, D.V. Adversarial robustness assessment: Why in evaluation both L0 and L∞ attacks are necessary. PLoS ONE 2022, 17, e0265723. [Google Scholar] [CrossRef] [PubMed]
- Li, G.; Pattabiraman, K.; DeBardeleben, N. TensorFI: A Configurable Fault Injector for TensorFlow Applications. In Proceedings of the 2018 IEEE International Symposium on Software Reliability Engineering Workshops (ISSREW), Memphis, TN, USA, 15–18 October 2018; pp. 313–320. [Google Scholar] [CrossRef]
- Foldy-Porto, T.; Venkatesha, Y.; Panda, P. Activation Density Driven Efficient Pruning in Training. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 8929–8936. [Google Scholar] [CrossRef]
- Moskalenko, V.V. Model-Agnostic Meta-Learning for Resilience Optimization of Artificial Intelligence System. Radio Electron. Comput. Sci. Control 2023, 2, 79. [Google Scholar] [CrossRef]
- Chen, Z.; Narayanan, N.; Fang, B.; Li, G.; Pattabiraman, K.; DeBardeleben, N. TensorFI: A Flexible Fault Injection Framework for TensorFlow Applications. In Proceedings of the 2020 IEEE 31st International Symposium on Software Reliability Engineering (ISSRE), Coimbra, Portugal, 12–15 October 2020; IEEE: Piscataway, NJ, USA. [Google Scholar] [CrossRef]
- Lysenko, S.; Kharchenko, V.; Bobrovnikova, K.; Shchuka, R. Computer systems resilience in the presence of cyber threats: Taxonomy and ontology. Radioelectron. Comput. Syst. 2020, 1, 17–28. [Google Scholar]
Dynamic Compression Rate | Pretrained Model under Normal Condition | Pretrained Model under Fault Injection | Pretrained Model Optimized for Resilience | ||||||
---|---|---|---|---|---|---|---|---|---|
MFLOPs | MFLOPs | MFLOPs | |||||||
1.0 | 510.48 | 85.1% | 0.774 | 510.48 | 87.5% | 0.944 | 510.48 | 89.1% | 0.973 |
0.8 | 408.56 | 86.5% | 0.792 | 410.44 | 87.6% | 0.949 | 416.11 | 89.2% | 0.986 |
0.6 | 306.64 | 87.0% | 0.791 | 310.66 | 87.9% | 0.953 | 312.30 | 89.9% | 0.988 |
0.4 | 204.72 | 81.7% | 0.771 | 206.55 | 85.5% | 0.950 | 208.52 | 89.1% | 0.975 |
0.2 | 102.80 | 75.4% | 0.694 | 103.33 | 80.1% | 0.942 | 104.69 | 87.9% | 0.957 |
Dynamic Compression Rate | Pretrained Model under Normal Condition | Pretrained Model under Fault Injection | Pretrained Model Optimized for Resilience | ||||||
---|---|---|---|---|---|---|---|---|---|
GFLOPs | GFLOPs | GFLOPs | |||||||
1.0 | 10.56 | 86.3% | 0.763 | 10.56 | 88.1% | 0.933 | 10.56 | 90.6% | 0.961 |
0.8 | 7.92 | 87.2% | 0.781 | 8.11 | 88.8% | 0.941 | 8.80 | 91.1% | 0.970 |
0.6 | 6.16 | 88.3% | 0.788 | 6.55 | 89.2% | 0.949 | 7.04 | 92.2% | 0.979 |
0.4 | 3.52 | 84.6% | 0.764 | 4.07 | 87.1% | 0.943 | 4.40 | 90.5% | 0.965 |
0.2 | 1.76 | 77.2% | 0.688 | 1.98 | 83.3% | 0.931 | 2.64 | 88.8% | 0.944 |
Dynamic Compression Rate | Pretrained Model under Normal Condition | Pretrained Model under Adversarial Attack | Pretrained Model Optimized for Resilience | ||||||
---|---|---|---|---|---|---|---|---|---|
MFLOPs | MFLOPs | MFLOPs | |||||||
1.0 | 510.48 | 83.1% | 0.754 | 510.48 | 85.5% | 0.801 | 510.48 | 88.1% | 0.902 |
0.8 | 409.99 | 83.5% | 0.772 | 412.13 | 86.1% | 0.822 | 424.01 | 88.2% | 0.910 |
0.6 | 309.71 | 84.0% | 0.784 | 312.01 | 86.8% | 0.853 | 319.54 | 88.8% | 0.917 |
0.4 | 206.88 | 75.7% | 0.781 | 207.43 | 84.4% | 0.850 | 208.52 | 87.4% | 0.905 |
0.2 | 103.10 | 73.4% | 0.685 | 103.98 | 78.4% | 0.841 | 104.55 | 87.1% | 0.888 |
Dynamic Compression Rate | Pretrained Model under Normal Condition | Pretrained Model under Adversarial Attack | Pretrained Model Optimized for Resilience | ||||||
---|---|---|---|---|---|---|---|---|---|
GFLOPs | GFLOPs | GFLOPs | |||||||
1.0 | 10.56 | 85.2% | 0.744 | 10.56 | 85.9% | 0.781 | 10.56 | 88.9% | 0.911 |
0.8 | 7.92 | 85.7% | 0.751 | 8.61 | 86.8% | 0.812 | 9.42 | 88.9% | 0.923 |
0.6 | 6.50 | 85.7% | 0.759 | 6.95 | 87.1% | 0.844 | 7.26 | 90.1% | 0.923 |
0.4 | 3.22 | 77.6% | 0.731 | 3.87 | 86.3% | 0.839 | 4.22 | 88.5% | 0.920 |
0.2 | 1.76 | 75.0% | 0.705 | 1.88 | 80.7% | 0.833 | 2.24 | 88.1% | 0.914 |
Dynamic Compression Rate | Fine-Tuning of ResNet-110-Based AI System | Fine-Tuning of DeiT-S-Based AI System | ||
---|---|---|---|---|
Pretrained on the Base Dataset under Normal Condition | Meta-Trained on Result of Adaptation to Disturbances for Resilience Optimization | Pretrained on Base Dataset under Normal Condition | Meta-Trained on Result of Adaptation to Disturbances for Resilience Optimization | |
1.0 | 0.791 | 0.962 | 0.761 | 0.921 |
0.8 | 0.809 | 0.978 | 0.762 | 0.930 |
0.6 | 0.805 | 0.988 | 0.759 | 0.935 |
0.4 | 0.761 | 0.955 | 0.732 | 0.902 |
0.2 | 0.731 | 0.892 | 0.711 | 0.803 |
Disturbance Type | Fine-Tuning of ResNet-110-Based AI System | Fine-Tuning of DeiT-S-Based AI System | ||
---|---|---|---|---|
ReLU | LeakyReLU6 | ReLU | LeakyReLU6 | |
Fault Injection | 0.953 | 0.988 | 0.950 | 0.979 |
Adversarial Attack | 0.891 | 0.917 | 0.884 | 0.923 |
Task Change | 0.978 | 0.988 | 0.911 | 0.935 |
Meta-Trained Backbone | Fine-Tuning Setting | (%) under Adversarial Attack | (%) under Fault Injection Attack |
---|---|---|---|
ResNet-110 | Only Supervised Fine-Tuning | 3.1 | 2.9 |
Only Test-Time-Adaptation | 9.7 | 11.3 | |
DeiT-S | Only Supervised Fine-Tuning | 2.6 | 1.9 |
Only Test-Time-Adaptation | 11.7 | 10.5 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Moskalenko, V.; Kharchenko, V.; Semenov, S. Model and Method for Providing Resilience to Resource-Constrained AI-System. Sensors 2024, 24, 5951. https://doi.org/10.3390/s24185951
Moskalenko V, Kharchenko V, Semenov S. Model and Method for Providing Resilience to Resource-Constrained AI-System. Sensors. 2024; 24(18):5951. https://doi.org/10.3390/s24185951
Chicago/Turabian StyleMoskalenko, Viacheslav, Vyacheslav Kharchenko, and Serhii Semenov. 2024. "Model and Method for Providing Resilience to Resource-Constrained AI-System" Sensors 24, no. 18: 5951. https://doi.org/10.3390/s24185951
APA StyleMoskalenko, V., Kharchenko, V., & Semenov, S. (2024). Model and Method for Providing Resilience to Resource-Constrained AI-System. Sensors, 24(18), 5951. https://doi.org/10.3390/s24185951