Generative Adversarial Network-Based Anomaly Detection and Forecasting with Unlabeled Data for 5G Vertical Applications
Abstract
:1. Introduction
- First, to achieve anomaly detection in 5G vertical applications, the strengths of the autoencoder and the GANs are combined. The autoencoder efficiently distills the significant characteristics of the data, while GAN provides robustness by generating adversarial examples. This innovative combination greatly enhances both the accuracy and stability of our anomaly detection model, ensuring the precise identification and consistent performance across diverse data contexts.
- Second, in order to proactively prepare for anomalies, we employ an LSTM model to predict the 5G network quality data, followed by the use of unsupervised anomaly detection methods to identify anomalies in the predicted data. By leveraging the predictive power of the LSTM model, our approach seeks to anticipate anomalous behavior in advance, and subsequently mitigate any potential negative impacts on the network.
- Finally, the simulation results show that our method outperforms traditional anomaly detection algorithms. Moreover, our method demonstrates a remarkable accuracy in detecting and forecasting anomalies in unlabeled network data. Therefore, it has good performance in network quality monitoring for 5G vertical applications.
2. System Model and Problem Formulation
- Data process: For the unlabeled network data, we express the set of time periods and that of data dimensions as and , respectively. On this basis, the time series data of length T can be expressed as , where denotes the network data at a specific time point t with dimensions of M. To capture the relationship between the current time point and its previous ones, a processing procedure is necessary for the network data. This procedure transforms the data into a series of window data, denoted by . Each window data at time point t is defined as a sequence of from to t, expressed as . It is worth noting that the length of each window data is K.
- Anomaly detection: Anomaly detection is applied to the processed window data using an anomaly detection model. The underlying principle of anomaly detection involves training the model on known data samples and assessing the abnormality level of new data samples based on this learned representation. By establishing a threshold , when the abnormality level of a new data sample surpasses the threshold at a given time point, it is classified as an anomaly. Hence, to identify the presence of anomalies in the new data samples, it is crucial to obtain a series of anomaly labels, which can be represented as , where indicates whether there is an anomaly in the sample at time point t.
- Anomaly forecasting: In contrast to anomaly detection, anomaly forecasting focuses on identifying anomalies in future time data. To accomplish this, the processed window data is initially used as input for a prediction model, which generates the forecasts of the network data for the upcoming time points denoted by . Subsequently, an anomaly detection model is applied to the predicted data, enabling the detection of abnormal patterns and providing a series of anomaly labels. Consequently, this approach facilitates accurate anomaly forecasting.
3. Network Anomaly Detection
3.1. Data Acquisition and Process
3.1.1. Feature Library Construction Framework
3.1.2. Feature Library Construction Methodology
3.2. The Structure of Network Anomaly Detection
3.2.1. Autoencoder-Based Feature Transformation
3.2.2. Enhancement via Adversarial-Based High-Dimensional Feature Transformation
3.2.3. Acquisition of Anomaly Score
3.3. The Algorithm of Network Anomaly Detection
3.3.1. The Two-Phase Training Process
- Phase 1: Autoencoder training. Initially, the input data, represented by W, is subject to compression by the encoder, symbolized as E. This procedure reduces the input data into a latent space, denoted by Z. Following this, each of the two decoders embarks on the task of reconstruction, with the outcomes manifesting as and , correspondingly. The errors incurred in this reconstruction process, can hence be individually articulated as follows:
- Phase 2: Adversarial training. Subsequent to the initial reconstruction by the two decoders, the output derived from is subjected to another compression by E, transitioning once more into the latent space Z. This compressed output is then conveyed to for another round of reconstruction, resulting in the output . This procedure allows to be conditioned in such a way as to deceive , whereas is trained to differentiate between the original data and the data generated from . Consequently, aspires to minimize the disparity between the original data W and the outputs from , while endeavors to maximize this distinction. Within this adversarial framework, the reconstruction errors can be represented as follows:
3.3.2. The Inference Process
Algorithm 1 Anomaly detection algorithm. |
Input: Historical window data , new window data , epoch N, threshold , weight parameter Output: Anomaly labels
|
4. Network Anomaly Forecasting
4.1. The Structure of Network Anomaly Forecasting
4.1.1. Network Data Prediction
4.1.2. Anomaly Forecasting Function
4.2. The Algorithm of Network Anomaly Forecasting
Algorithm 2 Anomaly forecasting algorithm. |
Input: Historical window data , threshold , weight parameter Output: anomaly labels of predicted data
|
- Data Input: Utilizing the historical data in the past time as the input, the input gate of a memory block in LSTM can make full use of temporal information in all time points to identify changes in features and provide a basis for subsequent feature screening by updating the cell state. In other words, input gates determine what feature from the network input at the current moment is saved to the cell state . How it updates is shown as follows:
- Feature screening: In the time windows of input, there are some KPI features that contribute little to prediction because they are correlated to other features, which can be inferred by other important features. Thus, a forgetting gate can filter out the features that contribute more to prediction, where the weight matrix is used to achieve this function, and then combine these features and input information at the current moment to complete the prediction. The forgetting gate determines how much of the unit state at the previous moment is retained to the current state . The definition of a forgetting gate is:
- Predicted results output: Combining the filtered features and current input, the memory block output the predicted data at current time and hidden state through output gate. The hidden state is transmitted to the next memory block to predict the data of the next time point. After traversing multiple memory blocks, the final memory block receives the cell state and hidden states which contain all the information on historical time, and it outputs the final result of prediction. The output gate decides the next , which will be pass with the new cell state to the next memory block. The process of updating is presented as follows:
5. Experiments and Results
5.1. Experimental Setup
5.2. 5G Vertical Applications Platform
- Network equipment: This includes data collected from base stations, switches, routers, and other hardware elements that make up the network infrastructure.
- Network management systems: These systems monitor and manage the operation of the network, and can provide data about network traffic, faults, performance, and security.
- User equipment: This includes data from user devices connected to the network, which can provide data on usage patterns, performance, and service quality.
5.3. Performance of the Proposed Anomaly Detection Method
5.4. Performance of the Proposed Anomaly Forecasting Method
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Rao, S.K.; Prasad, R. Impact of 5G technologies on industry 4.0. Wirel. Pers. Commun. 2018, 100, 145–159. [Google Scholar] [CrossRef]
- Zhou, K.; Liu, T.; Zhou, L. Industry 4.0: Towards future industrial opportunities and challenges. In Proceedings of the 2015 IEEE 12th International Conference on Fuzzy Systems and Knowledge Discovery (FSKD), Zhangjiajie, China, 15–17 August 2015; pp. 2147–2152. [Google Scholar]
- Malanchini, I.; Agostini, P.; Alam, K.; Baumgart, M.; Kasparick, M.; Liao, Q.; Lipp, F.; Marchenko, N.; Michailow, N.; Pries, R.; et al. Leveraging machine learning for industrial wireless communications. arXiv 2021, arXiv:2105.02051. [Google Scholar]
- Rostami, A. Private 5G networks for vertical industries: Deployment and operation models. In Proceedings of the 2019 IEEE 2nd 5G World Forum (5GWF), Dresden, Germany, 30 September–2 October 2019; pp. 433–439. [Google Scholar]
- Norp, T. 5G requirements and key performance indicators. J. Ict Stand. 2018, 15–30. [Google Scholar] [CrossRef]
- Morgado, A.; Huq, K.M.S.; Mumtaz, S.; Rodriguez, J. A survey of 5G technologies: Regulatory, standardization and industrial perspectives. Digit. Commun. Netw. 2018, 4, 87–97. [Google Scholar] [CrossRef]
- Lee, S.; Levanti, K.; Kim, H.S. Network monitoring: Present and future. Comput. Netw. 2014, 65, 84–98. [Google Scholar] [CrossRef]
- Ding, A.Y.; Janssen, M. Opportunities for applications using 5G networks: Requirements, challenges, and outlook. In Proceedings of the Seventh International Conference on Telecommunications and Remote Sensing, Barcelona, Spain, 8–9 October 2018; pp. 27–34. [Google Scholar]
- Zhang, S. An overview of network slicing for 5G. IEEE Wirel. Commun. 2019, 26, 111–117. [Google Scholar] [CrossRef]
- Banchs, A.; Gutierrez-Estevez, D.M.; Fuentes, M.; Boldi, M.; Provvedi, S. A 5G mobile network architecture to support vertical industries. IEEE Commun. Mag. 2019, 57, 38–44. [Google Scholar] [CrossRef]
- Cayamcela, M.E.M.; Lim, W. Artificial intelligence in 5G technology: A survey. In Proceedings of the IEEE 2018 International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 19–21 October 2018; pp. 860–865. [Google Scholar]
- He, Z.; Xu, X.; Deng, S. Discovering cluster-based local outliers. Pattern Recognit. Lett. 2003, 24, 1641–1650. [Google Scholar] [CrossRef]
- Idé, T.; Papadimitriou, S.; Vlachos, M. Computing correlation anomaly scores using stochastic nearest neighbors. In Proceedings of the Seventh IEEE International Conference on Data Mining (ICDM 2007), Omaha, NE, USA, 28–31 October 2007; pp. 523–528. [Google Scholar]
- Sarafijanovic-Djukic, N.; Davis, J. Fast distance-based anomaly detection in images using an inception-like autoencoder. In Proceedings of the International Conference on Discovery Science, Split, Croatia, 28–30 October 2019; Springer: Cham, Switzerland, 2019; pp. 493–508. [Google Scholar]
- Chaovalitwongse, W.A.; Fan, Y.J.; Sachdeo, R.C. On the time series k-nearest neighbor classification of abnormal brain activity. IEEE Trans. Syst. Man Cybern.-Part A Syst. Hum. 2007, 37, 1005–1016. [Google Scholar] [CrossRef]
- Hautamaki, V.; Karkkainen, I.; Franti, P. Outlier detection using k-nearest neighbour graph. In Proceedings of the IEEE 17th International Conference on Pattern Recognition, Cambridge, UK, 26 August 2004; Volume 3, pp. 430–433. [Google Scholar]
- Ma, Q.; Zheng, J.; Li, S.; Cottrell, G.W. Learning representations for time series clustering. Adv. Neural Inf. Process. Syst. 2019, 32. [Google Scholar] [CrossRef]
- Kiss, I.; Genge, B.; Haller, P.; Sebestyén, G. Data clustering-based anomaly detection in industrial control systems. In Proceedings of the 2014 IEEE 10th International Conference on Intelligent Computer Communication and Processing (ICCP), Cluj-Napoca, Cluj, Romania, 4–6 September 2014; pp. 275–281. [Google Scholar]
- Divekar, A.; Parekh, M.; Savla, V.; Mishra, R.; Shirole, M. Benchmarking datasets for anomaly-based network intrusion detection: KDD CUP 99 alternatives. In Proceedings of the 2018 IEEE 3rd International Conference on Computing, Communication and Security (ICCCS), Kathmandu, Nepal, 25–27 October 2018; pp. 1–8. [Google Scholar]
- Ma, J.; Perkins, S. Time-series novelty detection using one-class support vector machines. In Proceedings of the IEEE International Joint Conference on Neural Networks, Portland, OR, USA, 20–24 July 2003; Volume 3, pp. 1741–1745. [Google Scholar]
- Manevitz, L.M.; Yousef, M. One-class SVMs for document classification. J. Mach. Learn. Res. 2001, 2, 139–154. [Google Scholar]
- Javaheri, D.; Gorgin, S.; Jeong-A, L.; Masdari, M. Fuzzy logic-based DDoS attacks and network traffic anomaly detection methods: Classification, overview, and future perspectives. Inf. Sci. 2023, 626, 315–338. [Google Scholar] [CrossRef]
- Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. ACM Comput. Surv. 2009, 41, 1–58. [Google Scholar] [CrossRef]
- Gupta, M.; Gao, J.; Aggarwal, C.C.; Han, J. Outlier detection for temporal data: A survey. IEEE Trans. Knowl. Data Eng. 2013, 26, 2250–2267. [Google Scholar] [CrossRef]
- Zhang, C.; Song, D.; Chen, Y.; Feng, X.; Lumezanu, C.; Cheng, W.; Ni, J.; Zong, B.; Chen, H.; Chawla, N.V. A deep neural network for unsupervised anomaly detection and diagnosis in multivariate time series data. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 1409–1416. [Google Scholar]
- Malhotra, P.; Ramakrishnan, A.; Anand, G.; Vig, L.; Agarwal, P.; Shroff, G. LSTM-based encoder-decoder for multi-sensor anomaly detection. arXiv 2016, arXiv:1607.00148. [Google Scholar]
- Yuan, Y.; Xun, G.; Ma, F.; Wang, Y.; Du, N.; Jia, K.; Su, L.; Zhang, A. Muvan: A multi-view attention network for multivariate temporal data. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018; pp. 717–726. [Google Scholar]
- Su, Y.; Zhao, Y.; Niu, C.; Liu, R.; Sun, W.; Pei, D. Robust anomaly detection for multivariate time series through stochastic recurrent neural network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 2828–2837. [Google Scholar]
- Bonetto, R.; Soldan, M.; Lanaro, A.; Milani, S.; Rossi, M. Seq2Seq RNN based gait anomaly detection from smartphone acquired multimodal motion data. arXiv 2019, arXiv:1911.08608. [Google Scholar]
- Imtiaz, U.M.; Qusay, H. Design and development of RNN anomaly detection model for IoT networks. IEEE Access 2022, 10, 62722–62750. [Google Scholar]
- Nenavath, C.; Mummadi, U.K. Metaheuristic feature selection with deep learning enabled cascaded recurrent neural network for anomaly detection in Industrial Internet of Things environment. Clust. Comput. 2023, 26, 1801–1819. [Google Scholar]
- Wei, Y.; Jaccard-Jang, J.; Xu, W.; Fariza, S.; Camtepe, S.; Boulic, M. LSTM-autoencoder-based anomaly detection for indoor air quality time-series data. IEEE Sens. J. 2023, 23, 3787–3800. [Google Scholar] [CrossRef]
- Qin, Y.; Song, D.; Chen, H.; Cheng, W.; Jiang, G.; Cottrell, G. A dual-stage attention-based recurrent neural network for time series prediction. arXiv 2017, arXiv:1704.02971. [Google Scholar]
- Geiger, A.; Liu, D.; Alnegheimish, S.; Cuesta-Infante, A.; Veeramachaneni, K. TadGAN: Time series anomaly detection using generative adversarial networks. In Proceedings of the 2020 IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; pp. 33–43. [Google Scholar]
- Audibert, J.; Michiardi, P.; Guyard, F.; Marti, S.; Zuluaga, M.A. Usad: Unsupervised anomaly detection on multivariate time series. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Online, 6–10 July 2020; pp. 3395–3404. [Google Scholar]
- Minango, P.I.; Chuma, Y.; Vaz, E.L.; de Oliveira Caumo, G.; Gomes, G.; Minango, J. Revision of the 5G concept rollout and its application in smart cities: A study case in South America. In Proceedings of the Brazilian Technology Symposium, Campinas, Brazil, 19–21 October 2021; Springer: Cham, Switzerland, 2021; pp. 229–238. [Google Scholar]
Symbol | Description |
---|---|
The set of time periods | |
The set of data dimensions | |
, | Time series data, is the network data at a specific time point t |
, | Window data, is the window data at a specific time point t |
, | The series of anomaly labels, is the anomaly label at a specific time point t |
, | Forecasts of the network data, is the forecast data at a specific time point t |
Cell Index | The Proposed Scheme | Autoencoder-Only | ||||||
---|---|---|---|---|---|---|---|---|
P | R | F1 | Average F1 | P | R | F1 | Average F1 | |
Internet of Vehicles | ||||||||
0 | 0.933 | 0.921 | 0.927 | 0.926 | 0.836 | 0.821 | 0.828 | 0.830 |
1 | 0.938 | 0.913 | 0.925 | 0.827 | 0.831 | 0.829 | ||
2 | 0.917 | 0.941 | 0.929 | 0.851 | 0.817 | 0.834 | ||
3 | 0.926 | 0.926 | 0.926 | 0.826 | 0.832 | 0.829 | ||
4 | 0.942 | 0.908 | 0.925 | 0.831 | 0.826 | 0.828 | ||
Intelligent manufacturing | ||||||||
0 | 0.864 | 0.919 | 0.891 | 0.894 | 0.811 | 0.876 | 0.842 | 0.840 |
1 | 0.863 | 0.923 | 0.892 | 0.817 | 0.874 | 0.845 | ||
2 | 0.860 | 0.937 | 0.897 | 0.815 | 0.871 | 0.842 | ||
3 | 0.856 | 0.939 | 0.896 | 0.809 | 0.867 | 0.837 | ||
4 | 0.849 | 0.944 | 0.894 | 0.808 | 0.866 | 0.836 | ||
Industrial Internet | ||||||||
0 | 0.861 | 0.918 | 0.889 | 0.888 | 0.796 | 0.867 | 0.830 | 0.825 |
1 | 0.858 | 0.925 | 0.890 | 0.795 | 0.865 | 0.829 | ||
2 | 0.853 | 0.927 | 0.888 | 0.791 | 0.862 | 0.825 | ||
3 | 0.849 | 0.928 | 0.887 | 0.786 | 0.859 | 0.821 | ||
4 | 0.847 | 0.932 | 0.887 | 0.785 | 0.857 | 0.819 |
Method | F1 | |
---|---|---|
The Proposed scheme | The Baseline Scheme | |
Internet of Vehicles | 0.912 | 0.701 |
Intelligent manufacturing | 0.899 | 0.687 |
Industrial Internet | 0.869 | 0.669 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhang, Q.; Chen, B.; Zhang, T.; Cao, K.; Ding, Y.; Gao, T.; Zhao, Z. Generative Adversarial Network-Based Anomaly Detection and Forecasting with Unlabeled Data for 5G Vertical Applications. Appl. Sci. 2023, 13, 10745. https://doi.org/10.3390/app131910745
Zhang Q, Chen B, Zhang T, Cao K, Ding Y, Gao T, Zhao Z. Generative Adversarial Network-Based Anomaly Detection and Forecasting with Unlabeled Data for 5G Vertical Applications. Applied Sciences. 2023; 13(19):10745. https://doi.org/10.3390/app131910745
Chicago/Turabian StyleZhang, Qing, Bin Chen, Taoye Zhang, Kang Cao, Yuming Ding, Tianhang Gao, and Zhongyuan Zhao. 2023. "Generative Adversarial Network-Based Anomaly Detection and Forecasting with Unlabeled Data for 5G Vertical Applications" Applied Sciences 13, no. 19: 10745. https://doi.org/10.3390/app131910745
APA StyleZhang, Q., Chen, B., Zhang, T., Cao, K., Ding, Y., Gao, T., & Zhao, Z. (2023). Generative Adversarial Network-Based Anomaly Detection and Forecasting with Unlabeled Data for 5G Vertical Applications. Applied Sciences, 13(19), 10745. https://doi.org/10.3390/app131910745