LAMBERT: Leveraging Attention Mechanisms to Improve the BERT Fine-Tuning Model for Encrypted Traffic Classification
Abstract
:1. Introduction
- We comprehensively analyse the feasibility of applying artificial intelligence techniques and BERT models to encrypted traffic classification and reveal the important impact of long-distance dependencies between byte sequences on the accuracy of encrypted traffic classification.
- We propose a new encrypted traffic classification model, LAMBERT, which uses the long-term byte sequence modelling to model the long and short distances of the byte feature sequences extracted by BERT and introduces the multiattention of the key enhancement module to solve the problem of sequence loss between long-distance dependencies. Furthermore, additive attention is used to reduce the dimensionality of the features output by multiattention while retaining the original features.
- We test the capabilities of the LAMBERT model by comparing it with ten other advanced models using three types of typical datasets. These data include imbalanced sample datasets (USTC-TFC and VPN-App), no pretrained datasets (USTC-TFC)) and large classification datasets (CSTNET-TLS 1.3). The experimental results show that the LAMBERT model can achieve high accuracy on these datasets. We have published the source code of the LAMBERT model on GitHub (https://github.com/MysteryObstacle/Lambert, accessed on 29 March 2024).
2. Related Work
3. Motivation and Objective
4. Methodology
4.1. Workflow
4.2. Preprocessing
4.3. Pre-Training Model
4.4. Fine-Tuning Model
4.4.1. Packet Feature Extraction Module
4.4.2. Long-Term Byte Sequence Modeling Module
4.4.3. Key Enhancement Module
4.4.4. Classification Output Module
4.4.5. Model Enhancement
5. Experiments and Results Analysis
5.1. Experimental Setup
5.1.1. Experimental Environment
5.1.2. Datasets and Downstream Tasks
5.1.3. Evaluation Metrics
5.2. Comparison with State-of-the-Art Methods
5.3. Rationality Analysis of Long-Distance Dependency between Bytes
5.4. Ablation Study
- Without BiGRU: LAMBERT model without BiGRU, where the output of BERT is directly fed to the attention block for fully connected classification, aiming to evaluate the effect of BiGRU.
- Without Multi-head Self-Attention (MA):LAMBERT model without multihead self-attention passes the time step output by the BiGRU to the full connection for classification after additive attention dimensionality reduction and is used to evaluate the effect of multihead self-attention.
- Without Additive Attention (AA): LAMBERT model without additive attention takes the horizontal average of the feature vectors obtained by multihead self-attention and then passes it to the fully connected layer for classification, which is used to evaluate the effect of additive attention.
- Without Fully Connected Layer (FC): A baseline model with only BERT and FC, which is a native BERT fine-tuned classification model. It feeds the [CLS] vector output by BERT directly into FC for classification, which is used as a baseline control along with LAMBERT (Table 3).
5.5. Unbalanced Sample Analysis
5.6. Analysis of Samples without Pre-Training
5.7. Large Classification Sample Analysis
6. Discussion
- Improve the fine-tuning methodIn this paper, we have mainly focused on the structural improvement of the fine-tuning model, but it has been shown that improving the fine-tuning method can also help to improve the classification performance. Some possible directions for improvement are as follows:
- (1)
- Further pre-training: For those scenarios where there is a lack of pre-training data, consider further pre-training on the original pretrained model. This can make the model adaptable to new domain data and improve its generalizability.
- (2)
- Multitask tuning: Multitask learning has shown promising performance in deep learning. Applying LAMBERT to multiple encrypted traffic classification tasks without pre-training datasets and sharing knowledge between these tasks may further improve the classification performance.
- Research on classification under a single complex encryption sysytemDifferent cryptographic implementations have different degrees of randomness [11]. According to the interpretability analysis of ET-BERT, the classification accuracy decreases significantly on the encryption system with a single encryption algorithm and strong randomness of the encryption algorithm. The experimental results of LAMBERT further confirm this view. The distribution of ciphertext is mainly affected by the original text, the encryption algorithm and the encryption configuration. It is a task full of potential and challenges to realize classification by perceiving the change law of the original text in the ciphertext distribution without using the randomness difference between different encryption algorithms.
- Classification research in superlarge target classification scenariosFrom the experimental results, we find that the classification accuracy decreases for datasets with more classification targets. On three classification tasks, ETVS, ETVA and ETMA, with 10 to 20 classification targets, the F1 score reaches more than 99%, while on the ETTA with 120 classification targets, the F1 score decreases to 97%. However, in some real classification scenarios, the number of classification targets often reaches thousands. It can be assumed that in this case, the accuracy of classification will decrease significantly. Therefore, it is a great challenge to maintain high classification accuracy in such superlarge object classification scenarios.
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Mohassel, P.; Zhang, Y. Secureml: A system for scalable privacy-preserving machine learning. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 19–38. [Google Scholar]
- Available online: https://transparencyreport.google.com/https/overview?hl=zh_CN (accessed on 1 January 2024).
- Ma, Y.; Chai, X.; Gan, Z.; Zhang, Y. Privacy-preserving TPE-based JPEG image retrieval in cloud-assisted internet of things. IEEE Internet Things J. 2023, 11, 4842–4856. [Google Scholar] [CrossRef]
- Ning, J.; Poh, G.S.; Loh, J.C.; Chia, J.; Chang, E.C. PrivDPI: Privacy-preserving encrypted traffic inspection with reusable obfuscated rules. In Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, London, UK, 11–15 November 2019; pp. 1657–1670. [Google Scholar]
- Fauvel, K.; Chen, F.; Rossi, D. A lightweight, efficient and explainable-by-design convolutional neural network for internet traffic classification. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Long Beach, CA, USA, 6–10 August 2023; pp. 4013–4023. [Google Scholar]
- Shi, H.; Li, H.; Zhang, D.; Cheng, C.; Cao, X. An efficient feature generation approach based on deep learning and feature selection techniques for traffic classification. Comput. Netw. 2018, 132, 81–98. [Google Scholar] [CrossRef]
- Bujlow, T.; Carela-Español, V.; Barlet-Ros, P. Independent comparison of popular DPI tools for traffic classification. Comput. Netw. 2015, 76, 75–89. [Google Scholar] [CrossRef]
- Rezaei, S.; Liu, X. Deep learning for encrypted traffic classification: An overview. IEEE Commun. Mag. 2019, 57, 76–81. [Google Scholar] [CrossRef]
- Suo, H.; Liu, Z.; Wan, J.; Zhou, K. Security and privacy in mobile cloud computing. In Proceedings of the 2013 9th International Wireless Communications and Mobile Computing Conference (IWCMC), Sardinia, Italy, 1–5 July 2013; pp. 655–659. [Google Scholar] [CrossRef]
- Xie, R.; Wang, Y.; Cao, J.; Dong, E.; Xu, M.; Sun, K.; Li, Q.; Shen, L.; Zhang, M. Rosetta: Enabling robust tls encrypted traffic classification in diverse network environments with tcp-aware traffic augmentation. In Proceedings of the ACM Turing Award Celebration Conference, Wuhan, China, 28–30 July 2023; pp. 131–132. [Google Scholar]
- Velan, P.; Čermák, M.; Čeleda, P.; Drašar, M. A survey of methods for encrypted traffic classification and analysis. Int. J. Netw. Manag. 2015, 25, 355–374. [Google Scholar] [CrossRef]
- He, W.; Li, S.; Wang, W.; Wei, M.; Qiu, B. Cryptoeyes: Privacy preserving classification over encrypted images. In Proceedings of the IEEE INFOCOM 2021-IEEE Conference on Computer Communications, Vancouver, BC, Canada, 10–13 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–10. [Google Scholar]
- Crotti, M.; Dusi, M.; Gringoli, F.; Salgarelli, L. Traffic classification through simple statistical fingerprinting. ACM SIGCOMM Comput. Commun. Rev. 2007, 37, 5–16. [Google Scholar] [CrossRef]
- Shapira, T.; Shavitt, Y. Flowpic: Encrypted internet traffic classification is as easy as image recognition. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference on Computer Communications Workshops (INFOCOM WKSHPS), Paris, France, 29 April–2 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 680–687. [Google Scholar]
- Yang, Y.; Kang, C.; Gou, G.; Li, Z.; Xiong, G. TLS/SSL encrypted traffic classification with autoencoder and convolutional neural network. In Proceedings of the 2018 IEEE 20th International Conference on High Performance Computing and Communications; IEEE 16th International Conference on Smart City; IEEE 4th International Conference on Data Science and Systems (HPCC/SmartCity/DSS), Exeter, UK, 28–30 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 362–369. [Google Scholar]
- Taylor, V.F.; Spolaor, R.; Conti, M.; Martinovic, I. Robust smartphone app identification via encrypted network traffic analysis. IEEE Trans. Inf. Forensics Secur. 2017, 13, 63–78. [Google Scholar] [CrossRef]
- Lin, X.; Xiong, G.; Gou, G.; Li, Z.; Shi, J.; Yu, J. Et-bert: A contextualized datagram representation with pre-training transformers for encrypted traffic classification. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; pp. 633–642. [Google Scholar]
- Aceto, G.; Ciuonzo, D.; Montieri, A.; Pescapè, A. MIMETIC: Mobile encrypted traffic classification using multimodal deep learning. Comput. Netw. 2019, 165, 106944. [Google Scholar] [CrossRef]
- Liu, C.; He, L.; Xiong, G.; Cao, Z.; Li, Z. Fs-net: A flow sequence network for encrypted traffic classification. In Proceedings of the IEEE INFOCOM 2019-IEEE Conference On Computer Communications, Paris, France, 29 April–2 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1171–1179. [Google Scholar]
- Ren, X.; Gu, H.; Wei, W. Tree-RNN: Tree structural recurrent neural network for network traffic classification. Expert Syst. Appl. 2021, 167, 114363. [Google Scholar] [CrossRef]
- Lotfollahi, M.; Jafari Siavoshani, M.; Shirali Hossein Zade, R.; Saberian, M. Deep packet: A novel approach for encrypted traffic classification using deep learning. Soft Comput. 2020, 24, 1999–2012. [Google Scholar] [CrossRef]
- Wang, W.; Zhu, M.; Wang, J.; Zeng, X.; Yang, Z. End-to-end encrypted traffic classification with one-dimensional convolution neural networks. In Proceedings of the 2017 IEEE International Conference on Intelligence and Security Informatics (ISI), Beijing, China, 22–24 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 43–48. [Google Scholar]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
- Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
- He, H.Y.; Yang, Z.G.; Chen, X.N. PERT: Payload encoding representation from transformer for encrypted traffic classification. In Proceedings of the 2020 ITU Kaleidoscope: Industry-Driven Digital Transformation (ITU K), Ha Noi, Vietnam, 7–11 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–8. [Google Scholar]
- Hu, X.; Gu, C.; Chen, Y.; Wei, F. CBD: A deep-learning-based scheme for encrypted traffic classification with a general pre-training method. Sensors 2021, 21, 8231. [Google Scholar] [CrossRef] [PubMed]
- Van Ede, T.; Bortolameotti, R.; Continella, A.; Ren, J.; Dubois, D.J.; Lindorfer, M.; Choffnes, D.; Van Steen, M.; Peter, A. Flowprint: Semi-supervised mobile-app fingerprinting on encrypted network traffic. In Proceedings of the Network and Distributed System Security Symposium (NDSS), San Diego, CA, USA, 23–26 February 2020; Volume 27. [Google Scholar]
- Al-Naami, K.; Chandra, S.; Mustafa, A.; Khan, L.; Lin, Z.; Hamlen, K.; Thuraisingham, B. Adaptive encrypted traffic fingerprinting with bi-directional dependence. In Proceedings of the 32nd Annual Conference on Computer Security Applications, Los Angeles, CA, USA, 5–9 December 2016; pp. 177–188. [Google Scholar]
- Sirinam, P.; Imani, M.; Juarez, M.; Wright, M. Deep fingerprinting: Undermining website fingerprinting defenses with deep learning. In Proceedings of the 2018 ACM SIGSAC Conference on Computer and Communications Security, Toronto, ON, Canada, 15–19 October 2018; pp. 1928–1943. [Google Scholar]
- Shi, Z.; Luktarhan, N.; Song, Y.; Tian, G. BFCN: A novel classification method of encrypted traffic based on BERT and CNN. Electronics 2023, 12, 516. [Google Scholar] [CrossRef]
- Ma, X.; Liu, T.; Hu, N.; Liu, X. Bi-ETC: A Bidirectional Encrypted Traffic Classification Model Based on BERT and BiLSTM. In Proceedings of the 2023 8th International Conference on Data Science in Cyberspace (DSC), Hefei, China, 18–20 August 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 197–204. [Google Scholar]
- Sengupta, S.; Ganguly, N.; De, P.; Chakraborty, S. Exploiting diversity in android tls implementations for mobile app traffic classification. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 1657–1668. [Google Scholar]
- Canard, S.; Diop, A.; Kheir, N.; Paindavoine, M.; Sabt, M. BlindIDS: Market-compliant and privacy-friendly intrusion detection system over encrypted traffic. In Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security, Abu Dhabi, United Arab Emirates, 2–6 April 2017; pp. 561–574. [Google Scholar]
- Weber, J.; Senior, A.E. Catalytic mechanism of F1-ATPase. Biochim. Biophys. Acta (BBA)—Bioenerg. 1997, 1319, 19–58. [Google Scholar] [CrossRef]
- Peng, Z.; Huang, W.; Gu, S.; Xie, L.; Wang, Y.; Jiao, J.; Ye, Q. Conformer: Local features coupling global representations for visual recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 367–376. [Google Scholar]
- Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
- Niu, D.; Yu, M.; Sun, L.; Gao, T.; Wang, K. Short-term multi-energy load forecasting for integrated energy systems based on CNN-BiGRU optimized by attention mechanism. Appl. Energy 2022, 313, 118801. [Google Scholar] [CrossRef]
- Rahman, A.; Srikumar, V.; Smith, A.D. Predicting electricity consumption for commercial and residential buildings using deep recurrent neural networks. Appl. Energy 2018, 212, 372–385. [Google Scholar] [CrossRef]
- Zhang, Y.; Ai, Q.; Lin, L.; Yuan, S.; Li, Z. A very short-term load forecasting method based on deep LSTM RNN at zone level. Power Syst. Technol. 2019, 43, 1884–1892. [Google Scholar]
- Cinar, Y.G.; Mirisaee, H.; Goswami, P.; Gaussier, E.; Aït-Bachir, A. Period-aware content attention RNNs for time series forecasting with missing values. Neurocomputing 2018, 312, 177–186. [Google Scholar] [CrossRef]
- Zhang, Z.; Sabuncu, M. Generalized cross entropy loss for training deep neural networks with noisy labels. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 30. [Google Scholar]
- Zhang, Z.; Zhang, Z. Artificial neural network. In Multivariate Time Series Analysis in Climate and Environmental Research; Springer: Cham, Switzerland, 2018; pp. 1–35. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Jin, X.; Lan, C.; Zeng, W.; Chen, Z.; Zhang, L. Style normalization and restitution for generalizable person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3143–3152. [Google Scholar]
- Gil, G.D.; Lashkari, A.H.; Mamun, M.; Ghorbani, A.A. Characterization of encrypted and VPN traffic using time-related features. In Proceedings of the 2nd International Conference on Information Systems Security and Privacy (ICISSP 2016), Rome, Italy, 19–21 February 2016; SciTePress: Setúbal, Portugal, 2016; pp. 407–414. [Google Scholar]
- Wang, W.; Zhu, M.; Zeng, X.; Ye, X.; Sheng, Y. Malware traffic classification using convolutional neural network for representation learning. In Proceedings of the 2017 International Conference on Information Networking (ICOIN), Da Nang, Vietnam, 11–13 January 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 712–717. [Google Scholar]
- Available online: https://stratosphereips.org/category/dataset.html (accessed on 15 June 2019).
- Available online: https://www.ixiacom.com/products/breakingpoint (accessed on 7 May 2022).
- Available online: https://www.alexa.com/topsites (accessed on 10 September 2021).
Task | Dataset | Packet | Label | Encryption Algorithm |
---|---|---|---|---|
ETVS | ISCX-VPN-Service | 60,000 | 12 | AES |
ETVA | ISCX-VPN-App | 77,163 | 17 | AES |
ETMA | USTC-TFC | 97,115 | 20 | AES RC4 3DES |
ETTA | CSTNET-TLS 1.3 | 581,709 | 120 | AES CHACHA20 |
Task | ETVS | ETVA | ETMA | ETTA | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AC | PR | RC | F1 | AC | PR | RC | F1 | AC | PR | RC | F1 | AC | PR | RC | F1 | |
FlowPrint | 0.7962 | 0.8042 | 0.7812 | 0.7820 | 0.8767 | 0.6697 | 0.6651 | 0.6531 | 0.8146 | 0.6434 | 0.7002 | 0.6573 | 0.1261 | 0.1354 | 0.1272 | 0.1116 |
AppScanner | 0.7182 | 0.7339 | 0.7225 | 0.7197 | 0.6266 | 0.4864 | 0.5198 | 0.4935 | 0.8954 | 0.8984 | 0.8968 | 0.8892 | 0.6662 | 0.6246 | 0.6327 | 0.6201 |
BIND | 0.7534 | 0.7583 | 0.7488 | 0.7420 | 0.6767 | 0.5152 | 0.5153 | 0.4965 | 0.8457 | 0.8681 | 0.8382 | 0.8396 | 0.7964 | 0.7605 | 0.7650 | 0.7560 |
DF | 0.7154 | 0.7192 | 0.7104 | 0.7102 | 0.6116 | 0.5706 | 0.4752 | 0.4799 | 0.7787 | 0.7883 | 0.7819 | 0.7593 | 0.7936 | 0.7721 | 0.7573 | 0.7602 |
FS-Net | 0.7205 | 0.7502 | 0.7238 | 0.7131 | 0.6647 | 0.4819 | 0.4848 | 0.4737 | 0.8846 | 0.8846 | 0.8920 | 0.8840 | 0.8639 | 0.8404 | 0.8349 | 0.8322 |
Deep Packet | 0.9329 | 0.9377 | 0.9306 | 0.9321 | 0.9758 | 0.9785 | 0.9745 | 0.9765 | 0.9640 | 0.9650 | 0.9631 | 0.9641 | 0.8019 | 0.4315 | 0.2689 | 0.4022 |
PERT | 0.9352 | 0.9400 | 0.9349 | 0.9368 | 0.8229 | 0.7092 | 0.7173 | 0.6992 | 0.9909 | 0.9911 | 0.9910 | 0.9911 | 0.8915 | 0.8846 | 0.8719 | 0.8741 |
ET-BERT | 0.9887 | 0.9888 | 0.9887 | 0.9887 | 0.9959 | 0.9932 | 0.9922 | 0.9927 | 0.9909 | 0.9910 | 0.9911 | 0.9910 | 0.9714 | 0.9716 | 0.9719 | 0.9717 |
BFCN | - | - | - | - | 0.9965 | 0.9936 | 0.9947 | 0.9941 | - | - | - | - | - | - | - | - |
Bi-ETC | - | - | - | - | 0.9970 | 0.9934 | 0.9951 | 0.9943 | - | - | - | - | - | - | - | - |
LAMBERT | 0.9915 | 0.9917 | 0.9915 | 0.9915 | 0.9970 | 0.9952 | 0.9952 | 0.9952 | 0.9930 | 0.9931 | 0.9930 | 0.9930 | 0.9743 | 0.9742 | 0.9742 | 0.9741 |
Class | Index | VPN-Service | VPN-App | USTC-TFC | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
BiGRU | MA | AA | FC | AC | PR | RC | F1 | AC | PR | RC | F1 | AC | PR | RC | F1 | |
LAMBERT (full mode) | ✓ | ✓ | ✓ | ✓ | 0.9915 | 0.9917 | 0.9915 | 0.9915 | 0.9970 | 0.9952 | 0.9952 | 0.9952 | 0.9930 | 0.9931 | 0.9930 | 0.9930 |
Without BiGRU | ✗ | ✓ | ✓ | ✓ | 0.9905 | 0.9906 | 0.9905 | 0.9905 | 0.9965 | 0.9929 | 0.9941 | 0.9935 | 0.9917 | 0.9918 | 0.9918 | 0.9917 |
Without Multi-head Self-Attention | ✓ | ✗ | ✓ | ✓ | 0.9898 | 0.9899 | 0.9898 | 0.9898 | 0.9966 | 0.9936 | 0.9948 | 0.9942 | 0.9914 | 0.9915 | 0.9915 | 0.9914 |
Without Additive Attention | ✓ | ✓ | ✗ | ✓ | 0.9905 | 0.9907 | 0.9905 | 0.9905 | 0.9968 | 0.9924 | 0.9955 | 0.9939 | 0.9920 | 0.9921 | 0.9920 | 0.9920 |
Without FC | ✗ | ✗ | ✗ | ✓ | 0.9887 | 0.9888 | 0.9887 | 0.9887 | 0.9959 | 0.9932 | 0.9922 | 0.9927 | 0.9909 | 0.9910 | 0.9911 | 0.9910 |
Label | Class | Count |
---|---|---|
1 | AIM_Chat | 1340 |
2 | 5000 | |
3 | 5000 | |
4 | Gmail | 5000 |
5 | Hangout | 5000 |
6 | ICQ | 823 |
7 | Netflix | 5000 |
8 | SCP | 5000 |
9 | Skype | 5000 |
10 | Spotify | 5000 |
11 | Tor | 5000 |
12 | Torrent | 5000 |
13 | Vimeo | 5000 |
14 | VoipBuster | 5000 |
15 | VPN-FTPS | 5000 |
16 | VPN-SFTP | 5000 |
17 | YouTube | 5000 |
Label | Class | Count |
---|---|---|
1 | Cridex | 5000 |
2 | Geodo | 5000 |
3 | Htbot | 5000 |
4 | Miuref | 5000 |
5 | Neris | 5000 |
6 | Nsis-ay | 823 |
7 | Shifu | 5000 |
8 | Tinba | 5000 |
9 | Virut | 5000 |
10 | Zeus | 5000 |
11 | BitTorrent | 5000 |
12 | FaceTime | 5000 |
13 | FTP | 1903 |
14 | Gmail | 5000 |
15 | MySQL | 5000 |
16 | Outlook | 5000 |
17 | Skype | 5000 |
18 | SMB | 5000 |
19 | 5000 | |
20 | WorldOfWarcraft | 5000 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, T.; Ma, X.; Liu, L.; Liu, X.; Zhao, Y.; Hu, N.; Ghafoor, K.Z. LAMBERT: Leveraging Attention Mechanisms to Improve the BERT Fine-Tuning Model for Encrypted Traffic Classification. Mathematics 2024, 12, 1624. https://doi.org/10.3390/math12111624
Liu T, Ma X, Liu L, Liu X, Zhao Y, Hu N, Ghafoor KZ. LAMBERT: Leveraging Attention Mechanisms to Improve the BERT Fine-Tuning Model for Encrypted Traffic Classification. Mathematics. 2024; 12(11):1624. https://doi.org/10.3390/math12111624
Chicago/Turabian StyleLiu, Tao, Xiting Ma, Ling Liu, Xin Liu, Yue Zhao, Ning Hu, and Kayhan Zrar Ghafoor. 2024. "LAMBERT: Leveraging Attention Mechanisms to Improve the BERT Fine-Tuning Model for Encrypted Traffic Classification" Mathematics 12, no. 11: 1624. https://doi.org/10.3390/math12111624
APA StyleLiu, T., Ma, X., Liu, L., Liu, X., Zhao, Y., Hu, N., & Ghafoor, K. Z. (2024). LAMBERT: Leveraging Attention Mechanisms to Improve the BERT Fine-Tuning Model for Encrypted Traffic Classification. Mathematics, 12(11), 1624. https://doi.org/10.3390/math12111624