Co-Training Method Based on Semi-Decoupling Features for MOOC Learner Behavior Prediction
Abstract
:1. Introduction
- (1)
- Calculate the feature importance of the dataset and rank them, and then select a number of the most important features as the shared features of the dual views;
- (2)
- Calculate the correlation coefficients between features, process the remaining features according to the correlation coefficients, and then add them to the dual views as other features. While not reducing the data and ensuring the sufficiency of the data, separate the two views as independently as possible according to the feature correlation and difference;
- (3)
- Combine shared features and “divergent” features to complete feature semi-decoupling, and then construct independent and redundant dual views for co-training.
2. Related Work
3. Materials and Methods
3.1. Co-Training
3.2. Semi-Decoupled Features
3.2.1. Feature Importance
3.2.2. Correlation Coefficient between Features and Features
- Those with weak feature correlation do not co-exist in the same view;
- Those with strong feature correlation must co-exist in the same view.
3.3. The Algorithm of Semi-Decoupled Feature Co-Training
Algorithm 1: Semi-decoupled feature co-training algorithm. |
Input: Single-view View, View1, and View2; ρmax and ρmin; w1 and w2; Test. Output: final_prediction. 1: importance = F.feature_importances_(). //F = F.drop(F.fy) 2: Sort importance. 3: Achieve View’s features sharing. 4: ρfx, ρfy = F.corr(fx, fy). //F includes the feature of fy 5: Divide remaining features. 6: Get View1, View2 with divergence. 7: for iteration = 1, 2, … in iterations: 8: predict the Test. 9: if prediction > ρmax 10: prediction = 1 11: if prediction < ρmin 12: prediction = 0 13: end if 14: both classifiers no longer change or reach a predetermined 15: number of iterative rounds. 16: end for |
4. Results
4.1. Evaluation Metrics
- F1: The combination of Precision and Recall. Precision and Recall influence each other. If Precision increases then Recall decreases; if Recall increases then Precision decreases; if both need to be balanced, then F1 measure is needed.
- Precision: The precision rate, which indicates the percent of the positive category samples that were actually positive. TP means that the original case was positive and was predicted to be positive; FP means that the original case was negative but was predicted to be positive.
- Recall: The recall rate, which also refers to as the check-all rate, means the percentage of positive class samples marked as positive. FN represents cases that were originally positive but were predicted to be negative.
4.2. Comparative Experiment
4.2.1. Test on the edX Dataset
4.2.2. Test on the Breast Cancer Wisconsin Dataset
4.3. Iteration Analysis and Discussion
4.4. Hyperparameters
5. Discussion
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Han, S.; Han, Q. Review of semi-supervised learning research. Comput. Eng. Appl. 2020, 56, 19–27. [Google Scholar]
- Liu, Y.; Zheng, Y.; Jiang, L.; Li, G.; Zhang, W. A survey on pseudo-labeling methods in deep semi-supervised learning. Comput. Sci. Explor. 2022, 15, 1–15. [Google Scholar]
- Saravanan, R.; Sujatha, P. A state of art techniques on machine learning algorithms: A perspective of supervised learning approaches in data classification. In Proceedings of the 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 14–15 June 2018; pp. 945–949. [Google Scholar]
- Guo, H.; Zhuang, X.; Chen, P. Stochastic deep collocation method based on neural architecture search and transfer learning for heterogeneous porous media. Eng. Comput. 2022. [Google Scholar] [CrossRef]
- Guo, H.; Zhuang, X.; Chen, P. Analysis of three-dimensional potential problems in non-homogeneous media with physics-informed deep collocation method using material transfer learning and sensitivity analysis. Eng. Comput. 2022. [Google Scholar] [CrossRef]
- Ghosh, I.; Chaudhuri, T.D. FEB-stacking and FEB-DNN models for stock trend prediction: A performance analysis for pre and post COVID-19 periods. Decis. Mak. Appl. Manag. Eng. 2021, 4, 51–84. [Google Scholar] [CrossRef]
- Dike, H.U.; Zhou, Y.; Deveerasetty, K.K.; Wu, Q. Unsupervised learning based on artificial neural network: A review. In Proceedings of the 2018 IEEE International Conference on Cyborg and Bionic Systems (CBS), Shenzhen, China, 25–27 October 2018; pp. 322–327. [Google Scholar]
- Li, N.; Shepperd, M.; Guo, Y. A systematic review of unsupervised learning techniques for software defect prediction. Inf. Softw. Technol. 2020, 122, 106287. [Google Scholar] [CrossRef] [Green Version]
- Madiraju, N.S.; Sadat, S.M.; Fisher, D.; Karimabadi, H. Deep temporal clustering: Fully unsupervised learning of time-domain features. arXiv 2018, arXiv:1802.01059. [Google Scholar]
- Halimi, O.; Litany, O.; Rodola, E.; Bronstein, A.; Kimmel, R. Unsupervised learning of dense shape correspondence. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4370–4379. [Google Scholar]
- Yan, K.; Huang, J.; Shen, W. Unsupervised learning for fault detection and diagnosis of air handling units. Energy Build. 2020, 210, 109689. [Google Scholar] [CrossRef]
- Bnou, K.; Hakim, A.; Raghay, S. A wavelet denoising approach based on unsupervised learning model. EURASIP J. Adv. Signal Process. 2020, 36, 1–26. [Google Scholar] [CrossRef]
- Engelen, J.V.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef] [Green Version]
- Tu, E.; Yang, J. A review of semi-supervised learning theories and recent advances. J. Shanghai Jiaotong Univ. 2018, 52, 1280. [Google Scholar]
- Schmarje, L.; Santarossa, M.; Schroder, S.M.; Koch, R. A survey on semi-, self- and unsupervised learning for image classification. IEEE Access 2021, 9, 82146–82168. [Google Scholar] [CrossRef]
- Park, S.; Lee, J.; Kim, K. Semi-supervised distributed representations of documents for sentiment analysis. Neural Netw. 2019, 119, 139–150. [Google Scholar] [CrossRef]
- Yang, Y.; Bie, R.; Wu, H. Semi-supervised power iteration clustering. Procedia Comput. Sci. 2019, 147, 588–595. [Google Scholar] [CrossRef]
- Ying, Z.; Wang, F.; Zhai, Y.; Wang, W. Semi-supervised generative adversarial network based on self-attention feature fusion for SAR target recognition. J. Signal Process. 2022, 38, 258–267. [Google Scholar]
- Kang, Z.; Peng, C.; Cheng, Q.; Liu, X.; Tian, L. Structured graph learning for clustering and semi-supervised classification. Pattern Recognit. 2021, 110, 107627. [Google Scholar] [CrossRef]
- Wu, J.; Dong, T.; Jiang, P. Protein structural classes prediction by using laplace support vector machine and based on semi-supervised method. Microcomput. Appl. 2020, 36, 4. [Google Scholar]
- Luo, S.; Wang, Z.; Wang, X. Semi-supervised soft sensor on two-subspace co-training model. J. Chem. Eng. 2022, 73, 1270–1279. [Google Scholar]
- Zhang, D.; Chen, H.; Wang, J. Summary of semi-supervised feature selection. Comput. Appl. Res. 2021, 38, 21–329. [Google Scholar]
- Zhai, X.; Oliver, A.; Kolesnikov, A.; Beyer, L. S4l: Self-supervised semi-supervised learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 1476–1485. [Google Scholar]
- Shi, B.; Sun, M.; Kao, C.; Rozgic, V.; Matsoukas, S.; Wang, C. Semi-supervised acoustic event detection based on tri-training. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 750–754. [Google Scholar]
- Sohn, K.; Berthelot, D.; Carlini, N. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; Volume 33, pp. 596–608. [Google Scholar]
- Wen, Y.; Liu, S.; Miao, Y.; Yi, X.; Liu, C. Survey on Semi-supervised classification of data streams with concept drifts. J. Softw. 2022, 33, 1287–1314. [Google Scholar]
- Li, J.; Xiong, C.; Hai, S. CoMatch: Semi-supervised learning with contrastive graph regularization. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 9475–9484. [Google Scholar]
- Ning, X.; Wang, X.; Xu, S. A review of research on co-training. Concurr. Comput. Pract. Exp. 2021, e6276. [Google Scholar] [CrossRef]
- Xing, Y.; Yu, G.; Domeniconi, C.; Wang, J.; Zhang, Z. Multi-label co-training. In Proceedings of the 27th International Joint Conference on Artificial Intelligence, Stockholm, Sweden, 13–19 July 2018; pp. 2882–2888. [Google Scholar]
- Wang, D.; Lin, Y.; Chen, J. Application of collaborative training algorithm in fault diagnosis of rolling bearing. Comput. Eng. Appl. 2020, 56, 273–278. [Google Scholar]
- Liang, Y.; Liu, Z.; Liu, W. A co-training style semi-supervised artificial neural network modeling and its application in thermal conductivity prediction of polymeric composites filled with BN sheets. Energy AI 2021, 4, 100052. [Google Scholar] [CrossRef]
- Karlos, S.; Kostopoulos, G.; Kotsiantis, S. A soft-voting ensemble based Co-training scheme using static selection for binary classification problems. Algorithms 2020, 13, 26. [Google Scholar] [CrossRef] [Green Version]
- Wang, Y.; Li, T. Improving semi-supervised Co-forest algorithm in evolving data streams. Appl. Intell. 2018, 48, 3248–3262. [Google Scholar] [CrossRef]
- Hou, F.F.; Lei, W.T.; Li, H. Impr-Co-Forest: The improved Co-forest algorithm based on optimized decision tree and dual-confidence estimation method. J. Comput. 2019, 30, 110–122. [Google Scholar]
- Zhang, M.; Zhou, Z. CoTrade: Confident Co-training with data editing. IEEE Press 2011, 41, 1612–1626. [Google Scholar]
- Liu, A.; Xu, N.; Nie, W. Benchmarking a multimodal and multiview and interactive dataset for human action recognition. IEEE Trans. Cybern. 2017, 47, 1781–1794. [Google Scholar] [CrossRef]
- Guo, W.; Yao, J.; Wang, S. Comparison and research progress of multi-view semi-supervised classification algorithms. J. Fuzhou Univ. 2021, 49, 626–637. [Google Scholar]
- Havlíček, V.; Córcoles, A.D.; Temme, K. Supervised learning with quantum-enhanced feature spaces. Nature 2019, 567, 209–212. [Google Scholar] [CrossRef] [Green Version]
- Liu, J.; Wang, G. Pharmacovigilance from social media: An improved random subspace method for identifying adverse drug events. Int. J. Med. Inform. 2018, 117, 33–43. [Google Scholar] [CrossRef] [PubMed]
- Babaei, M.; Roozbahani, A.; Shahdany, S. Risk assessment of agricultural water conveyance and delivery systems by fuzzy fault tree analysis method. Water Resour. Manag. 2018, 32, 4079–4101. [Google Scholar] [CrossRef]
- Jiang, B.; Zhang, Z.; Lin, D.; Tang, J.; Luo, B. Semi-supervised learning with graph learning-convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11313–11320. [Google Scholar]
- Wang, S.; Li, C.; Wang, R.; Liu, Z.; Wang, M.; Tan, H.; Zheng, H. Annotation-efficient deep learning for automatic medical image segmentation. Nat. Commun. 2021, 12, 5915. [Google Scholar] [CrossRef] [PubMed]
- Wang, H.; Zhu, Y.; Xu, W.; Xu, L.; Huang, Y.; Lu, W. Extraction and importance ranking of features for gait recognition. Chin. J. Med. Phys. 2019, 36, 811–817. [Google Scholar]
- Cai, J.; Luo, J.; Wang, S. Feature selection in machine learning: A new perspective. Neurocomputing 2018, 300, 70–79. [Google Scholar] [CrossRef]
- Kirasich, K.; Smith, T.; Sadler, B. Random forest vs logistic regression: Binary classification for heterogeneous datasets. SMU Data Sci. Rev. 2018, 1, 9. [Google Scholar]
- Xia, Z.; Xue, S.; Wu, L. ForeXGBoost: Passenger car sales prediction based on XGBoost. Distrib. Parallel Databases 2020, 38, 713–738. [Google Scholar] [CrossRef]
- Zou, Y.; Yu, Z.; Liu, X.; Kumar, B.V.K.V.; Wang, J. Confidence regularized self-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 5981–5990. [Google Scholar]
- Lv, J.; Li, T. A summary of semi-supervised self-training methods. J. Chongqing Norm. Univ. 2021, 38, 98–106. [Google Scholar]
- HarvardX Person-Course Academic Year 2013 De-Identified Dataset, Version 3.0. Available online: https://dataverse.harvard.edu/dataset.xhtml?persistentId=doi:10.7910/DVN/26147 (accessed on 11 January 2022).
- Sun, L.; Zhang, Q.; Zheng, Y. Learners’ online learning behavior analysis based on edX open data. Softw. Guide 2020, 19, 190–194. [Google Scholar]
- Street, W.N.; Wolberg, W.H.; Mangasarian, O.L. Nuclear feature extraction for breast tumor diagnosis. In Proceedings of the IS&T/SPIE 1993 International Symposium on Electronic Imaging: Science and Technology, San Jose, CA, USA, 31 January–5 February 1993; Volume 1905, pp. 861–870. Available online: http://archive.ics.uci.edu/ml/datasets/breast+cancer+wisconsin+(diagnostic) (accessed on 1 May 2022).
Feature | Feature Description | Feature | Feature Description |
---|---|---|---|
course_id | The id of all courses | grade | The grade of a course |
userid_DI | The id of all users | start_time_DI | The start time of registration |
registered | Whether to register for the course | last_event_DI | The last time of visit |
viewed | Whether to access the courseware | nchapters | The learning chapter |
explored | Whether to explore the process | ndays_act | The days of interaction |
certified | Whether to obtain a certificate | nforum_posts | The number of forum posts |
final_cc_cname_ | Nationality | nplay_video | The number of videos played |
LoE_DI | Academic qualifications | nevents | The number of interactions in the course |
YoB | Birthday | roles | The role that MOOC learner played |
gender | Gender (male or female) | incomplete_flag | Whether the information is filled in completely |
K-Fold Cross-Validation | Division Proportion | Metric | LR | RF | LGB | GBDT | Self-Training | Co-Training | Tri-Training | Ours |
---|---|---|---|---|---|---|---|---|---|---|
K = 2 | 5:5 | F1 | 0.7345 | 0.8944 | 0.9273 | 0.9754 | 0.9669 | 0.9772 | 0.9781 | 0.9946 |
Precision | 0.6562 | 0.8646 | 0.8995 | 0.9658 | 0.9519 | 0.9597 | 0.9712 | 0.9908 | ||
Recall | 0.8339 | 0.9263 | 0.9568 | 0.9851 | 0.9823 | 0.9953 | 0.9851 | 0.9983 | ||
K = 3 | 1:2 | F1 | 0.7351 | 0.8859 | 0.9314 | 0.9720 | 0.9686 | 0.9755 | 0.9712 | 0.9940 |
Precision | 0.6581 | 0.8550 | 0.9046 | 0.9609 | 0.9574 | 0.9569 | 0.9585 | 0.9898 | ||
Recall | 0.8326 | 0.9191 | 0.9600 | 0.9833 | 0.9801 | 0.9948 | 0.9843 | 0.9982 | ||
2:1 | F1 | 0.7350 | 0.8927 | 0.9314 | 0.9769 | 0.9684 | 0.9777 | 0.9810 | 0.9941 | |
Precision | 0.6613 | 0.8708 | 0.9046 | 0.9675 | 0.9574 | 0.9594 | 0.9719 | 0.9893 | ||
Recall | 0.8274 | 0.9158 | 0.9600 | 0.9865 | 0.9798 | 0.9966 | 0.9903 | 0.9990 | ||
K = 5 | 2:8 | F1 | 0.7326 | 0.8753 | 0.9198 | 0.9674 | 0.9590 | 0.9727 | 0.9626 | 0.9935 |
Precision | 0.6581 | 0.8497 | 0.8912 | 0.9553 | 0.9457 | 0.9536 | 0.9518 | 0.9902 | ||
Recall | 0.8261 | 0.9025 | 0.9504 | 0.9799 | 0.9727 | 0.9927 | 0.9737 | 0.9968 | ||
8:2 | F1 | 0.7350 | 0.8942 | 0.9322 | 0.9759 | 0.9685 | 0.9771 | 0.9832 | 0.9953 | |
Precision | 0.6603 | 0.8668 | 0.9028 | 0.9677 | 0.9538 | 0.9584 | 0.9754 | 0.9920 | ||
Recall | 0.8288 | 0.9234 | 0.9636 | 0.9842 | 0.9837 | 0.9965 | 0.9910 | 0.9986 |
K-Fold Cross-Validation | Train:Test | Metric | LR | RF | LGB | GBDT | Self-Training | Co-Training | Tri-Training | Ours |
---|---|---|---|---|---|---|---|---|---|---|
K = 2 | 5:5 | F1 | 0.9541 | 0.9721 | 0.9720 | 0.9625 | 0.9760 | 0.9803 | 0.9791 | 0.9859 |
Precision | 0.9506 | 0.9721 | 0.9724 | 0.9560 | 0.9674 | 0.9765 | 0.9777 | 0.9748 | ||
Recall | 0.9581 | 0.9721 | 0.9721 | 0.9693 | 0.9847 | 0.9842 | 0.9804 | 0.9974 | ||
K = 3 | 1:2 | F1 | 0.9503 | 0.9678 | 0.9644 | 0.9477 | 0.9835 | 0.9831 | 0.9630 | 0.9872 |
Precision | 0.9485 | 0.9665 | 0.9599 | 0.9313 | 0.9798 | 0.9755 | 0.9584 | 0.9747 | ||
Recall | 0.9524 | 0.9692 | 0.9692 | 0.9650 | 0.9873 | 0.9908 | 0.9678 | 1.0000 | ||
2:1 | F1 | 0.9587 | 0.9723 | 0.9767 | 0.9555 | 0.9911 | 0.9909 | 0.9696 | 0.9860 | |
Precision | 0.9439 | 0.9617 | 0.9597 | 0.9247 | 0.9875 | 0.9847 | 0.9591 | 0.9725 | ||
Recall | 0.9748 | 0.9832 | 0.9944 | 0.9888 | 0.9949 | 0.9974 | 0.9804 | 1.0000 | ||
K = 5 | 2:8 | F1 | 0.9432 | 0.9572 | 0.9536 | 0.9288 | 0.9755 | 0.9666 | 0.9578 | 0.9758 |
Precision | 0.9344 | 0.9518 | 0.9498 | 0.9072 | 0.9660 | 0.9628 | 0.9531 | 0.9553 | ||
Recall | 0.9524 | 0.9629 | 0.9580 | 0.9517 | 0.9860 | 0.9718 | 0.9629 | 0.9974 | ||
8:2 | F1 | 0.9574 | 0.9767 | 0.9794 | 0.9622 | 0.9899 | 0.9869 | 0.9767 | 0.9885 | |
Precision | 0.9491 | 0.9676 | 0.9729 | 0.9422 | 0.9900 | 0.9819 | 0.9678 | 0.9772 | ||
Recall | 0.9667 | 0.9861 | 0.9861 | 0.9833 | 0.9899 | 0.9921 | 0.9861 | 1.0000 |
K-Fold Cross-Validation | Train:Test | Metric | Iteration 1 | Iteration 2 | Iteration 3 | Iteration 4 |
---|---|---|---|---|---|---|
K = 2 | 5:5 | F1 | 0.9820 | 0.9946 | 0.9902 | 0.9891 |
Precision | 0.9729 | 0.9908 | 0.9840 | 0.9815 | ||
Recall | 0.9912 | 0.9983 | 0.9964 | 0.9969 | ||
K = 3 | 1:2 | F1 | 0.9804 | 0.9940 | 0.9892 | 0.9871 |
Precision | 0.9710 | 0.9898 | 0.9827 | 0.9785 | ||
Recall | 0.9900 | 0.9982 | 0.9957 | 0.9959 | ||
2:1 | F1 | 0.9825 | 0.9941 | 0.9906 | 0.9891 | |
Precision | 0.9735 | 0.9893 | 0.9845 | 0.9804 | ||
Recall | 0.9917 | 0.9990 | 0.9967 | 0.9980 | ||
K = 5 | 2:8 | F1 | 0.9733 | 0.9935 | 0.9860 | 0.9857 |
Precision | 0.9615 | 0.9902 | 0.9797 | 0.9791 | ||
Recall | 0.9853 | 0.9968 | 0.9923 | 0.9925 | ||
8:2 | F1 | 0.9838 | 0.9953 | 0.9899 | 0.9895 | |
Precision | 0.9763 | 0.9920 | 0.9825 | 0.9806 | ||
Recall | 0.9914 | 0.9986 | 0.9974 | 0.9986 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, H.; Xu, L.; Huang, Z.; Wang, J. Co-Training Method Based on Semi-Decoupling Features for MOOC Learner Behavior Prediction. Axioms 2022, 11, 223. https://doi.org/10.3390/axioms11050223
Wang H, Xu L, Huang Z, Wang J. Co-Training Method Based on Semi-Decoupling Features for MOOC Learner Behavior Prediction. Axioms. 2022; 11(5):223. https://doi.org/10.3390/axioms11050223
Chicago/Turabian StyleWang, Huanhuan, Libo Xu, Zhenrui Huang, and Jiagong Wang. 2022. "Co-Training Method Based on Semi-Decoupling Features for MOOC Learner Behavior Prediction" Axioms 11, no. 5: 223. https://doi.org/10.3390/axioms11050223
APA StyleWang, H., Xu, L., Huang, Z., & Wang, J. (2022). Co-Training Method Based on Semi-Decoupling Features for MOOC Learner Behavior Prediction. Axioms, 11(5), 223. https://doi.org/10.3390/axioms11050223