Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction
Abstract
:1. Introduction
- (1)
- Grounded in psychological theory and management theory of personnel performance, we constructed a high-quality dataset with performance prediction characteristics. This is key to ensuring data quantity and quality;
- (2)
- To the best of our knowledge, this is first time deep learning has been applied to the field of personnel performance prediction, which fills the gap in this field;
- (3)
- Instead of considering each attribute equally, a self-attention mechanism was used to automatically select the informative features of personnel performance. Our proposed CHRNNA framework can be viewed as a universal framework for personnel performance prediction, which has greatly improved classification performance.
2. Related Work
2.1. Personnel Performance Prediction
2.2. Deep Learning
3. Methods
- (1)
- Two kinds of neural networks, CNNs and RNNs, where the latter refers to LSTM, are employed in personnel prediction.This hybrid model combining CNN with LSTM is called the CRNN model;
- (2)
- The self-attention mechanism is introduced to a hybrid CRNN model, which aims to automatically capture the informative features;
- (3)
- The learned features, extracted from the last layer of CRNN model based on self-attention, are directly fed into the KNN classifier as inputs.
3.1. Personnel Performance Prediction Data
3.1.1. Data Description
3.1.2. Data Collection
3.1.3. Data Preprocessing
3.2. First Stage: A Hybrid CRNN Model with Self-Attention
3.2.1. CNN for Feature Extraction
3.2.2. LSTM for Contextual Information
3.2.3. Self-Attention
3.2.4. Model Training
3.3. Second Stage: Classification
- (1)
- The raw instances are propagated via our proposed network, and their feature vectors are extracted from the last layer of the hybrid CRNN model with self-attention;
- (2)
- The learned features mentioned above are fed into the KNN classifier as inputs;
- (3)
- The distances of the samples are computed, and the nearest training samples belonging to the test samples are selected;
- (4)
- The conventional KNN classification is carried out within these chosen data.
4. Experiments
4.1. Experimental Setup
4.2. Comparison Algorithms
- C4.5 is a popular and powerful decision tree classifier;
- RIPPER is a traditional rule-based classifier, whose accuracy may not be as high in most cases;
- CBA is an important association rule-based classifier, which generates all the association rules with certain support and confidence thresholds;
- CMAR is an association rule-based classifier, which uses multiple rules for prediction;
- MKNN is an approach based on local learning which first introduces a new similarity function;
- EEKNN is a performance modeling method based on data mining. A proposed improved KNN algorithm was used to handle the personnel performance prediction problem;
- CNN is a popular model which is used for feature learning;
- RNN is a deep learning model for sequential data modeling. Here, we considered it as a comparison method;
- LSTM is a variant of RNN. It is an advanced RNN architecture which prevents vanishing gradients or the phenomena of exploding;
- CHRNN is our method. It is a general framework without a self-attention mechanism;
- CHRNNA is our method. It is our final selected framework which integrates the self-attention mechanism.
4.3. Evaluation Criterion
4.4. Experimental Results
4.4.1. Experimental Results on a Personnel Performance Prediction Dataset
4.4.2. The Time Efficiency Analysis
4.5. Ablation Study of Various Modules
5. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
References
- Bal, M.; Bal, Y.; Ustundag, A. Knowledge Representation and Discovery Using Formal Concept Analysis: An HRM Application. World Congress Eng. 2011, 2, 1068–1073. [Google Scholar]
- Chien, C.-F.; Chen, L.-F. Data mining to improve personnel selection and enhance human capital: A case study in high-technology industry. Expert Syst. Appl. 2008, 2, 280–290. [Google Scholar] [CrossRef]
- Karahoca, A.; Karahoca, D.; Kaya, O. Data mining to cluster human performance by using online self regulating clustering method. In Proceedings of the Wseas International Conference on Multivariate Analysis & Its Application in Science & Engineering, Istanbul, Turkey, 27–30 May 2008. [Google Scholar]
- Gobert, J.D.; Sao Pedro, M.A.; Baker, R.S.J.D.; Toto, E.; Montalvo, O. Leveraging Educational Data Mining for Real-Time Performance Assessment of Scientific Inquiry Skills within Microworlds. J. Educ. Data Mining 2012, 4, 104–143. [Google Scholar]
- Li, N.; Kong, H.; Ma, Y.; Gong, G.; Huai, W. Human performance modeling for manufacturing based on an improved KNN algorithm. Int. J. Adv. Manuf. Technol. 2016, 84, 473–483. [Google Scholar] [CrossRef]
- Wang, Q.; Li, B.; Hu, J. Feature Selection for Human Resource Selection Based on Affinity Propagation and SVM Sensitivity Analysis. In Proceedings of the World Congress on Nature & Biologically Inspired Computing, Coimbatore, India, 9–11 December 2009. [Google Scholar]
- Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
- LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Wang, H.; Feng, J.; Zhang, Z.; Su, H.; Cui, L.; He, H.; Liu, L. Breast mass classification via deeply integrating the contextual information from multi-view data. Pattern Recognit. 2018, 80, 42–52. [Google Scholar] [CrossRef]
- Xu, X.; Li, G.; Xie, G.; Ren, J.; Xie, X. Weakly Supervised Deep Semantic Segmentation Using CNN and ELM with Semantic Candidate Regions. Complexity 2019, 2019, 9180391. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2016), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 91–99. [Google Scholar]
- Satt, A.; Rozenberg, S.; Hoory, R. Efficient Emotion Recognition from Speech Using Deep Learning on Spectrograms. In Proceedings of the 18th Annual Conference of the International Speech Communication Association (Interspeech 2017), Stockholm, Sweden, 20–24 August 2017; pp. 1089–1093. [Google Scholar]
- Bartz, C.; Herold, T.; Yang, H.; Meinel, C. Language Identification Using Deep Convolutional Recurrent Neural Networks. In Proceedings of the 24th International Conference on Neural Information Processing (ICONIP 2017), Guangzhou, China, 14–18 November 2017; pp. 880–889. [Google Scholar]
- Zhang, J.; Zhu, Y.; Zhang, X.; Ye, M.; Yang, J. Developing a Long Short-Term Memory (LSTM) based model for predicting water table depth in agricultural areas. J. Hydrol. 2018, 561, 918–929. [Google Scholar] [CrossRef]
- Sun, X. Drug-Drug Interaction Extraction via Recurrent Hybrid Convolutional Neural Networks with an Improved Focal Loss. Entropy 2019, 21, 37. [Google Scholar] [CrossRef] [Green Version]
- Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
- Barrick, M.R.; Mount, M.K. The big five personality dimensions and job performance: A meta-analysis. Personnel Psychology 1991, 44, 1–26. [Google Scholar] [CrossRef]
- Borman, W.C.; Motowidlo, S.J. Task Performance and Contextual Performance: The Meaning for Personnel Selection Research. Hum. Perform. 1997, 10, 99–109. [Google Scholar] [CrossRef]
- Cho, V.; Ngai, E.W.T. Data mining for selection of insurance sales agents. Expert Syst. 2003, 20, 123–132. [Google Scholar] [CrossRef]
- Delgado-Gómez, D.; Aguado, D.; Lopez-Castroman, J.; Cruz, C.S.; Artés-Rodríguez, A. Improving sale performance prediction using support vector machines. Expert Syst. Appl. 2011, 38, 5129–5132. [Google Scholar] [CrossRef]
- Valle, M.A.; Varas, S.; Ruz, G.A. Job performance prediction in a call center using a naive Bayes classifier. Expert Syst. Appl. 2012, 39, 9939–9945. [Google Scholar] [CrossRef]
- Thakur, G.S.; Gupta, A.; Gupta, S. Data Mining for Prediction of Human Performance Capability in the Software Industry. Int. J. Data Mining Knowl. Manag. Process 2015, 5, 53–64. [Google Scholar] [CrossRef] [Green Version]
- Sarker, A.; Shamim, S.M.; Zama, P.D.M.S.; Rahman, M.M. Employee’s Performance Analysis and Prediction Using K-means Clustering & Decision Tree Algorithm. Glob. J. Comput. Sci. Technol. 2018, 18, 1–6. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems 2012, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
- Kalchbrenner, N.; Grefenstette, E.; Blunsom, P. A Convolutional Neural Network for Modelling Sentences. In Proceedings of the Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014, Baltimore, MD, USA, 22–27 June 2014. [Google Scholar]
- Ren, Y.; Zhang, Y.; Zhang, M.; Ji, D. Context-Sensitive Twitter Sentiment Classification Using Neural Network. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
- Zhao, Z.; Yang, Z.; Luo, L.; Lin, H.; Wang, J. Drug drug interaction extraction from biomedical literature using syntax convolutional neural network. Bioinformatics 2016, 32, 3444–3453. [Google Scholar] [CrossRef]
- Ombabi, A.H.; Lazzez, O.; Ouarda, W.; Alimi, A.M. Deep learning framework based on Word2Vec and CNN for users interests classification. In Proceedings of the 2017 Sudan Conference on Computer Science and Information Technology (SCCSIT), Elnihood, Sudan, 17–19 November 2017; pp. 1–7. [Google Scholar]
- Li, X.; Ye, M.; Liu, Y.; Zhang, F.; Liu, D.; Tang, S. Accurate object detection using memory-based models in surveillance scenes. Pattern Recognit. 2017, 67, 73–84. [Google Scholar] [CrossRef]
- Yang, Y.; Zhou, J.; Ai, J.; Bin, Y.; Hanjalic, A.; Shen, H.T.; Ji, Y. Video Captioning by Adversarial LSTM. IEEE Trans. Image Process. 2018, 27, 5600–5611. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Huang, D.; Jiang, Z.; Zou, L.; Li, L. Drug–drug interaction extraction from biomedical literature using support vector machine and long short term memory networks. Inf. Sci. 2017, 415–416, 100–109. [Google Scholar] [CrossRef]
- Piech, C.; Bassen, J.; Huang, J.; Ganguli, S.; Sahami, M.; Guibas, L.; Sohl-Dickstein, J. Deep knowledge tracing. In Proceedings of the Annual Conference on Neural Information Processing Systems 2015, Montreal, QC, Canada, 7–12 December 2015; pp. 505–513. [Google Scholar]
- Su, Y.; Liu, Q.; Liu, Q.; Huang, Z.; Yin, Y.; Chen, E.; Ding, C.H.Q.; Wei, S.; Hu, G. Exercise-Enhanced Sequential Modeling for Student Performance Prediction. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI-18), New Orleans, LA, USA, 2–7 February 2018; pp. 2435–2443. [Google Scholar]
- Wei, X.; Lin, H.; Yang, L.; Yu, Y. A Convolution-LSTM-Based Deep Neural Network for Cross-Domain MOOC Forum Post Classification. Information 2017, 8, 92. [Google Scholar] [CrossRef] [Green Version]
- Zapata-Impata, B.S.; Gil, P.; Torres, F. Learning Spatio Temporal Tactile Features with a ConvLSTM for the Direction Of Slip Detection. Sensors 2019, 19, 523. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Ordóñez, F.J.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Baldominos, A.; Saez, Y.; Isasi, P. Evolutionary Design of Convolutional Neural Networks for Human Activity Recognition in Sensor-Rich Environments. Sensors 2018, 18, 1288. [Google Scholar] [CrossRef] [Green Version]
- Guo, X.; Zhang, H.; Yang, H.; Xu, L.; Ye, Z. A Single Attention-Based Combination of CNN and RNN for Relation Classification. IEEE Access 2019, 7, 12467–12475. [Google Scholar] [CrossRef]
- Salgado, J.F. The five factor model of personality and job performance in the European Community. J. Appl. Psychol. 1997, 82, 30–43. [Google Scholar] [CrossRef]
- Güngör, Z.; Serhadlioglu, G.; Kesen, S.E. A fuzzy AHP approach to personnel selection problem. Appl. Soft Comput. 2009, 9, 641–646. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. In Proceedings of the 3rd International Conference on Learning Representations (ICLR 2015), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Bhattacharya, G.; Ghosh, K.; Chowdhury, A.S. An affinity-based new local distance function and similarity measure for kNN algorithm. Pattern Recognit. Lett. 2012, 33, 356–363. [Google Scholar] [CrossRef]
Datasets | Instances | Attributes | Classes |
---|---|---|---|
Personnel Performance | 1139 | 22 | 2 |
Category | Attributes | Value | Description |
---|---|---|---|
Capability | Perception | 1–5 | 5 levels of perception |
Learning | 1–5 | 5 levels of learning | |
Memory | 1–5 | 5 levels of memory | |
Creativity | 1–5 | 5 levels of creativity | |
Speech | 1–5 | 5 levels of speech | |
Logic | 1–5 | 5 levels of logic | |
Conscientiousness | Dedication | 1–5 | 5 levels of dedication |
Engagement | 1–5 | 5 levels of engagement | |
Self control | 1–5 | 5 levels of self control | |
Decision making | 1–5 | 5 levels of decision making | |
Achievement | Goal | 1–5 | 5 levels of goal |
Initiative | 1–5 | 5 levels of initiative | |
Independence | 1–5 | 5 levels of independence | |
Confidence | 1–5 | 5 levels of confidence | |
Emotion | Emotion awareness | 1–5 | 5 levels of emotion awareness |
Emotion expression | 1–5 | 5 levels of emotion expression | |
Self-motivation | 1–5 | 5 levels of self-motivation | |
Emotion management | 1–5 | 5 levels of emotion management | |
Complementary quality | Education | 1–5 | 5 levels of education |
Health | 1–5 | 5 levels of health | |
Experience | 1–5 | 5 levels of experience | |
Stress tolerance | 1–5 | 5 levels of stress tolerance | |
Personnel performance | 0, 1 | 0, not achieve performance; 1, achieve performance |
Parameters | Parameters Value |
---|---|
epochs | 200 |
batch size | 128 |
convolution 1 layer: filters number | 13 |
convolution 2 layer: filters number | 26 |
filters size | 1 × 2 |
LSTM dimension | 64 |
Dropout rate | 0.5 |
Metrics | C4.5 | RIPPER | CBA | CMAR | MKNN | EEKNN | CNN | RNN | LSTM | Ours | |
---|---|---|---|---|---|---|---|---|---|---|---|
CHRNN | CHRNNA | ||||||||||
P | 71.47 | 82.52 | 80.86 | 57.16 | 67.22 | 71.93 | 92.14 | 89.41 | 92.06 | 96.02 | 97.07 |
R | 78.92 | 85.01 | 88.36 | 79.28 | 60.54 | 67.12 | 92.54 | 89.04 | 91.67 | 94.61 | 96.14 |
F1 | 74.32 | 83.70 | 84.11 | 60.02 | 62.68 | 69.12 | 92.34 | 89.22 | 91.86 | 95.31 | 96.60 |
C4.5 | RIPPER | CBA | CMAR | MKNN | EEKNN | CNN | RNN | LSTM | CHRNN | CHRNNA |
---|---|---|---|---|---|---|---|---|---|---|
0.82 | 5.44 | 4.52 | 11.31 | 1878.56 | 52.04 | 5.07 | 3.27 | 6.56 | 12.29 | 8.89 |
Strategy | ACC | Δ1 | P | Δ2 | R | Δ3 | F1 | Δ4 |
---|---|---|---|---|---|---|---|---|
CHRNNA (Our Model) | 96.49 | - | 97.07 | - | 96.14 | - | 96.60 | - |
- Self-attention | 95.13 | −1.36 | 96.02 | −1.05 | 94.61 | −1.53 | 95.31 | −1.29 |
- LSTM | 95.83 | −0.66 | 96.04 | −1.03 | 95.61 | −0.53 | 95.81 | −0.79 |
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xue, X.; Feng, J.; Gao, Y.; Liu, M.; Zhang, W.; Sun, X.; Zhao, A.; Guo, S. Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction. Entropy 2019, 21, 1227. https://doi.org/10.3390/e21121227
Xue X, Feng J, Gao Y, Liu M, Zhang W, Sun X, Zhao A, Guo S. Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction. Entropy. 2019; 21(12):1227. https://doi.org/10.3390/e21121227
Chicago/Turabian StyleXue, Xia, Jun Feng, Yi Gao, Meng Liu, Wenyu Zhang, Xia Sun, Aiqi Zhao, and Shouxi Guo. 2019. "Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction" Entropy 21, no. 12: 1227. https://doi.org/10.3390/e21121227
APA StyleXue, X., Feng, J., Gao, Y., Liu, M., Zhang, W., Sun, X., Zhao, A., & Guo, S. (2019). Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction. Entropy, 21(12), 1227. https://doi.org/10.3390/e21121227