Success Prediction of Crowdsourced Projects for Competitive Crowdsourced Software Development
Abstract
:1. Introduction
- An automated approach is proposed to predict the success of CSPs.
- The proposed approach minimizes training time by utilizing the BertTokenizer, thereby eliminating the need for feature engineering.
- The proposed approach is accurate, with average accuracy, precision, recall, and F1-score up to 93.75%, 93.76%, 93.75%, and 93.75%, respectively.
2. Literature Review
2.1. Success Prediction on the CCSD Platform
2.2. Recommendations for CCSD Decisions
2.3. Succcess Factors for CCSD Projects/Platforms
2.4. Simulation Methods for Failure Prediction and Task Scheduling
2.5. Quality Assessment in CCSD Platform
3. Methods
3.1. Overview
- First, it performs input formatting for BERT. Proper input formatting is essential for BERT to generate accurate and meaningful contextualized embeddings for classification. The input formatting process involves tokenizing the text input and adding special tokens to mark the beginning and end of the text. It also involves padding and truncating the input to a fixed length and creating attention masks to differentiate between padding and actual input.
- Second, it fine-tunes the BERT classifier for the project’s success prediction task.
- Finally, it predicts the success of an upcoming software project using a fine-tuned BERT classifier.
3.2. Illustrating an Example of a CCSD Project
- Project Name: “Google Drive Legal Hold Time Capsule Management Tool Prototype Challenge”.
- Project Type: “UI Prototype Competition”.
- Detailed Requirements: (a snippet from the requirements) “This challenge is to create a working HTML Prototype for the Legal Hold Time Capsule based on the provided requirements document. This is currently envisioned as working within a browser environment. It is not expected that users will access this application beyond a secure environment, and not typically on a mobile device. The clickable prototype should be accessible from traditional browsers, allowing movement among views and basic navigation, dynamically rendered on laptop; desktop screens of various sizes. (There are no device resolution constraints at this time, however, extra points will be awarded for adaptive design for device screen size flexibility). Possible future challenges related to this concept include application development using APIs.”
- Project Id: “30049261”
- Required Technologies: “CSS, HTML, and JavaScript” are the technologies specified for this project’s development.
- Required Platform: “GoogleEC2” is the platform on which the completed project will run (could be more than one).
- Prize: “[1300, 650]” in the detailed information section of a TopCoder project refers to the prize money that will be awarded to the competition winners. In this case, there will be two winners, and the first-place winner will receive a prize of USD 1300, while the second-place winner will receive a prize of USD 650.
- Current Status: “Completed” is a project status that indicates whether the project was successfully completed.
3.3. Problem Definition
3.4. Input Formatting
- (1)
- Tokenization and Adding Special Tokens: BertTokenizer splits the text of the CCSD project into a sequence of subwords, which are then mapped to integer IDs using a pre-defined vocabulary. The tokenizer also adds special tokens, i.e., CLS (classification) and SEP (separator), to mark the beginning and end of the project description, respectively. Its tokenization process can be formulated as follows:
- (2)
- Token IDs: Every w of S is converted to a numerical ID using a pre-defined vocabulary that can be formulated as:
- (3)
- Padding and Truncation: In this step of input formatting, token IDs are padded and truncated to 256 tokens. If the total number of IDs for a p is shorter than 256 tokens, it is padded with the special token ‘0’. In the case of more than 256 tokens, the remaining tokens are truncated as follows:
- (4)
- Attention Masks: Attention masks differentiate between actual tokens and padding tokens in the input sequence. This is crucial since the attention mechanism in the transformer architecture employs these masks to focus on the real tokens and disregard the padding tokens. The attention mask for input sequence can be formulated as
- (1)
- Input Text: “Google Drive Legal Hold Time Capsule Management Tool Prototype Challenge”.
- (2)
- Tokenization and Adding Special Tokens: [CLS, ‘Google’, ‘Drive’, ‘Legal’, ‘Hold’, ‘Time’, ‘Capsule’, ‘Management’, ‘Tool’, ‘Prototype’, ‘Challenge’, SEP]
- (3)
- Token IDs: [101, 8224, 3140, 3424, 2982, 2051, 13149, 3259, 6999, 9817, 4721, 102]
- (4)
- Padding and Truncation: [101, 8224, 3140, 3424, 2982, 2051, 13149, 3259, 6999, 9817, 4721, 102, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
- (5)
- Attention Masks: [101, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 102, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]
3.5. Training and Prediction
3.6. Evaluation
3.6.1. Research Questions
- RQ1. Is the proposed approach superior to the current state-of-the-art (SOTA)?
- RQ2. What is the effect, if any, of re-sampling on the proposed approach?
- RQ3. When predicting project success, is BERT more effective than other classifiers?
- RQ4. What impact does non-textual data have on the process of fine-tuning BERT?
3.6.2. Dataset
3.6.3. Process
- First, we choose the set of training projects.
- Second, the training samples are used to train and test the BERT model.
- Third, Long Short-Term Memory (LSTM), Convolutional Neural Networks (CNN), Gradient Boosting Classifier (GBC), and XGBoost Classifier (XGC) are trained on the training samples and tested on testing samples.
- As a final step, we calculate each classifier’s accuracy, precision, recall, and F1 score to compare their performance.
3.6.4. Evaluation Metrics
3.6.5. Results and Discussions
Comparison between the Proposed Approach and SOTA Deep Learning Approaches
- The BERT model outperforms other models. The proposed approach gives an accuracy of 93.75% and improves the precision, recall, and F1 score upon the baseline deep learning results by 13.46% = (93.76 − 82.64)/82.64) × 100%, 8.83% = (93.75 − 86.16)/86.16) × 100%, and 11.13% = (93.75 − 84.36)/84.36) × 100%, respectively. Moreover, the proposed approach gives an accuracy of 96.8% and improves the precision, recall, and F1 score upon the baseline deep learning results by 17.30% = (96.97 − 82.64)/82.64) × 100%, 12.29% = (96.8 − 86.16)/86.16) × 100%, and 14.80% = (96.81 − 84.36)/84.36) × 100%, respectively. One possible reason is that BERT’s ability to capture complex word relationships makes it a powerful tool in text classification.
- The lowest precision, recall, and F1 scores obtained by the proposed approach surpass the highest precision, recall, and F1 scores achieved by SOTA deep learning methods, as illustrated in Table 2. The improvement is credited to employing the BERT model and its embeddings in training LSTM and CNN classifiers. BERT, a robust language model, creates embeddings that grasp the nuanced meaning of the text. Leveraging LSTM and CNN capabilities, often used in text classification, enables these classifiers to learn complex patterns and connections in input data, extracting pertinent features for classification. Integrating BERT embeddings, with their rich semantic information, further boosts LSTM and CNN performance, resulting in exceptional outcomes for text classification tasks.
Importance of Re-Sampling
- The application of the over-sampling technique has shown improvements in accuracy, precision, recall, and F1 score. The increase in accuracy is calculated as 3.26% = (96.80 − 93.75%)/93.75%, precision as 3.42% = (96.97 − 93.76%)/93.76%, recall as 3.26% = (96.80 − 93.75%)/93.75%, and F1 score as 3.27% = (96.81 − 93.75%)/93.75%. This improvement can be attributed to BERT’s exposure to more data, allowing it to learn meaningful patterns more effectively.
- On the other hand, under-sampling reduces the number of samples in the majority class, resulting in a loss of information. Consequently, the majority and minority classes in the fine-tuned BERT exhibit lower performance when under-sampling is applied.
- By examining the confusion matrix (Figure 5), it becomes evident that the true positive and true negative rates are high when over-sampling is used. This indicates that the fine-tuned BERT correctly identifies positive and negative examples. However, when under-sampling is employed, the true positive and true negative rates are relatively low. This suggests that the model may misclassify some examples and require further optimization.
Comparison of BERT with M/DL Algorithms
- The BERT outperforms LSTM, CNN, XGC, and GBC. The possible reason for this performance improvement is that it uses pre-training on a large amount of data, which allows it to learn more meaningful and generalized representations of text data. Additionally, BERT uses attention mechanisms to capture long-range dependencies and contextual information, making it more effective in natural language understanding and processing.
- DL models, i.e., BERT, LSTM, and CNN outperform ML models, i.e., XGC and GBC. The possible reason for this improvement is that DL models can learn complex and hierarchical representations of data, leading to improved performance in tasks involving natural language understanding and processing. Additionally, DL models can handle large amounts of data and automatically learn relevant features from the data, eliminating the need for feature engineering.
- Recall of BERT is lower than other models. This is because feature-based M/DL models have a higher false positive rate that can also lead to a higher recall rate, where these models may identify some negative cases as positive. One of the key explanations for this behavior is that feature-based models, which depend on pre-defined and engineered features, may encounter challenges in effectively capturing intricate contextual relationships and nuanced patterns [53]. Often depending on handcrafted features, these models might misclassify instances not fitting well within their predefined feature space [53]. For instance, in sentiment analysis, such models might base sentiment polarity on simple features like word presence and miss contextual nuances where words convey varied sentiments based on their surrounding context. This oversight could lead to mislabeling instances, impacting the model’s recall. On the other hand, BERT, being a contextual language model, considers the entire context of words and sentences due to its bidirectional nature and deep contextual embeddings [32]. It can capture intricate relationships between words and their context, resulting in more accurate classifications and potentially a lower recall than feature-based models. As shown in Figure 6 LSTM and CNN are predicting 92.7%(450) of unsuccessful (485) projects as successful. Whereas, BERT is predicting 21.6%(525) of unsuccessful (2429) projects as successful.
- XGC achieves higher precision than LSTM, CNN, and GBC as shown in Figure 7. This is because it has the ability to handle imbalanced datasets, minimize false positives, and prevent overfitting using regularization techniques. Its gradient-boosting algorithm assigns higher weights to misclassified instances, reducing false positives and improving precision.
Impact of Non-Textual Inputs
- Disabling non-textual features decreases accuracy, precision, recall, and F1 score by 0.10%, 0.10%, 0.10%, and 0.09%, respectively.
- Disabling non-text features in the proposed approach resulted in slightly lower scores across all metrics compared to the default settings. This outcome is attributed to BERT’s ability to learn meaningful patterns and relationships in numerical and textual data, thanks to its pre-training on diverse data types. Unlike traditional machine learning models, BERT’s deep neural network architecture allows it to handle complex relationships. It performs well on a combination of alpha-numeric and textual data. Still, it is unsuitable for handling only numerical data, as it requires a mix of text and non-textual data for optimal performance.
3.6.6. Threats to Validity
4. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Storey, M.A.; Zagalsky, A.; Figueira Filho, F.; Singer, L.; German, D.M. How social and communication channels shape and challenge a participatory culture in software development. IEEE Trans. Softw. Eng. 2016, 43, 185–204. [Google Scholar] [CrossRef]
- Mao, K.; Capra, L.; Harman, M.; Jia, Y. A survey of the use of crowdsourcing in software engineering. J. Syst. Softw. 2017, 126, 57–84. [Google Scholar] [CrossRef]
- Dwarakanath, A.; Chintala, U.; Shrikanth, N.; Virdi, G.; Kass, A.; Chandran, A.; Sengupta, S.; Paul, S. Crowd build: A methodology for enterprise software development using crowdsourcing. In Proceedings of the 2015 IEEE/ACM 2nd International Workshop on CrowdSourcing in Software Engineering, Florence, Italy, 19 May 2015; pp. 8–14. [Google Scholar]
- Illahi, I.; Liu, H.; Umer, Q.; Niu, N. Machine learning based success prediction for crowdsourcing software projects. J. Syst. Softw. 2021, 178, 110965. [Google Scholar] [CrossRef]
- Tunio, M.Z.; Luo, H.; Cong, W.; Fang, Z.; Gilal, A.R.; Abro, A.; Wenhua, S. Impact of personality on task selection in crowdsourcing software development: A sorting approach. IEEE Access 2017, 5, 18287–18294. [Google Scholar] [CrossRef]
- Fu, Y.; Sun, H.; Ye, L. Competition-aware task routing for contest based crowdsourced software development. In Proceedings of the 2017 6th International Workshop on Software Mining (SoftwareMining), Urbana, IL, USA, 3 November 2017; pp. 32–39. [Google Scholar]
- Brabham, D.C. Moving the crowd at Threadless: Motivations for participation in a crowdsourcing application. Inf. Commun. Soc. 2010, 13, 1122–1145. [Google Scholar] [CrossRef]
- Dwarakanath, A.; Shrikanth, N.; Abhinav, K.; Kass, A. Trustworthiness in enterprise crowdsourcing: A taxonomy & evidence from data. In Proceedings of the 38th International Conference on Software Engineering Companion, Austin, TX, USA, 14–22 May 2016; pp. 41–50. [Google Scholar]
- Boehm, B.W. Software Engineering Economics; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
- Beecham, S.; Baddoo, N.; Hall, T.; Robinson, H.; Sharp, H. Motivation in Software Engineering: A systematic literature review. Inf. Softw. Technol. 2008, 50, 860–878. [Google Scholar] [CrossRef]
- Sun, Y.; Wang, N.; Peng, Z. Working for one penny: Understanding why people would like to participate in online tasks with low payment. Comput. Hum. Behav. 2011, 27, 1033–1041. [Google Scholar] [CrossRef]
- Kaufmann, N.; Schulze, T.; Veit, D. More Than Fun and Money. Worker Motivation in Crowdsourcing—A Study on Mechanical Turk; University of Mannhein: Mannhein, Germany, 2011. [Google Scholar]
- Martinez, M.G.; Walton, B. The wisdom of crowds: The potential of online communities as a tool for data analysis. Technovation 2014, 34, 203–214. [Google Scholar] [CrossRef]
- Dubey, A.; Abhinav, K.; Taneja, S.; Virdi, G.; Dwarakanath, A.; Kass, A.; Kuriakose, M.S. Dynamics of software development crowdsourcing. In Proceedings of the 2016 IEEE 11th International Conference on Global Software Engineering (ICGSE), Orange County, CA, USA, 2–5 August 2016; pp. 49–58. [Google Scholar]
- Fitzgerald, B.; Stol, K.J. The dos and dont’s of crowdsourcing software development. In Proceedings of the International Conference on Current Trends in Theory and Practice of Informatics, Snezkou, Czech Republic, 24–29 January 2015; pp. 58–64. [Google Scholar]
- Khanfor, A.; Yang, Y.; Vesonder, G.; Ruhe, G.; Messinger, D. Failure prediction in crowdsourced software development. In Proceedings of the 2017 24th Asia-Pacific Software Engineering Conference (APSEC), Nanjing, China, 4–8 December 2017; pp. 495–504. [Google Scholar]
- Afridi, H.G. Empirical investigation of correlation between rewards and crowdsource-based software developers. In Proceedings of the 2017 IEEE/ACM 39th International Conference on Software Engineering Companion (ICSE-C), Buenos Aires, Argentina, 20–28 May 2017; pp. 80–81. [Google Scholar]
- Illahi, I.; Liu, H.; Umer, Q.; Zaidi, S.A.H. An empirical study on competitive crowdsource software development: Motivating and inhibiting factors. IEEE Access 2019, 7, 62042–62057. [Google Scholar] [CrossRef]
- Yang, Y.; Karim, M.R.; Saremi, R.; Ruhe, G. Who should take this task? Dynamic decision support for crowd workers. In Proceedings of the 10th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement, Ciudad Real, Spain, 8–9 September 2016; pp. 1–10. [Google Scholar]
- Stol, K.J.; Fitzgerald, B. Two’s company, three’s a crowd: A case study of crowdsourcing software development. In Proceedings of the 36th International Conference on Software Engineering, Hyderabad, India, 31 May–7 June 2014; pp. 187–198. [Google Scholar]
- Huang, Y.; Nazir, S.; Wu, J.; Hussain Khoso, F.; Ali, F.; Khan, H.U. An efficient decision support system for the selection of appropriate crowd in crowdsourcing. Complexity 2021, 2021, 5518878. [Google Scholar] [CrossRef]
- Yin, X.; Huang, J.; He, W.; Guo, W.; Yu, H.; Cui, L. Group task allocation approach for heterogeneous software crowdsourcing tasks. Peer Peer Netw. Appl. 2021, 14, 1736–1747. [Google Scholar] [CrossRef]
- Yuen, M.C.; King, I.; Leung, K.S. Temporal context-aware task recommendation in crowdsourcing systems. Knowl. Based Syst. 2021, 219, 106770. [Google Scholar] [CrossRef]
- Wang, J.; Yang, Y.; Wang, S.; Hu, J.; Wang, Q. Context-and Fairness-Aware In-Process Crowdworker Recommendation. ACM Trans. Softw. Eng. Methodol. TOSEM 2022, 31, 1–31. [Google Scholar] [CrossRef]
- Messinger, D. Elements of Good Crowdsourcing. In Proceedings of the 3rd International Workshop, Austin, TX, USA, 16 May 2016. [Google Scholar]
- Saremi, R.; Yang, Y.; Vesonder, G.; Ruhe, G.; Zhang, H. Crowdsim: A hybrid simulation model for failure prediction in crowdsourced software development. arXiv 2021, arXiv:2103.09856. [Google Scholar]
- Saremi, R.; Yagnik, H.; Togelius, J.; Yang, Y.; Ruhe, G. An evolutionary algorithm for task scheduling in crowdsourced software development. arXiv 2021, arXiv:2107.02202. [Google Scholar]
- Hu, Z.; Wu, W.; Luo, J.; Wang, X.; Li, B. Quality assessment in competition-based software crowdsourcing. Front. Comput. Sci. 2020, 14, 1–14. [Google Scholar] [CrossRef]
- Jung, H.J. Quality assurance in crowdsourcing via matrix factorization based task routing. In Proceedings of the 23rd International Conference on World Wide Web, Seoul, Republic of Korea, 7–11 April 2014; pp. 3–8. [Google Scholar]
- Amelio, A.; Bonifazi, G.; Corradini, E.; Di Saverio, S.; Marchetti, M.; Ursino, D.; Virgili, L. Defining a deep neural network ensemble for identifying fabric colors. Appl. Soft Comput. 2022, 130, 109687. [Google Scholar] [CrossRef]
- Anceschi, E.; Bonifazi, G.; De Donato, M.C.; Corradini, E.; Ursino, D.; Virgili, L. Savemenow. AI: A machine learning based wearable device for fall detection in a workplace. In Enabling Applications in Data Science; Springer: Berlin/Heidelberg, Germany, 2021; pp. 493–514. [Google Scholar]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
- Mazzola, E.; Piazza, M.; Perrone, G. How do different network positions affect crowd members’ success in crowdsourcing challenges? J. Prod. Innov. Manag. 2023, 40, 276–296. [Google Scholar] [CrossRef]
- Yin, X.; Wang, H.; Wang, W.; Zhu, K. Task recommendation in crowdsourcing systems: A bibliometric analysis. Technol. Soc. 2020, 63, 101337. [Google Scholar] [CrossRef]
- Wang, J.; Yang, Y.; Wang, S.; Chen, C.; Wang, D.; Wang, Q. Context-aware personalized crowdtesting task recommendation. IEEE Trans. Softw. Eng. 2021, 48, 3131–3144. [Google Scholar] [CrossRef]
- He, H.R.; Liu, Y.; Gao, J.; Jing, D. Investigating Business Sustainability of Crowdsourcing Platforms. IEEE Access 2022, 10, 74291–74303. [Google Scholar] [CrossRef]
- Borst, I. Understanding Crowdsourcing: Effects of Motivation and Rewards on Participation and Performance in Voluntary online Activities; Number EPS-2010-221-LIS; Erasmus University Rotterdam: Rotterdam, The Netherlands, 2010. [Google Scholar]
- Yang, Y.; Saremi, R. Award vs. worker behaviors in competitive crowdsourcing tasks. In Proceedings of the 2015 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM), Beijing, China, 22–23 October 2015; pp. 1–10. [Google Scholar]
- Kamar, E.; Horvitz, E. Incentives for truthful reporting in crowdsourcing. AAMAS 2012, 12, 1329–1330. [Google Scholar]
- Machado, L.; Melo, R.; Souza, C.; Prikladnicki, R. Collaborative Behavior and Winning Challenges in Competitive Software Crowdsourcing. Proc. ACM Hum. Comput. Interact. 2021, 5, 1–25. [Google Scholar] [CrossRef]
- Al Haqbani, O.; Alyahya, S. Supporting Coordination among Participants in Crowdsourcing Software Design. In Proceedings of the 2022 IEEE/ACIS 20th International Conference on Software Engineering Research, Management and Applications (SERA), Las Vegas, NV, USA, 22–25 May 2022; pp. 132–139. [Google Scholar]
- Alabdulaziz, M.S.; Hassan, H.F.; Soliman, M.W. The effect of the interaction between crowdsourced style and cognitive style on developing research and scientific thinking skills. EURASIA J. Math. Sci. Technol. Educ. 2022, 18, em2162. [Google Scholar] [CrossRef] [PubMed]
- Xu, H.; Wu, Y.; Hamari, J. What determines the successfulness of a crowdsourcing campaign: A study on the relationships between indicators of trustworthiness, popularity, and success. J. Bus. Res. 2022, 139, 484–495. [Google Scholar] [CrossRef]
- Feng, Y.; Yi, Z.; Yang, C.; Chen, R.; Feng, Y. How do gamification mechanics drive solvers’ Knowledge contribution? A study of collaborative knowledge crowdsourcing. Technol. Forecast. Soc. Change 2022, 177, 121520. [Google Scholar] [CrossRef]
- Shi, X.; Evans, R.D.; Shan, W. What Motivates Solvers’ Participation in Crowdsourcing Platforms in China? A Motivational–Cognitive Model. IEEE Trans. Eng. Manag. 2022, 1–13. [Google Scholar] [CrossRef]
- Mejorado, D.M.; Saremi, R.; Yang, Y.; Ramirez-Marquez, J.E. Study on patterns and effect of task diversity in software crowdsourcing. In Proceedings of the 14th ACM/IEEE International Symposium on Empirical Software Engineering and Measurement (ESEM), Bari, Italy, 5–9 October 2020; pp. 1–10. [Google Scholar]
- Urbaczek, J.; Saremi, R.; Saremi, M.L.; Togelius, J. Scheduling tasks for software crowdsourcing platforms to reduce task failure. arXiv 2020, arXiv:2006.01048. [Google Scholar]
- Wu, W.; Tsai, W.T.; Li, W. An evaluation framework for software crowdsourcing. Front. Comput. Sci. 2013, 7, 694–709. [Google Scholar] [CrossRef]
- Sarzynska-Wawer, J.; Wawer, A.; Pawlak, A.; Szymanowska, J.; Stefaniak, I.; Jarkiewicz, M.; Okruszek, L. Detecting formal thought disorder by deep contextualized word representations. Psychiatry Res. 2021, 304, 114135. [Google Scholar] [CrossRef] [PubMed]
- Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
- Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
- Joulin, A.; Grave, E.; Bojanowski, P.; Mikolov, T. Bag of tricks for efficient text classification. arXiv 2016, arXiv:1607.01759. [Google Scholar]
- Choi, Y.; Cardie, C.; Riloff, E.; Patwardhan, S. Identifying sources of opinions with conditional random fields and extraction patterns. In Proceedings of the Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing, Vancouver, BC, Canada, 10–12 October 2005; pp. 355–362. [Google Scholar]
- Jo, H.; Bang, Y. Factors influencing continuance intention of participants in crowdsourcing. Humanit. Soc. Sci. Commun. 2023, 10, 824. [Google Scholar] [CrossRef]
No. of CSPs | 16,894 |
---|---|
Successful CSPs | 14,465 (85.61%) |
Unsuccessful CSPs | 2429 (14.39%) |
CSPs Classification | 2 |
Minimum Words in CSPs Text | 1 |
Maximum Words in CSPs Text | 3214 |
Duration of Projects | Upto July 2018 |
Model | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
BERT | 93.75 | 93.76 | 93.75 | 93.75 |
LSTM | 86.62 | 86.54 | 99.93 | 92.75 |
CNN | 86.56 | 86.53 | 99.86 | 92.72 |
ML-SP (Baseline) | — | 82.64 | 86.16 | 84.36 |
Re-Sampling | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
No (Default) | 93.75 | 93.76 | 93.75 | 93.75 |
Yes (Under-Sampling) | 90.26 | 90.27 | 90.26 | 91.28 |
Yes (Over-Sampling) | 96.80 | 96.97 | 96.80 | 96.81 |
Model Name | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
BERT | 93.75 | 93.76 | 93.75 | 93.75 |
LSTM | 86.62 | 86.54 | 99.93 | 92.75 |
CNN | 86.56 | 86.53 | 99.86 | 92.72 |
GradientBoost | 86.5 | 86.54 | 99.76 | 92.68 |
XGBoost | 86.21 | 87.91 | 97.27 | 92.36 |
Inputs | Accuracy | Precision | Recall | F1 Score |
---|---|---|---|---|
Default | 93.75 | 93.76 | 93.75 | 93.75 |
Textual Only | 93.66 | 93.67 | 93.66 | 93.67 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Rashid, T.; Anwar, S.; Jaffar, M.A.; Hakami, H.; Baashirah, R.; Umer, Q. Success Prediction of Crowdsourced Projects for Competitive Crowdsourced Software Development. Appl. Sci. 2024, 14, 489. https://doi.org/10.3390/app14020489
Rashid T, Anwar S, Jaffar MA, Hakami H, Baashirah R, Umer Q. Success Prediction of Crowdsourced Projects for Competitive Crowdsourced Software Development. Applied Sciences. 2024; 14(2):489. https://doi.org/10.3390/app14020489
Chicago/Turabian StyleRashid, Tahir, Shumaila Anwar, Muhammad Arfan Jaffar, Hanadi Hakami, Rania Baashirah, and Qasim Umer. 2024. "Success Prediction of Crowdsourced Projects for Competitive Crowdsourced Software Development" Applied Sciences 14, no. 2: 489. https://doi.org/10.3390/app14020489
APA StyleRashid, T., Anwar, S., Jaffar, M. A., Hakami, H., Baashirah, R., & Umer, Q. (2024). Success Prediction of Crowdsourced Projects for Competitive Crowdsourced Software Development. Applied Sciences, 14(2), 489. https://doi.org/10.3390/app14020489