Subject-Specialized Chatbot in Higher Education as a Tutor for Autonomous Exam Preparation: Analysis of the Impact on Academic Performance and Students’ Perception of Its Usefulness
Abstract
:1. Introduction
- RQ1: How effective is the specialized chatbot in resolving students’ doubts and delivering clear and precise explanations to support exam preparation?
- RQ2: What is the frequency and patterns of chatbot use among students, and what types of questions do they typically ask?
- RQ3: How do students and teachers perceive the effectiveness and usefulness of the chatbot as a complementary learning tool?
2. Materials and Methods
2.1. Participants
2.2. Research Procedure
2.2.1. Pre-Intervention Phase
- Ease of use: Chatbase provides an intuitive interface, allowing chatbot creation without advanced programming knowledge. This facilitated the quick setup and customization of the chatbot for the educational context.
- Compatibility with advanced AI: The tool uses natural language processing (NLP) technologies based on advanced models, such as GPT, which enabled the chatbot to answer complex questions coherently and accurately.
- Integration with knowledge bases: Chatbase allows the chatbot to be trained with specific course files and documents.
- Interaction recording: The platform logs and analyzes all user interactions with the chatbot, which proved valuable for subsequent data analysis.
2.2.2. Intervention Phase
2.2.3. Post-Intervention Phase
2.3. Data Collection Instruments
2.3.1. Instruments for Evaluating the Chatbot’s Quality
- Response Accuracy (≥90%): The chatbot must provide correct answers 90% of the time, ensuring that responses are largely accurate and contribute to quality learning.
- Response Consistency (≥95%): A 95% consistency rate ensures that responses are aligned with repeated concepts in different contexts, avoiding student confusion.
- Interaction Fluency (≥85%): Fluent conversation is critical for a positive experience, with 85% representing a comprehensible and natural interaction in most cases.
- Problem-Solving Ability (≥90%): The chatbot should effectively resolve complex questions at least 90% of the time, ensuring it can guide students in understanding difficult concepts.
- Exam Preparation Effectiveness (≥85%): To be perceived as useful, the chatbot must effectively help students prepare for exams, achieving an 85% user satisfaction rate.
- Reliability (100%): The chatbot is expected to have 100% availability, ensuring uninterrupted operation throughout the usage period.
2.3.2. Pre- and Post-Intervention Surveys
2.3.3. In-Depth Interviews
2.3.4. Chatbot Interaction Analysis
2.4. Ethical Considerations
3. Results
3.1. Overall Chatbot Performance
- Response Accuracy: The chatbot achieved 98% accuracy, surpassing the methodological threshold of ≥90%. This indicator assessed the system’s ability to provide correct answers to simple questions such as true/false or concept definitions. Zero-shot prompting was used for these types of questions, a technique where the chatbot generated answers without prior examples. This approach was effective for direct inquiries and required no additional adjustments to improve accuracy in this context.
- Response Consistency: Response consistency was measured based on the uniformity of responses in different contexts, especially for repeated questions or those involving related concepts. The chatbot achieved 96% consistency, exceeding the minimum standard of ≥95%. This result was attained through the application of few-shot prompting, where specific examples were provided to guide responses. This technique significantly improved response alignment in multiple-choice scenarios and questions that required concept matching.
- Interaction Fluency: The chatbot scored 89% in interaction fluency, meeting the methodological threshold of ≥85%. This metric was assessed based on the chatbot’s ability to maintain understandable and natural interactions, particularly in open-ended or unstructured conversations. Chain-of-thought prompting (CoT) was used in the evaluation, allowing the chatbot to break down its reasoning into clear steps, improving the structure and clarity of responses in essay-type or more complex questions.
- Problem-Solving Ability: The chatbot’s problem-solving ability, assessed in terms of its skill in answering complex questions and guiding the understanding of difficult concepts, reached 92%, surpassing the threshold of ≥90%. This result was achieved through a combination of few-shot prompting and CoT prompting, techniques that enabled the chatbot to provide more detailed and explanatory answers. These iterative adjustments to the chatbot’s configuration enhanced its ability to address deeper conceptual issues, aligning with the established pedagogical goals.
- Exam-Preparation Effectiveness: The research team also evaluated the chatbot’s effectiveness in exam preparation, achieving an 88% satisfaction rate, exceeding the ≥85% standard. This indicator measured the perceived usefulness of the chatbot in content review and exam simulations. Few-shot prompting and CoT prompting techniques were essential to ensure the chatbot provided precise and relevant feedback, helping students better prepare for their academic assessments.
- Reliability: The chatbot demonstrated 100% reliability during the lab testing period, meeting the proposed standard of full availability. This metric evaluated the system’s ability to remain operational without interruptions, even under high-demand conditions. The absence of failures or downtime during testing ensured that the system could respond continuously and stably, solidifying its technical robustness.
3.2. Analysis of Chatbot Interactions: Most Consulted Content Types and Relationship with Learning Outcomes
3.3. Perceptions of the Chatbot’s Effectiveness and Usefulness
Overall Perception of the Chatbot’s Usefulness
4. Discussion and Conclusions
5. Limitations and Future Research
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Abas, M. A., Arumugam, S. E., Yunus, M. M., & Rafiq, K. R. M. (2023). ChatGPT and personalized learning: Opportunities and challenges in higher education. International Journal of Academic Research in Business and Social Sciences, 13, 3936–3945. [Google Scholar] [CrossRef] [PubMed]
- Agbong-Coates, I. J. (2024). ChatGPT integration significantly boosts personalized learning outcomes: A Philippine study. International Journal of Educational Management and Development Studies, 5, 165–186. [Google Scholar] [CrossRef]
- Alenezi, M. (2023). Digital learning and digital institution in higher education. Education Sciences, 13, 88. [Google Scholar] [CrossRef]
- Boone, H. N., & Boone, D. A. (2012). Analyzing likert data. Journal of Extension, 50(2), 48. [Google Scholar] [CrossRef]
- Chang, D. H., Lin, M. P., Hajian, S., & Wang, Q. Q. (2023). Educational design principles of using AI chatbot that supports self-regulated learning in education: Goal setting, feedback, and personalization. Sustainability, 15, 12921. [Google Scholar] [CrossRef]
- Chen, K., Tallant, A. C., & Selig, I. (2024). Exploring generative AI literacy in higher education: Student adoption, interaction, evaluation, and ethical perceptions. Information and Learning Science. ahead-of-print. [Google Scholar] [CrossRef]
- Chukwuere, J. E. (2024). The future of generative AI chatbots in higher education. arXiv. [Google Scholar] [CrossRef]
- Cramarenco, R., Burcă-Voicu, M., & Dabija, D. (2023). Student perceptions of online education and digital technologies during the COVID-19 pandemic: A systematic review. Electronics, 12, 319. [Google Scholar] [CrossRef]
- Dabis, A., & Csáki, C. (2024). AI and ethics: Investigating the first policy responses of higher education institutions to the challenge of generative AI. Humanities & Social Sciences Communications, 11, 1006. [Google Scholar] [CrossRef]
- Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. Management Information Systems Research Center, 13(3), 319–340. [Google Scholar] [CrossRef]
- Deng, X., & Yu, Z. (2023). A meta-analysis and systematic review of the effect of chatbot technology use in sustainable education. Sustainability, 15, 2940. [Google Scholar] [CrossRef]
- Dhawan, S. (2022). Higher education quality and student satisfaction: Meta-analysis subgroup analysis and meta-regression. Metamorphosis: A Journal of Management Research, 21(1), 48–66. [Google Scholar] [CrossRef]
- Diederich, S., Lembcke, T., Brendel, A. B., & Kolbe, L. M. (2021). Understanding the impact that response failure has on how users perceive anthropomorphic conversational service agents: Insights from an online experiment. AIS Transactions on Human-Computer Interaction, 13(1), 82–103. [Google Scholar] [CrossRef]
- España. (2018). Ley orgánica 3/2018, de 5 de diciembre, de protección de datos personales y garantía de los derechos digitales. Boletín Oficial del Estado, 294, 119788–119857. Available online: https://www.boe.es/buscar/act.php?id=BOE-A-2018-16673 (accessed on 1 September 2024).
- Følstad, A., Nordheim, C. B., & Bjørkli, C. A. (2018, September 24). What makes users trust a chatbot for customer service? An exploratory interview study. International Conference on Internet Science (pp. 194–208), St. Petersburg, Russia. [Google Scholar] [CrossRef]
- Guttierrez-Aguilar, O., Huarsaya-Rodriguez, E., & Duche-Pérez, A. (2024). The mediating effect of academic performance on ChatGPT satisfaction in university students. In G. F. Olmedo Cifuentes, D. G. Arcos Avilés, & H. V. Lara Padilla (Eds.), Emerging research in intelligent systems (Volume 903, pp. 155–168). Springer. [Google Scholar] [CrossRef]
- Halaweh, M. (2023). ChatGPT in education: Strategies for responsible implementation. Contemporary Educational Technology, 15, ep421. [Google Scholar] [CrossRef]
- Hang, C. N., Wei Tan, C., & Yu, P. -D. (2024). MCQGen: A large language model-driven mcq generator for personalized learning. IEEE Access, 12, 102261–102273. [Google Scholar] [CrossRef]
- Holmes, W., & Tuomi, I. (2022). State of the art and practice in AI in education. European Journal of Education, 57(4), 542–570. [Google Scholar] [CrossRef]
- Holmes, W., Bialik, M., & Fadel, C. (2019). Artificial intelligence in education: Promises and implications for teaching and learning. Centre for Curriculum Redesign. [Google Scholar]
- Joshi, A., Kale, S., Chandel, S., & Pal, D. K. (2015). Likert scale: Explored and explained. Current Journal of Applied Science and Technology, 7(4), 396–403. [Google Scholar] [CrossRef]
- Kasneci, E., Sessler, K., Küchemann, S., Bannert, M., Dementieva, D., Fischer, F., & Gasser, U. (2023). ChatGPT for good? On opportunities and challenges of large language models for education. Learning and Individual Differences, 103, 102274. [Google Scholar] [CrossRef]
- Knoth, N., Tolzin, A., Janson, A., & Leimeister, J. M. (2024). AI literacy and its implications for prompt engineering strategies. Computer and Education: Artificial Intelligence, 6, 100225. [Google Scholar] [CrossRef]
- Kvale, S. (2007). Doing interviews. SAGE Publications. [Google Scholar] [CrossRef]
- Labadze, L., Grigolia, M., & Machaidze, L. (2023). Role of AI chatbots in education: Systematic literature review. International Journal of Educational Technology in Higher Education, 20, 56. [Google Scholar] [CrossRef]
- Lin, M. P., & Chang, D. (2023). CHAT-ACTS: A pedagogical framework for personalized chatbot to enhance active learning and self-regulated learning. Computer and Education: Artificial Intelligence, 5, 100167. [Google Scholar] [CrossRef]
- Martins, L., Fernández-Ferrer, M., & Puertas, E. (2024). Analysing conversation pathways with a chatbot tutor to enhance self-regulation in higher education. Education Sciences, 14, 590. [Google Scholar] [CrossRef]
- Mat Dangi, M. R., & Mohamed Saat, M. (2021). 21st-century educational technology adoption in accounting education: Does institutional support moderate accounting educators’ acceptance behavior? International Journal Academic Research in Business and Social Sciences, 11(1), 8288. [Google Scholar] [CrossRef] [PubMed]
- Mironova, J., Riashchenko, V., Kinderis, R., Djakona, V., & Dimitrova, S. (2024). Ethical concerns in using generative tools in higher education: Cross-country study. Environment Technology Resources, 2, 444–447. [Google Scholar] [CrossRef]
- Mungai, B. K., Omieno, P. K. K., Egessa, M., & Manyara, P. N. (2024). AI chatbots in LMS: A pedagogical review of cognitive, constructivist, and adaptive principles. Engineering and Technology Journal, 9(8), 4709–4715. [Google Scholar] [CrossRef]
- Ng, D. T. K., & Leung, J. K. L. (2024). Empowering student self-regulated learning and science education through ChatGPT: A pioneering pilot study. British Journal of Educational Technology, 55(4), 1328–1353. [Google Scholar] [CrossRef]
- Patton, M. Q. (2015). Qualitative research & evaluation methods (4th ed.). SAGE Publications. [Google Scholar]
- Rahman, M. M., & Watanobe, Y. (2023). ChatGPT for education and research: Opportunities, threats, and strategies. Applied Sciences, 13(9), 5783. [Google Scholar] [CrossRef]
- Sánchez-Vera, F. (2024). Developing effective educational chatbots with GPT: Insights from a pilot study in a university subject. Trends in Higher Education, 3(1), 155–168. [Google Scholar] [CrossRef]
- Schei, O. M., Møgelvang, A., & Ludvigsen, K. (2024). Perceptions and use of AI chatbots among students in higher education: A scoping review of empirical studies. Education Science, 14, 922. [Google Scholar] [CrossRef]
- Schwab, K. (2017). The fourth industrial revolution. World Economic Forum. [Google Scholar]
- Su, J. H., & Yang, W. P. (2023). Unlocking the power of ChatGPT: A framework for applying generative AI in education. ECNU Review of Education, 6(3), 355–366. [Google Scholar] [CrossRef]
- Tan, C. W. (2023). Large language model-driven classroom flipping: Empowering student-centric peer questioning with flipped interaction. arXiv. [Google Scholar] [CrossRef]
- Valverde-Berrocoso, J., Acevedo-Borrega, J., & Cerezo-Pizarro, M. (2022). Educational technology and student performance: A systematic review. Frontier in Education, 7, 916502. [Google Scholar] [CrossRef]
- Winkler, R., & Söllner, M. (2018). Unleashing the potential of chatbots in education: A state-of-the-art analysis. Academy of Management Annual Meeting (AOM), 15, 44. [Google Scholar] [CrossRef]
- Yigci, D., Eryilmaz, M., Yetisen, A. K., Tasoglu, S., & Ozcan, A. (2024). Large language model-based chatbots in higher education. Advanced Intelligent Systems, 2400429. [Google Scholar] [CrossRef]
- Zhang, X., Jiang, H., Qiao, Z., & Li, P. (2024). Students’ response to ChatGPT: An adaptive technology-to-performance model. Journal of Computer Information Systems, 1–18. [Google Scholar] [CrossRef]
Category | Description | Number of Interactions | Percentage (%) |
---|---|---|---|
Definitions | Queries to clarify key terms and concepts | 244 | 34.66% |
Detailed Explanations | Requests for in-depth information on theories and frameworks | 157 | 22.30% |
Practical Applications | Questions about applying pedagogical theories in real contexts | 105 | 14.91% |
Comparisons and Differences | Queries to contrast concepts or methodologies | 87 | 12.36% |
Regulations and Legislation | Questions about educational laws and regulations | 61 | 8.66% |
Other Queries | Technical questions or requests for additional resources | 50 | 7.10% |
Usage (Interaction Range) | Number of Students | Average Interactions | Final Exam Average Score (Range 1 to 10) |
---|---|---|---|
Low (1–10) | 12 | 9.83 | 4.92 |
Moderate (11–20) | 26 | 13.15 | 6.58 |
High (21–30) | 2 | 22.00 | 8.00 |
Very High (>31) | 4 | 50.00 | 6.00 |
Chatbot Usefulness Aspect | Expectations (%) | Perceived Usefulness (%) |
---|---|---|
Specific Question Resolution | 84.2 | 91.4 |
Concept Comprehension | 100 | 95.7 |
Practical Examples | 42.1 | 61.4 |
Exam Training | 52.6 | 45.4 |
Content Review | 73.7 | 42.9 |
Motivation and Study Tips | 10.5 | 0 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sánchez-Vera, F. Subject-Specialized Chatbot in Higher Education as a Tutor for Autonomous Exam Preparation: Analysis of the Impact on Academic Performance and Students’ Perception of Its Usefulness. Educ. Sci. 2025, 15, 26. https://doi.org/10.3390/educsci15010026
Sánchez-Vera F. Subject-Specialized Chatbot in Higher Education as a Tutor for Autonomous Exam Preparation: Analysis of the Impact on Academic Performance and Students’ Perception of Its Usefulness. Education Sciences. 2025; 15(1):26. https://doi.org/10.3390/educsci15010026
Chicago/Turabian StyleSánchez-Vera, Fulgencio. 2025. "Subject-Specialized Chatbot in Higher Education as a Tutor for Autonomous Exam Preparation: Analysis of the Impact on Academic Performance and Students’ Perception of Its Usefulness" Education Sciences 15, no. 1: 26. https://doi.org/10.3390/educsci15010026
APA StyleSánchez-Vera, F. (2025). Subject-Specialized Chatbot in Higher Education as a Tutor for Autonomous Exam Preparation: Analysis of the Impact on Academic Performance and Students’ Perception of Its Usefulness. Education Sciences, 15(1), 26. https://doi.org/10.3390/educsci15010026