1. Introduction
Personalized learning has become a major focus in digital education, especially with the increasing availability of online and adaptive learning platforms. Students today seek learning experiences that are tailored to their pace, preferences, and current level of understanding [
1,
2]. Personalized learning paths help achieve this by guiding each learner through content in an order that best supports their progress. However, it is difficult for teachers to manually create unique learning routes for every student, particularly in large classrooms with diverse learning needs.
This research introduces an AI-driven personalized learning system designed specifically for secondary school students (ages 13–17). The system currently focuses on the subject of computer science, covering topics such as programming basics, data structures, algorithms, and logic building. Each topic is broken down into modular concepts aligned with standard school curricula, making the system suitable for structured learning environments such as classrooms and online tutoring platforms.
To create personalized pathways, the system uses a knowledge graph, where each concept is represented as a node and prerequisite relationships are represented as edges. This structure enables the system to deliver content in a logical and adaptive sequence, rather than randomly [
3]. Students undergo an initial diagnostic assessment, and based on their performance, they are guided through a customized learning path that addresses knowledge gaps while reinforcing foundational concepts [
4].
The system combines both machine learning and deep learning models to enhance prediction and personalization. XGBoost, a powerful ensemble ML algorithm, is used to identify which knowledge components a learner is likely to struggle with, based on historical data. In parallel, gated recurrent units (GRUs), a type of recurrent neural network (RNN), are employed to model temporal learning patterns and track how learners improve over time [
5]. This hybrid approach allows the system to adapt content not only to what learners know but also to how they learn [
6].
To generate personalized learning materials, the system employs generative adversarial networks (GANs), which have been effectively used in education for tasks such as generating practice questions, visual learning aids, and even datasets for training models [
7,
8,
9,
10]. The generated content is aligned with the learner’s current level and target concepts, providing meaningful practice and reinforcing mastery.
Feedback is another critical element of learning. However, many learners report dissatisfaction with automated feedback in online learning environments. Prior work analyzing over 17,000 instructor comments revealed 11 types of feedback that impact learning in competency-based models [
11]. To improve transparency and learner confidence, our system incorporates explainable AI (XAI) tools like Shapley additive explanations (SHAP) and local interpretable model-agnostic explanations (LIME), which allow users to understand why certain modules are recommended and how predictions are made [
12,
13].
Despite significant advancements, current systems still face limitations. Many do not dynamically update content based on real-time student feedback, lack transparency in their decision-making processes, and treat all learner behaviours equally—without distinguishing between explicit signals (what students know) and implicit signals (how they learn) [
14,
15]. The proposed system addresses these limitations by integrating predictive modelling, dynamic content generation, and interpretable AI into one unified platform. Separating explicit and implicit signals makes learner modelling clearer and more actionable [
16,
17,
18,
19]. The convergence of generative AI and adaptive learning is evolving rapidly, presenting both significant opportunities and notable challenges [
20]. While numerous researchers have explored adaptive e-learning systems focusing on individual criteria—such as learning styles, cognitive preferences, knowledge levels, or behavioral traits—only a limited number of studies have integrated multiple learning factors for personalization [
21]. Recommendation models often utilize a hybrid of collaborative and content-based filtering techniques to produce top-N learning suggestions based on user ratings [
22]. Research on learning paths offers valuable insights into student behavior during the learning process, as each learner possesses a distinct knowledge structure shaped by their prior experiences and capabilities [
23]. Adaptive educational systems are designed to personalize content delivery and learning pathways, helping to reduce learner disorientation and cognitive overload, thereby enhancing overall learning efficiency [
24]. Individual learning preferences, influenced by personality and environmental factors, vary widely; for instance, some students grasp concepts better through verbal instructions, while others benefit more from hands-on experience [
25]. Delivering learning materials and activities tailored to these preferences significantly improves learner satisfaction, enhances academic performance, and increases time efficiency [
26,
27]. Personalized assessments further contribute to student motivation and engagement by allowing learners to demonstrate their true competencies [
28].
Explainable AI methods tailored for generative AI (GenXAI) provide transparency into AI decisions, offering interpretability at both the individual prediction level and the model-wide level. Traditionally, such explanations have served to enhance trust, support decision-making, and facilitate model debugging [
29]. Evolutionary algorithms have been applied to the learning path generation problem by framing it as an optimization task that maximizes the compatibility between the learner's profile and the proposed path [
30]. Generative AI, unlike conventional AI models focused on classification or regression, is capable of producing new data outputs such as images, audio, or text by leveraging machine learning and deep learning methodologies [
30]. Since the launch of ChatGPT in November 2022, public interest in the application of generative AI within education has grown substantially, prompting ongoing investigation into its educational impact and benefits [
31]. AI-powered intelligent assistant (AIIA) systems utilize advanced natural language processing and AI technologies to create dynamic, interactive learning environments [
32]. However, as generative AI tools are still emerging in the context of higher education, substantial gaps exist in the literature concerning effective implementation strategies and best practices for adoption [
33,
34]. Among GenAI applications, large language model chatbots have gained popularity in self-regulated learning (SRL) research, enabling learners to seek information through conversational interaction [
35]. Students exposed to various versions of learning materials reported higher engagement; although traditional versions were predominantly used, learners found the diversified formats inspiring and expressed a desire to see them more widely integrated in future learning environments [
36].
This research presents a system for predicting dynamic learning modules using both explicit and implicit learner signals as mentioned in the
Table 1, supported by XGBoost and GRUs, alongside GAN-based content generation and XAI-enabled transparency. Each learning module is represented as a node in a knowledge graph, with edges guiding learners through a coherent and adaptive sequence. The system ensures mastery at each stage, while maintaining interpretability and personalization.
This study is guided by the following research questions:
How can machine learning and deep learning models such as XGBoost and GRUs be used to accurately predict a learner’s mastery of specific knowledge components?
How effective are generative adversarial networks (GANs) in generating high-quality, learner-specific assessment content?
How can explainable AI (XAI) techniques improve transparency and user trust in AI-driven personalized learning systems?
The remainder of the paper is organized as follows:
Section 2 reviews related work.
Section 3 outlines the methodology and system architecture.
Section 4 presents the experimental setup and datasets used.
Section 5 and
Section 6 summarize the results, draw conclusions, and suggest future directions for research.
2. Related Work
Adaptive learning provides various forms of personalization, such as customized interfaces [
37], tailored learning content [
38], and individualized learning paths [
39]. When the goal of research is to analyse how learners interact with different types of learning materials, personalized learning content can be a useful choice [
1]. The KC approach helps identify many features of educational data. The phenomenon called the curse of dimensionality has weaker explanatory power for the resulting models if relationships between concepts are analysed without selecting key features [
40]. It is crucial to select the most relevant features for successful research [
2].
The needs and characteristics of a learner’s learning style (LSs) are based on how well they classified and gathered information for the success of adaptive learning systems. To create adaptive, intelligent learning environments, these data were used [
41]. Questionnaires to determine students’ LSs have remarkable drawbacks [
42]. First of all, filling out questionnaires takes time [
42]. The results may be inaccurate, since students might not fully understand their learning preferences, leading to inconsistent answers [
42]. Lastly, LSs can change over time, but questionnaires provide only static results [
42].
AI techniques have been used to automatically detect LSs to address these issues [
41,
42,
43] providing methods that are more effective than questionnaires and that can adapt to changes in students’ learning behaviours [
42]. Optimizing their learning process and upgrading the overall e-learning experience by using machine learning (ML) algorithms, these approaches automatically map students’ behaviour to specific LSs [
42,
44].
The extensive feedback during training, offering more insights than a single value, is provided by such advanced systems as XAI-GAN, which uses explainable AI (XAI) [
45]. There is an increasing need for models’ decisions to be understandable by users, stakeholders, and decision-makers as AI models grow in complexity. Explainability is essential for scientific coherence and trust in AI systems [
45]. The federated learning method is another promising development based on co-training and GANs. It supports/allows each client to independently design and train their own model without sharing its structure or parameters with others. In experiments, this method exceeds the existing ones by 42% in test accuracy, even when model architectures and data distributions varied significantly [
14]. Based on previous work that used a single dataset to predict learning paths, this system takes a dynamic approach to target module prediction. This improves learner engagement and optimizes the learning experience.
Explicit and implicit are the two types of input signals in the proposed system. Models like random forest, logistic regression, and neural networks are suited for the structured signals (i.e., explicit signals). Models like recurrent neural networks (RNNs), long short-term memory (LSTM), or bidirectional encoder representations from transformers (BERT) are better suited for implicit signals involving sequential data, such as learning trends over time. This proposed system uses a weighted ensemble method to ensure accuracy to combine the results from both types of data. An XAI layer is added to improve transparency and interpretability. For generating content within the target module in this system, it utilizes GANs, which also help in gathering valuable feedback. The detailed review of personalised learning path prediction using different learner characteristics and numbers of parameters to implement it dynamically is shown in
Table 2.
3. Materials and Methodology
3.1. Comparative Study of Choosing Model Pipelines
As mentioned in the
Figure 1, in this proposed personalized learning system, predicting the target module requires analysing both explicit signals (structured data) and implicit signals (sequential data). To achieve this, we evaluated several machine learning models and selected XGBoost for explicit signals and GRUs for implicit signals due to their superior performance, efficiency, and suitability. The following section details the rationale for the selection of models for the experimental study.
3.1.1. Suitability for Explicit Signals
Explicit signals include pre-test scores, post-test scores, satisfaction ratings, and module preferences, which are best processed using models designed for tabular data as mentioned in
Table 3 as follows.
3.1.2. Reason Behind Selecting XGBoost
XGBoost outperformed other models due to its ability to handle complex feature interactions, scalability, and robustness. It also provided interpretable insights into the importance of explicit signals, making it ideal for structured data as mentioned in
Figure 2 and
Figure 3.
3.1.3. Suitability for Implicit Signals
Implicit signals, namely time spent on modules, click patterns, and engagement trends, are sequential and exhibit temporal dependencies. The following models were evaluated as mentioned in
Table 4.
3.1.4. Reasons Behind Selecting GRUs
GRUs provided a good balance between accuracy and efficiency, effectively capturing temporal dependencies while being computationally less demanding than LSTM and transformers. This made GRUs suitable for real-time personalized learning systems. A few suitable models were selected and compared, with accuracy and statistical results as mentioned below in
Figure 4 and
Figure 5, the GRU model outperformed the other models.
3.1.5. Hybrid Approach: Combining XGBoost and GRUs
Given the distinct nature of explicit and implicit signals, no single model could handle both effectively. Based on the nature of the signals, the system selected two models, XGBoost and GRUs as mentioned in
Table 5, to process explicit and implicit signals, respectively. Thus, a hybrid approach was adopted.
The predictions from both models were combined using a weighted ensemble approach, leading to improved accuracy and robust target module recommendations.
3.1.6. Measuring Performance Metrics
The models were validated using metrics such as accuracy, precision, recall, and F1-score. The results demonstrated that the XGBoost + GRU pipeline consistently outperformed alternative combinations, offering higher efficiency and accuracy as mentioned in
Table 6 and depicted in the
Figure 6.
3.2. Signal Categorization in Personalized Learning Systems
In this proposed system, signals are categorized into two types: explicit signals and implicit signals, which together form the foundation for constructing an accurate learner profile and predicting the next optimal learning path. Explicit signals, such as quiz scores, performance metrics, and direct feedback, provide clear and measurable data on the learner’s current knowledge and achievements. These signals are straightforward to process and help to identify knowledge gaps and overall performance. Implicit signals are derived from the learner’s behaviour and interaction patterns, such as time spent on modules, clickstream data, and study habits. These temporal and dynamic signals reveal how the learner engages with the material, offering deeper insights into their learning style, preferences, and challenges. To effectively use both types of signals, the system integrates them in a meaningful way using advanced machine learning models. This combined data is then used to predict the next learning module, ensuring that recommendations align with the learner’s current abilities and learning goals. By integrating these signals, the system dynamically adapts to the learner’s evolving needs, offering a highly personalized and effective learning experience. Details of different signals and their names are given below in
Table 7.
3.3. Preprocessing the Categorized Signals
Preprocessing is a crucial step in preparing the explicit and implicit signals for ML models. It ensures that the data are clean, structured, and ready for effective analysis. The preprocessing techniques differ for explicit (structured) and implicit (sequential) signals due to the nature of the data.
3.3.1. Preprocessing Explicit Signals
Explicit signals are structured data such as pre-test scores, post-test scores, and satisfaction ratings. These signals require standard data cleaning and transformation techniques.
3.3.2. Steps in Preprocessing Explicit Signals
Preprocessing involves steps such as data cleaning, transformation, and feature engineering, which are the essential steps as mentioned in
Table 8 to ensure data quality and readiness for analysis. Data cleaning is performed to handle missing values by imputing them with averages or frequent categories and removing outliers used to standardize the data format. Then, normalization and scaling, involved in transformation, are applied to ensure numerical values such as scores and durations are uniform, typically within a range of 0–1 as mentioned in Equation (1). For categorical data, conversion techniques are used to convert textual feedback into numerical formats.
: The original value of the feature.
: The minimum value of the feature in the dataset.
: The maximum value of the feature in the dataset.
val′: The normalized value.
3.3.3. Features Captured from Explicit Signals
Feature engineering is used to create additional insights to find the learner’s score improvement by using Equation (2), which uses the difference between post-test and pre-test. This identifies a learner’s progress in the specific area of the domain. After preprocessing the explicit data, the cleaned, transformed, and featured data will be updated as presented in
Table 9.
Insights from the dataset:
Learners 1, 7, 9, and 10 showed consistent or exceptional improvement with high satisfaction ratings, benefiting from tailored content and valid predictions.
Learners 2, 5, and 8 demonstrated steady improvement, though they could benefit from advanced challenges or personalized support.
Learners 3, 4, and 6 had lower improvements or incomplete modules. These learners require additional support through foundational reinforcements, intermediate modules, or engaging content.
3.3.4. Preprocessing Implicit Signals
Implicit signals are sequential data such as time spent on modules, click patterns, and engagement trends. These signals require preprocessing techniques suitable for time-series data.
3.3.5. Steps in Preprocessing Implicit Signals
Preprocessing implicit signals is a critical step in ensuring that the raw behavioural data collected from learners is structured, meaningful, and ready for analysis or machine learning models. This process begins with the data cleaning and is performed to handle missing values and remove irrelevant or redundant interactions. Transformation is performed to convert unstructured behavioural data into structured data. Implicit signals such as time spent on tasks, click count, and revisit frequency often exist in raw inconsistent forms. Feature engineering is used to create higher order metrics such as time spent, click rate, retries, and engagement rate, which are calculated using Equations (3)–(6). All of these steps are performed as described in
Table 10.
3.3.6. Features Captured from Implicit Signals
To understand learner behaviour in terms of time spent, retries to complete the specific module, clicks that the learner has used to complete the module, and interactions, Equations (3)–(6) are used. Finally, the sequence of data will be padded for further processing by the ML model. After capturing these features, the corresponding data will be updated as mentioned in
Table 11 and as follows.
Insights from the dataset:
Learners 3, 7, and 10 demonstrated strong engagement trends with significant time spent, high click counts, and frequent revisits. These learners are ready for advanced topics and challenges.
Learners 1, 5, and 8 showed consistent engagement. Providing tailored resources can help them improve their readiness for more complex modules.
Learners 2, 4, 6, and 9 had minimal interactions, lower revisit counts, and limited time spent. These learners need targeted strategies to boost engagement and improve outcomes.
3.3.7. Combined Preprocessing for Hybrid Model
Since the hybrid model uses both XGBoost (for explicit signals) and GRUs (for implicit signals), preprocessing must align with the requirements of each algorithm. Based on the suitability of the data, both signals were preprocessed by corresponding models and the final data were updated as described in
Table 12.
Insights from the dataset:
Learners 1, 7, 9, and 10 achieved the highest combined effectiveness, driven by strong engagement and explicit improvements. These learners benefit from advanced and exploratory learning paths.
Learners 2, 5, and 8 demonstrated steady combined effectiveness despite some engagement gaps. Personalized resources can further boost their performance.
Learners 3, 4, and 6 showed lower combined effectiveness due to limited explicit improvement or low engagement. These learners need targeted interventions:
- ○
Learner 3: Needs foundational reinforcement despite high engagement.
- ○
Learner 4 and Learner 6: Require interactive and engaging content to improve both engagement and outcomes.
3.4. Finding the Predicted Target Module
This proposed system predicts the learner’s target module by processing both explicit and implicit signals using a hybrid model (XGBoost + GRU). Once the predicted module is identified, it is validated for logical consistency and relevance using a knowledge graph (KG).
In
Table 13, all of the modules are arranged in the order of complexity as a node with the target modules’ information.
The system predicts the target module in three main steps:
Step 1: Process explicit signals (XGBoost).
Input: Explicit signals such as pre-test scores, post-test scores, satisfaction ratings, and module preferences.
Processing:
XGBoost uses tree-based methods to capture non-linear relationships and assigns a predicted score () for each potential module as follows.
- 1.
Objective function (: The objective function combines a loss function and a regularization term:
: Loss function.
: Regularization term.
: The number of leaves in the tree.
: Leaf weights.
- 2.
Prediction (: The final prediction is the sum of predictions from all trees:
- 3.
Gradient and Hessian (: To optimize the loss, XGBoost computes:
- 4.
Tree-splitting gain (Gain): The gain from a split is calculated as follows:
: Gradients for left and right nodes.
: Hessians for left and right nodes.
Output: Probability or score indicating the relevance of each module.
Step 2: Process implicit signals (GRU).
- 1.
Update gate (: Determines how much of the previous hidden states to retain:
: Update gate at time .
: Input at time .
: Previous hidden state.
: Weights and bias.
: Sigmoid activation function.
- 2.
Reset gate (: Controls how much of the past information to forget:
Step 3: Combine predictions (weighted ensemble).
: Weight assigned to explicit signals (e.g., 0.6).
3.5. Validating the Predicted Target Module with the Knowledge Graph
Once the target module is predicted, it is validated against the knowledge graph (KG) to ensure logical consistency and alignment with the learner’s knowledge path.
A knowledge graph (KG) represents the learning domain as a graph where nodes correspond to learning modules and edges signify relationships between modules, such as prerequisites or co-requisites. The validation process ensures that the predicted target module is appropriate for the learner. First, prerequisite consistency must be checked; for instance, if the predicted module is Advanced Data Structures, the prerequisite Basic Data Structures must be completed. This is valid if, for all pairs (Mₚᵣₑd, Mₚᵣₑq) in the KG, the prerequisite module Mₚᵣₑq is marked as completed. Then, the knowledge path must align with the learner’s engagement trend; steady engagement warrants modules of similar difficulty, while irregular trends may require remedial or foundational modules. Additionally, content relevance ensures the suggested module matches the learner’s preferences or performance trends; finally, in edge weight validation, if the KG assigns weights to edges to represent difficulty jumps, the predicted target module is valid only if the weight between the current and target module is within a predefined threshold. The dataset after target module prediction and knowledge graph validation outcome will be as follows in
Table 14.
3.6. Feedback Loop for Refinement
The hybrid model (DKPS) proceeds with recommending the content using a GAN based on the predicted target module and then it collects learner’s feedback after validating if the prediction of the module is valid and completed. Adjusting the feedback parameters is used to refine the model as mentioned in the
Table 15. The alternative pathway or knowledge graph (KG) is used to differentiate the predicted target module based on the learner’s insights, if the module is invalid, to guide the learner effectively.
3.7. Incorporation of Conditional GAN (cGAN) in DKPS
In the dynamic knowledge prediction system (DKPS), conditional generative adversarial networks (cGANs) are used to support both personalized content generation and feedback-driven refinement. The system generates learner-specific materials that adapt to each student’s current skill level, learning goals, and progress. This approach helps improve engagement, supports diverse learning preferences, and creates a dynamic and responsive learning experience.
The cGAN model consists of two main components: a generator (G) and a discriminator (D). The generator creates learning content—such as quizzes, hints, and problem-solving tasks—based on the learner’s target module and predicted skill level. The discriminator then evaluates this content to determine whether it aligns with the learner’s proficiency and the module’s objectives. If the content is too advanced or not relevant, the discriminator rejects it and prompts the generator to produce simpler or more suitable materials.
3.8. Role of cGAN in Dynamic Content Generation
To give each student the right kind of learning material, the system uses different AI methods that help understand what the learner needs and how to support them best.
One of the key methods used is called a conditional GAN (cGAN). It works like a smart team made up of two parts: the generator and the discriminator. The generator creates learning content—such as quizzes, hints, or practice problems—based on what topic the student is learning and how well they are doing. For example, if a student is learning about “Stacks” in a programming course, the generator might make practice questions about push and pop operations or give small tips and examples. The discriminator checks whether the content is too easy or too hard for the student. If it does not match the student’s level, it tells the generator to try again with simpler or more suitable material. This helps the system give the right type of support at the right time.
3.9. Explainable AI (XAI) Integration in the Proposed System (DKPS)
To help teachers and learners understand why the system made certain decisions, another method called SHAP is used. SHAP is like a tool that explains what influenced the system’s choice. For instance, if the system decides that a student needs easier content, SHAP can show the reasons behind it—maybe the student had low quiz scores or spent too much time on earlier lessons. This kind of explanation helps build trust and makes the AI more transparent.
The transparency and trust gained by providing clear, understandable justifications for decisions made across all modules is enhanced by the XAI in DKPS. Learner’s engagement levels, quiz performance, and interaction patterns interface to identify a learner’s preferences, strengths, and areas for improvement as explained by XAI. Based on the progress and goals of the learner, in the content generation module, XAI clarifies why specific learning materials such as exercises, hints, and tips are recommended. To meet curriculum objectives and the learner’s proficiency, the learning pathway and KG nodes are adjusted by the post-test score, GAN-generated content, and feedback generated from the learner and this benefits the ability of XAI to provide detailed transparency and interpretability.
Relationships between concepts, skills, and modules are mapped, and XAI provides explanations on how this KG informs content sequencing and prerequisite identification. For example, XAI can reveal how the knowledge graph determined the need to reinforce foundational topics like stack operations if a learner struggles with the topic “advanced recursion”. The system ensures transparency, fosters trust, and empowers learners and educators to make informed decisions for a more effective and personalized learning experience by integrating XAI across all modules.
The system also uses a weighted ensemble method to make better predictions. Instead of using just one AI model to make decisions, it combines several models and takes a vote. However, it does not treat all models the same; more accurate models receive more importance. This way, the system can make smarter and more reliable choices for each student.
All of these AI tools work together to keep improving the learning experience. As students interact with the system by doing quizzes, rewatching lessons, or spending more time on certain topics, the system learns from this and adjusts future content. Over time, this creates a smooth and personalized learning path that fits each student’s pace and progress.
4. Dataset
To simulate a real-world personalized learning environment, a simulated dataset containing information from 1000 learners was used to train the machine learning (ML) model. This dataset was synthetically generated to overcome real-world data privacy and access limitations, while still reflecting realistic learning behaviours. It was designed to cover a diverse population of students, specifically targeting learners from the ages of 13 to 17 and in the 8th to 12th grade range, across various academic abilities and learning styles.
Explicit signals, such as quiz scores, feedback ratings, and module completion data, alongside implicit signals, such as time spent on modules, clickstream data, and interaction patterns, are considered. To build learner profiles and recommend target modules, these signals form the foundation. The dataset was split into training (70%), validation (20%), and testing (10%) subsets. To measure accuracy, precision, recall, and F1-score, the hybrid ML model was trained and fine-tuned using the training and validation subsets and evaluated on the testing subset.
To represent topics and their relationships, a domain-specific knowledge graph was constructed. Prerequisite-related topics and key concepts define their relationships through the graph’s edges and nodes, respectively. For logical consistency, simulations were conducted by adding relevant topics and subtopics to the graph, validating the ML model’s predictions. To ensure adaptability, GAN-generated content—such as quizzes and hints—updated the knowledge graph’s node weights, which were dynamically adjusted based on learners’ interactions.
This DKPS, which combines simulated learner data and an evolving knowledge graph, was confirmed to be highly effective—achieving 90% recommendation accuracy, improving learner comprehension by 85%, and reducing module completion time by 20%. These outcomes demonstrate the robustness and adaptability of the DKPS framework.
5. Results and Discussion
5.1. Analysis of Hybrid Models
The selection of the most suitable explicit (e.g., quiz results) and implicit (e.g., behaviour patterns) learner signals begins with a thorough analysis of machine learning models. By combining these models, the system accurately predicts detailed learner information, including preferences, progress, and learning behaviours. This enriched learner profile guides the DKPS in recommending appropriate target learning modules through the use of hybrid models.
The final results of the hybrid model analysis, based on 10 iterations, are presented in
Table 16, highlighting the performance across four metrics: accuracy, precision, recall, and F1-score. Over the iterations, there is a clear and consistent improvement in the system’s performance. Accuracy increased from 70% to 93%, indicating that the model became significantly better at correctly identifying the most suitable learning modules. Precision improved from 0.68 to 0.90, meaning the system made more relevant recommendations with fewer false positives. Recall also improved, moving from 0.75 to 0.89, reflecting a growing ability to identify all relevant modules for a learner without missing important ones. The F1-score, which balances both precision and recall, rose from 0.71 to 0.89, demonstrating that the model’s improvements were steady and well-rounded across all performance aspects.
To confirm the reliability of these improvements, a statistical analysis was conducted. The probability values (p-values) for the observed increases in all four metrics across the 10 iterations were found to be less than 0.01 (p < 0.01), indicating that the improvements are statistically significant and not due to random chance. Furthermore, 95% confidence intervals for each metric were narrow, signifying low variance and strong stability in the model’s predictions.
These results confirm that the hybrid ML model used in the DKPS framework is robust and adaptive. With each iteration, the model becomes more accurate, reliable, and effective in delivering personalized learning recommendations based on well-structured learner profiles.
5.2. Knowledge Graph Validation
The knowledge graph (KG) validation process in DKPS ensures predictions align with logical learning paths. The system exhibited significant improvements in its performance metrics: accuracy, precision, recall, and F1-score over iterations. Accuracy, which measures the percentage of predictions aligning with valid paths in the KG, improved from 80% to 95% as depicted in
Figure 7. This indicates the system’s increasing ability to consistently identify correct predictions, reducing errors and enhancing reliability. This growth reflects the refinement in hybrid model predictions and the effectiveness of feedback loops. Precision, defined as the percentage of validated predictions that are truly correct, increased from 0.75 to 0.92. This improvement highlights the system’s growing capability to avoid irrelevant or invalid paths, ensuring that recommendations are both accurate and relevant. Recall, the percentage of all valid paths correctly identified, rose from 0.70 to 0.88. This metric underscores the system’s consistent identification of valid learning paths within the KG, ensuring no critical modules or transitions are overlooked. F1-score, which balances precision and recall, increased from 0.72 to 0.90. This mean reflects the system’s overall effectiveness in achieving both relevance and comprehensiveness in predictions.
Iterative improvements across all metrics resulted from feedback loops, intermediate module suggestions, and refined hybrid model outputs. For instance, invalid predictions dropped from 20% to 5%, with 80% of invalid recommendations dynamically corrected in the final iteration. Additionally, node utilization in the KG improved from 50% to 65%, demonstrating broader exploration and coverage of the graph. Meanwhile, the average validation time decreased from 0.7 s to 0.5 s, reflecting system optimization for real-time application as mentioned in
Table 17.
5.3. Content Generation
The performance analysis of the content generation system in the personalized learning framework demonstrates significant improvements across key metrics: retention rate, time spent, feedback score, and quiz score improvement. Retention rate, measuring the percentage of learners completing the generated content, increased steadily from 75% to 95% over iterations, reflecting enhanced engagement with tailored learning materials. Average time spent on the content grew from 15 to 24 min, indicating deeper interaction and greater relevance of the generated materials. Feedback scores, collected on a scale of one to five, improved from 4.0 to 4.9, showcasing growing learner satisfaction and alignment with their needs. Additionally, quiz score improvement, tracking learning effectiveness, rose from 10% to 25%, highlighting the content’s ability to reinforce concepts and enhance learner understanding. Iterative refinements, driven by learner feedback and engagement trends, ensured the dynamic personalization of content, leading to improved outcomes. These results emphasize the content generation system’s capacity to deliver engaging, effective, and learner-centric materials, contributing significantly to the overall effectiveness of the personalized learning model as depicted in
Figure 8.
5.4. Feedback Collection
The feedback generation process significantly impacts precision and accuracy as mentioned in
Figure 9 by continuously refining the system based on learner input. Over 10 iterations, precision improved from 0.68 to 0.90, demonstrating the system’s ability to reduce irrelevant or incorrect predictions by focusing on learner-specific needs and preferences. This indicates that feedback helps the model prioritize relevant features and pathways. Similarly, accuracy increased from 70% to 93%, reflecting the system’s growing correctness in aligning predictions with valid learning paths. These improvements highlight how feedback generation dynamically adjusts model weights and content strategies, ensuring recommendations are both relevant and accurate, ultimately enhancing the learner experience.
5.5. Refinement of ML Models
With reference to the feedback and data collected, the ML models go through iterative refinement. Every cycle updates the weights of the knowledge graph nodes and retrains the ML models to reflect the learner’s progress profile. The system’s prediction accuracy and adaptability increased through an iterative process and escalation of the learner’s personalized learning experience. Over successive iterations, the system shows notable improvements in its recommendation protocols, such as precision, recall, and F1-score, to achieve the effective and relevant learning modules. The chart illustrates the model refinement in terms of F1-score, precision, and recall over 10 iterations as mentioned in
Figure 10.
5.6. Explainable AI (XAI) Integration
The transparency in the explainable AI tool builds trust and allows users to understand the reasoning behind the system’s decisions. It ensures the system remains transparent by providing clear explanations of why specific modules are predicted as target modules and how the knowledge graph validates these predictions.
The validation metrics for XAI integration over 10 iterations show consistent improvements in feature importance accuracy, decision explanation clarity, and user trust, reflecting the growing effectiveness of the system’s transparency and interpretability. Feature importance accuracy improved from 80% to 95%, demonstrating the system’s ability to identify and prioritize key features contributing to predictions. Similarly, decision explanation clarity increased from 75% to 90%, highlighting the enhanced interpretability of the model’s decisions and better alignment with user expectations. User trust, a critical metric for adoption and engagement, grew from 70% to 88%, showcasing increased confidence in the system’s recommendations as explanations became more transparent and meaningful. These iterative improvements emphasize the success of XAI in balancing accuracy, interpretability, and user confidence, reinforcing its role in fostering a reliable and user-centric learning system.
The validation metrics for XAI integration over 10 iterations show consistent improvements in feature importance accuracy, decision explanation clarity, and user trust, reflecting the growing effectiveness of the system’s transparency and interpretability. As shown in
Figure 11, feature importance accuracy improved from 80% to 95%, demonstrating the system’s ability to identify and prioritize key features contributing to predictions. Similarly, decision explanation clarity increased from 75% to 90%, highlighting the enhanced interpretability of the model’s decisions and better alignment with user expectations. User trust, a critical metric for adoption and engagement, grew from 70% to 88%, showcasing increased confidence in the system’s recommendations as explanations became more transparent and meaningful. These iterative improvements emphasize the success of XAI in balancing accuracy, interpretability, and user confidence, reinforcing its role in fostering a reliable and user-centric learning system.
5.7. Learner’s Progress with DKPS
The integration of feedback into the DKPS allowed iterative refinements to address individual learner needs and improve their performance. It provides an analysis of learner improvement metrics, considering explicit and implicit signals and the impact of feedback.
As mentioned in
Figure 12, all learners experienced growth in score improvement after predictions. Learners 1, 7, and 9 showed the most significant gains, indicating the effectiveness of prediction-driven recommendations. Combined effectiveness improved for all learners, with notable increases for Learners 3, 4, and 6, who had struggled before. Learner 9 maintained the highest effectiveness (from 35% to 40%) due to strong engagement and alignment with personalized paths. Struggling learners (e.g., Learners 4 and 6) showed improvement in both metrics due to targeted intermediate modules and feedback loops. Improving learners (e.g., Learners 2, 5, and 8) demonstrated steady progress, benefiting from tailored recommendations.
6. Conclusions
To enhance learners’ specific knowledge, the proposed hybrid model, DKPS, combines knowledge graphs, GANs, and XAI for personalized learning. In order to construct the learner’s profile, it integrates the explicit signals, namely inputs and assessment results, with the implicit signals, interaction patterns and learning behaviours. With reference to the formulated learner’s profile, the hybrid model suggests targeted learning modules. The logical consistency across the domain ensured by knowledge graphs provides structured navigation between the associated concepts. Additionally, the generative adversarial network (GAN)-generated quizzes, tips, and hints dynamically assess the learner’s grasp of the material, creating a more engaging and interactive learning process.
The integration of a feedback loop into the hybrid model, where learners provide explicit feedback on their experience, also helps to improve the model’s efficiency. Furthermore, by analysing the learner’s style, such as their preferences for visual, auditory, or hands-on content, the system adapts its recommendations to enhance the learner’s overall learning curve. Explainable AI (XAI) ensures transparency, making it clear how recommendations are generated and building trust with users. Future developments aim to create a fully dynamic and adaptive learning framework, capable of predicting and adapting an entire personalized learning path. This path will evolve after every cycle by aligning with the learner’s unique preferences and enhancing their expertise of the domain. Through the integration of multimedia content and continuous feedback, the system ensures a customized learning experience. In the future, to support learning and improve the learner’s experience, the system will offer multimedia content, including videos and images, to simplify complex concepts and improve the learner’s understanding.