Next Article in Journal
The Role of Students’ Perceptions of Educators’ Communication Accommodative Behaviors in Classrooms in China
Previous Article in Journal
Beyond Choice: Affective Representations of Economic and Moral Decisions
Previous Article in Special Issue
How Do Educational Settings Influence College Students’ Reading Behavior? An Empirical Study of China’s Top Universities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Behavioral Analysis of Postgraduate Education Satisfaction: Unveiling Key Influencing Factors with Bayesian Networks and Feature Importance

by
Sheng Li
1,
Ting Wang
2,
Hanqing Yin
1,
Shuai Ding
3 and
Zhiqiang Cai
2,*
1
Graduate School, Northwestern Polytechnical University, Xi’an 710072, China
2
Department of Industrial Engineering, Northwestern Polytechnical University, Xi’an 710072, China
3
School of Public Policy and Administration, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Behav. Sci. 2025, 15(4), 559; https://doi.org/10.3390/bs15040559
Submission received: 14 February 2025 / Revised: 18 April 2025 / Accepted: 19 April 2025 / Published: 21 April 2025
(This article belongs to the Special Issue Behaviors in Educational Settings—2nd Edition)

Abstract

:
Accurately evaluating postgraduate education satisfaction is crucial for improving higher education quality and optimizing management practices. Traditional methods often fail to capture the complex behavioral interactions among influencing factors. In this study, an innovative satisfaction indicator system framework is proposed that integrates a two-stage feature optimization method and the Tree Augmented Naive Bayes (TAN) model. The framework is designed to assess key satisfaction drivers across seven dimensions: course quality, research projects, mentor guidance, mentor’s role, faculty management, academic enhancement, and quality development. Using data from 8903 valid responses, Confirmatory Factor Analysis (CFA) was conducted to validate the framework’s reliability. The two-stage feature optimization method, including statistical pre-screening and XGBoost-based recursive feature selection, refined 49 features to 29 core indicators. The TAN model was used to construct a causal network, revealing the dynamic relationships between factors shaping satisfaction. The model outperformed four common machine learning algorithms, achieving an AUC value of 91.01%. The Birnbaum importance metric was employed to quantify the contribution of each feature, revealing the critical roles of academic resilience, academic aspirations, dedication and service spirit, creative ability, academic standards, and independent academic research ability. This study offers management recommendations, including enhancing academic support, mentorship, and interdisciplinary learning. Its findings provide data-driven insights for optimizing key indicators and improving postgraduate education satisfaction, contributing to behavioral sciences by linking satisfaction to outcomes and practices.

1. Introduction

Currently, the nation is actively advancing a high-quality education system, focusing on cultivating top innovative talents. This effort strengthens the foundation of education, science, and talent, essential for building a modernized socialist nation and enhancing the training of high-level professionals. Postgraduate education, as the highest level of education, plays a vital role in enhancing national competitiveness. Improving its quality is essential not only for universities to remain competitive but also for the broader development of the nation’s educational system. Postgraduate satisfaction, as a key indicator of educational quality, has become a focal point of attention. Satisfaction surveys provide valuable insights into students’ perceptions of various aspects such as curriculum design, teaching quality, and available resources, all of which are shaped by behavioral and psychological factors. These insights help administrators identify areas that need improvement, shape educational policies, and allocate resources in ways that are more aligned with students’ behavioral responses and their academic experiences.
Despite extensive research on educational satisfaction, significant limitations remain. Many studies rely on traditional statistical methods, such as surveys and correlation analyses, which may not fully capture the complex behavioral interactions among satisfaction-driving factors. Additionally, some methods depend on linear assumptions or subjective weighting, which can limit their ability to account for the behavioral complexity of educational systems. Additionally, existing research has not sufficiently addressed the dynamic, causal relationships between behavioral factors and satisfaction, nor has it effectively quantified the importance of individual features. This research gap is exacerbated by the underutilization of advanced artificial intelligence techniques, such as Bayesian networks, which can provide more comprehensive and robust solutions by uncovering underlying behavioral patterns.
To overcome these challenges, a new framework is introduced in this study to evaluate postgraduate education satisfaction, integrating behavioral insights. The framework uses a two-step feature optimization process, including statistical pre-screening and XGBoost-based feature elimination, to select the most important factors. A Tree Augmented Naive Bayes (TAN) model is used to build a causal network, revealing how these factors interact. The model’s reliability is confirmed through confusion matrices and performance metrics. Additionally, Birnbaum importance metrics are used to quantify the marginal contributions of individual features to satisfaction, providing deep behavioral insights into the critical roles of academic enhancement, quality development, and the mentor’s role. This approach not only enhances the interpretability and scientific rigor of satisfaction evaluations but also provides actionable, data-driven recommendations for improving key behavioral indicators in postgraduate education.
By integrating behavioral analysis with the two-stage feature optimization method and Bayesian networks, this study overcomes the limitations of traditional statistical approaches, offering an innovative solution for causal modeling and feature importance quantification. The findings will contribute to both theoretical advancements and practical applications, providing new pathways for improving education quality and ensuring the effectiveness of educational policies.

2. Literature Review

2.1. Measurement and Key Factors of Postgraduate Education Satisfaction

Postgraduate education satisfaction, as a critical indicator of higher education quality, is typically defined as students’ subjective evaluation of their overall learning experience, based on a comparison between their expectations and actual perceptions of educational services (Muijs & Bokhove, 2017). Existing research commonly employs survey questionnaires and multidimensional models to measure satisfaction, with core indicators encompassing course quality, mentor support, learning environment, and career development opportunities. However, these factors do not function in isolation; they interact with students’ cognitive, emotional, and behavioral responses, influencing their overall satisfaction.
The literature reveals the diverse and complex nature of influencing factors:
The first is physical and environmental factors. The quality of indoor environmental conditions significantly impacts postgraduate students’ learning comfort and satisfaction, and key factors such as thermal comfort, air quality, lighting, and noise levels are highlighted. Optimizing the design of learning spaces has been shown to effectively enhance satisfaction and learning efficiency (Al-Dmour, 2024). The second is career development and educational support. The effective use of career-related social networks (e.g., LinkedIn) and social media not only contributes to learning interactions but also plays a crucial role in improving career development prospects and satisfaction (Hazzam et al., 2024). The third set of factors are teaching and curriculum design. The modernization of course content and the integration of advanced technologies, such as 3D printing methods, have significantly enhanced postgraduate students’ learning experiences, thereby boosting their satisfaction (Allanqawi et al., 2023). The final factors are disciplinary and regional differences. Significant variations in satisfaction levels exist across different disciplines and geographic regions. For instance, fields with a strong clinical or applied focus report higher satisfaction levels due to clear career development pathways and practical skills training (Bush & Lowery, 2016).
In summary, postgraduate education satisfaction not only influences students’ academic performance but is also closely associated with their career success (Pérez Fuentes et al., 2023).

2.2. Methods for Postgraduate Education Satisfaction Evaluation

The evaluation process typically involves indicator weighting and data modeling.
In determining indicator weights, subjective weighting methods (e.g., analytic hierarchy process) rely on expert judgment, which introduces an inherent level of subjectivity. In contrast, objective weighting methods (e.g., entropy weight) are based on the characteristics of the data but may overlook the practical significance of indicators (Guo & Li, 2018). Recently, hybrid weighting approaches that combine subjective and objective methods have garnered attention, which aim to balance practical significance and result stability; however, they often face increased complexity during the integration process.
In terms of data modeling, traditional linear models (e.g., multiple regression and gray system models) perform well with small sample sizes but struggle to capture the nonlinear relationships inherent in satisfaction (Sharif et al., 2012). Meanwhile, modern machine learning methods (e.g., BP neural networks) excel in high-dimensional data analysis, demonstrating superior performance. However, these methods are prone to overfitting and often lack interpretability. In recent years, hybrid models have emerged as a mainstream approach. For instance, the combination of BP with a gray system leverages their complementary strengths, effectively handling both nonlinear relationships and sparse data (Tanveer et al., 2023).
Bayesian networks (BNs) have shown promise as a potential tool for postgraduate education satisfaction research due to their ability to model causal relationships and dynamically update predictions. Unlike traditional models, BNs can capture the behavioral dependencies among various satisfaction-related factors, offering a deeper understanding of how students’ cognitive and emotional responses influence their overall educational experience. Furthermore, they support real-time updates with dynamic data, enabling the long-term monitoring and optimization of satisfaction levels (R. Huang et al., 2014).
However, the application of BNs in the educational field faces several challenges. First, the construction of accurate models relies heavily on high-quality data and prior knowledge. Second, in scenarios involving a large number of variables, the computational resource demands can be substantial. Lastly, inferring causal relationships requires the integration of domain expertise and experimental validation to avoid overfitting and ensure the robustness of the conclusions (Weber et al., 2012).

2.3. Limitations of Existing Studies

Despite the progress made, there are still four notable limitations in research:
(1)
Insufficient exploration of causal relationships: traditional statistical methods struggle to reveal complex causal relationships and fail to comprehensively analyze the interactions among multidimensional factors.
(2)
A lack of multidimensional integrated analysis: existing studies rarely consider the combined effects of factors such as the environment and curriculum on satisfaction. This omission limits the understanding of how these crucial dimensions interact and influence satisfaction (Y. Zhang et al., 2024).
(3)
Insufficient model applicability and interpretability: modern machine learning methods, such as Support Vector Machines (SVMs) and neural networks, perform well in predictions. However, their “black-box” nature limits the interpretability of research findings and reduces their practical applicability as guidance for decision-making.
(4)
Limitations of feature selection techniques: many studies rely on subjective weighting or simple statistical methods for feature importance evaluation. These approaches lack rigorous and scientific quantitative foundations, which could lead to the omission of key variables.
This study contributes to global scholarship in three key areas: cross-cultural studies of graduate student satisfaction, policy-oriented research on higher education quality assurance, and behavioral modeling in institutional assessment frameworks:
(1)
Our findings align with international research on graduate satisfaction, particularly studies that explore satisfaction differences between online and blended doctoral programs (Erichsen et al., 2014). Similarly, research highlights how cultural and contextual factors, such as nationality, influence student satisfaction (Stewart et al., 2018), emphasizing the need to consider diverse satisfaction drivers (Crede & Borrego, 2014).
(2)
This study also engages with global discussions on educational policies. For instance, research has examined how academic dismissal policies impact dropout rates and satisfaction (Sneyers & De Witte, 2017). Our study emphasizes the critical role of academic support and mentorship, advocating for policies that promote holistic student experiences—similar to the policy issues surrounding international student housing (Ramia et al., 2022).
(3)
By utilizing Bayesian networks and behavioral modeling, this research advances the use of data-driven methods in higher education assessment. It builds on existing work that connects supervisory styles to student creativity through psychological factors (Walker & Palmer, 2011). Furthermore, it contributes to studies on the impact of student funding policies (Czarnecki & Litwiński, 2024), illustrating how behavioral models can offer deeper insights into educational practices (Gu et al., 2017).

3. Methods

3.1. Analysis of Questionnaire and Indicator System

To ensure the reliability and validity of the questionnaire, a systematic indicator system was developed, drawing on extensive literature and expert reviews. Seven core dimensions were identified: course quality, research projects, mentor guidance, mentor’s role, faculty management, academic enhancement, and quality development, each with corresponding indicators. CFA was used to assess the structural rationality and consistency of the indicators.
Reliability analysis was conducted by calculating Cronbach’s α (with values above 0.7 considered satisfactory) and Composite Reliability (CR), evaluating the internal consistency of the questionnaire and its dimensions. A validity analysis was carried out using Average Variance Extracted (AVE) to assess the explanatory power of latent variables for the observed variables (X. Zhang et al., 2023). The formula is as follows:
A V E = i = 1 n λ i 2 i = 1 n λ i 2 + i = 1 n θ i , i = 1 , 2 , , n
where n is the number of features, i denotes a specific feature, λ i represents the standardized factor loadings, and θ i denotes the error terms.
CFA was used to test the goodness-of-fit between the hypothesized factor structure and the actual data. Additionally, the Kaiser–Meyer–Olkin (KMO) test and Bartlett’s test of sphericity were employed to verify the suitability of the factor analysis, ensuring that the questionnaire design demonstrated robust scientific validity and reliability.

3.2. Two-Stage Feature Optimization Method

To identify the most representative features from the initial set of 49 for inclusion in the causal network modeling process, a two-stage feature optimization method with the following steps is proposed:
Stage One: Statistical Feature Pre-Screening
To enhance the efficiency of feature selection, this study employed three statistical methods: Variance Thresholding (VT), Correlation Analysis (CA), and Mutual Information (MI). Features with variances below the threshold were eliminated. Subsequently, using CA coefficients combined with MI values, features exhibiting minimal correlation or limited contribution to the target variable were excluded.
To ensure robustness in the screening criteria, fixed thresholds were replaced with data-driven dynamic thresholds. Features were retained when they simultaneously satisfied the following conditions:
(1)
Features with variance higher than the 15th percentile of the dataset’s variance distribution ( 0.61 ).
(2)
Features with an absolute correlation coefficient greater than the 15th percentile of the dataset’s correlation distribution with the target variable ( | C o r r | 0.43 ) and feature-to-feature correlation matrix values of | C o r r | 0.95 to reduce multicollinearity.
(3)
Features with mutual information values greater than the 15th percentile of the dataset (≥0.17).
Stage Two: XGBoost-Based Recursive Feature Elimination (RFE)
XGBoost was chosen as the base model for RFE due to its high performance and efficiency through parallel computation and regularization optimization. It effectively handles high-dimensional nonlinear data and provides clear feature importance evaluation, forming a solid foundation for recursive feature selection. In this stage, feature importance is initially assessed using XGBoost’s built-in method, ranking features based on their gain values (J. Chen et al., 2020).
The calculation formula is as follows:
G a i n ( T , i ) = 1 | T | t T G a i n ( t , i )
where T represents the set of all trees and G a i n ( T , i ) refers to the gain achieved by feature i during its split within tree t .
Subsequently, RFE combined with cross-validation was employed to iteratively remove features with the lowest gain, thereby optimizing the feature subset (Q. Chen et al., 2018). The ultimate objective is to maximize contribution while minimizing redundancy.
The formula for the feature scoring function is as follows:
S c o r e ( i ) = i = 1 n β i 2
where β i represents the regression coefficient of the i -th feature in the model, and n denotes the total number of features.
In addition, to validate the effectiveness, the ReliefF and PSO methods were included in a comparative experiment. This multi-faceted evaluation of different strategies ensures that the final selected feature subset demonstrates superior practicality and stability.

3.3. Bayesian Network Theory

Based on probabilistic graphical models, a Bayesian network (BN) represents the conditional dependencies and joint probability distributions among random variables. The core concept derives from Bayes’ theorem, which combines prior and conditional probabilities to infer posterior probabilities, allowing for dynamic belief updates in a hypothesis (Si et al., 2019). The formula is as follows:
P ( A | B ) = P ( B | A ) · P ( A ) P ( B )
where P ( A | B ) represents the posterior probability of event A given that event B has occurred; and P ( B | A ) is the conditional probability of B given that A has occurred. P ( A ) and P ( B ) are the prior and marginal probabilities of events A and B , respectively.
Naive Bayes (NB) applies Bayes’ theorem with the assumption of feature independence, simplifying computational complexity. The Tree-Augmented Naive Bayes (TAN) model enhances this by introducing dependencies between features through a tree structure. TAN maximizes the likelihood function to optimize the feature dependencies (Xiao et al., 2022). The formula is expressed as follows:
P C A P C · i = 1 n P ( A i | P a r e n t s ( A i ) )
where C represents the class, A = { A 1 , A 2 , , A n } denotes the set of n feature variables, and p a r e n t s ( A i ) refers to the parent node of feature A i .
Figure 1 illustrates examples of NB and TAN structures.
Compared to fully connected Bayesian networks, TAN significantly reduces computational complexity by constructing a maximum weight spanning tree while retaining dependency information among features. The modeling process is shown in Table 1.
This study evaluates the performance of the classifier using a confusion matrix (Y. Wang et al., 2020). The commonly used metrics and their specific formulas are as follows:
Accuracy ( A c c ) denotes the ratio of correctly classified samples to the total number of samples, which is the sum of True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN):
A c c = T P + T N T P + T N + F P + F N
Recall ( R ) denotes the ratio of actual positive samples that are correctly identified:
R = T P T P + F N
Precision ( P ) denotes the ratio of predicted positive samples that are truly positive:
P = T P T P + F P
F1-Score ( F 1 ) is the harmonic mean of precision and recall, considering both metrics:
F 1 = 2 R e c a l l P r e c i s i o n R e c a l l + P r e c i s i o n
True Positive Rate ( T P R ) and False Positive Rate ( F P R ) are defined as follows:
T P R = T P T P + F N , F P R = F P T N + F P
The ROC curve plots True Positive Rate (TPR) against False Positive Rate (FPR) to evaluate model performance across different classification thresholds, with the Area Under the Curve ( A U C ) serving as a key performance indicator. In multi-class tasks, the “one vs. all” approach is used, and in this study, five ROC curves were calculated to assess the model’s ability to distinguish between categories.
The experimental flow of this study is shown in Figure 2.

3.4. Importance Measurement Method

The theory of importance measurement, a key area in reliability mathematics, analyzes the contribution of individual components to overall system performance. This study uses the Birnbaum importance metric to quantitatively assess the marginal contributions of satisfaction indicators to overall satisfaction. Based on probabilistic inference, Birnbaum evaluates feature importance by comparing the impact of features on the posterior probability distribution of the target variable, focusing on system reliability sensitivity to individual feature states (Hu et al., 2017), expressed by the following formula:
I B i t = h t p i t , i = 1 , 2 , , n
where h t represents the system reliability, and p i t denotes the state probability of the i -th feature.
In this study, all indicators adopt a five-point Likert scale (ranging from 1 to 5), indicating that each variable has five discrete states. To accommodate satisfaction evaluation with multistate features, importance measurement theory was extended from binary systems to multistate systems (Dang et al., 2023). The formula is as follows:
I B M C i S = 1 w i 1 j = 1 w i P S = 5 C i = j P S = 5
where S represents the target variable, namely, the predicted outcome of satisfaction; C i denotes a feature, with each variable having w possible states (corresponding to the Likert scale, w = 5 ); P ( S = 5 ) refers to the prior probability of the target variable; and P S = 5 C i = j is the posterior probability given C i = j .
This study applied importance measurement theory to the five-point Likert scale, extending traditional binary systems to more precisely quantify the marginal contributions of features across multistate settings (Song et al., 2024). An in-depth analysis of the seven critical dimensions was conducted to calculate each dimension’s contribution to overall satisfaction under different states.

4. Data Processing and Feature Selection

4.1. Satisfaction Score Dataset

Indicator system overview: the final questionnaire encompasses seven primary dimensions, including course quality (M. Wang & Wu, 2019), research projects (Belash et al., 2015), mentor guidance (Sheng et al., 2024), mentor’s role (Avramkova et al., 2021), faculty management (H. Zhang et al., 2022), academic enhancement (Pineda, 2013), and quality development (J. Zhang et al., 2015). These dimensions are further subdivided into 49 secondary indicators, resulting in a total of 49 feature variables (labeled X 1 to X 49 ), with overall educational satisfaction ( Y ) serving as the overarching evaluation variable.
The detailed indicator structure is presented in Table 2.
Data source: The dataset used in this study was derived from a satisfaction survey conducted among all postgraduate students at the university. A total of 9122 samples were collected. The questionnaire was designed based on a five-point Likert scale, encompassing 49 features and one target. The variable values range from 1 to 5, corresponding to five levels of satisfaction from “Very Dissatisfied” to “Very Satisfied”.
Data preprocessing: After data collection, a thorough quality check was conducted. Specifically, samples with missing values and redundant entries were excluded. It was ensured that all variable values remained within the range of 1 to 5, and any outliers were eliminated. After these steps, 8903 valid samples were retained, accounting for 97.6% of the total dataset.

4.2. Reliability and Validity Analysis of the Questionnaire

CFA was employed to validate the structural rationality of the questionnaire. Each dimension was analyzed individually, and when a dimension exhibited low value, further examination was conducted to identify problematic indicators. The results, as shown in Table 3, indicate that the questionnaire on educational satisfaction demonstrated good performance overall, though it varied across different dimensions.
Reliability performance: With the exception of the research projects dimension, the Cronbach’s α and CR values for all other dimensions exceeded 0.8, indicating strong internal consistency across the questionnaire as a whole. Key dimensions such as mentor guidance, mentor’s role, academic enhancement, and quality development demonstrated particularly outstanding reliability.
Validity performance: The AVE values for all dimensions surpassed 0.5, signifying that latent factors exhibited strong explanatory power over observed variables. Notably, the dimensions mentor guidance and mentor’s role achieved AVE values exceeding 0.85, reflecting well-designed constructs and strong consistency among indicators.
Applicability performance: KMO values for all dimensions exceeded 0.5, with most surpassing 0.9, confirming that the questionnaire data were well-suited for factor analysis. Additionally, Bartlett’s sphericity test yielded significant results across all dimensions ( p < 0.001 ), further validating the suitability.
Overall, the questionnaire achieved a Cronbach’s α of 0.9755 and a KMO value of 0.9823, demonstrating high rationality in its overall design. These results affirm its suitability as a comprehensive tool for evaluating postgraduate education satisfaction.

4.3. Feature Selection for the Evaluation System

To enhance the accuracy of the educational satisfaction evaluation, a two-stage feature optimization method was adopted to screen the initial 49 features.
Stage 1: Statistical Feature Pre-Screening
This stage integrates VT, CA, and MI analyses, providing an assessment of both the redundancy between features and the correlation with Y .
After the screening process, 33 features were retained from the original 49. In Stage 1, the excluded features primarily came from the following dimensions: course quality ( X 1 , X 2 ), research projects ( X 9 , X 10 ), mentor guidance ( X 11 , X 12 , X 13 , X 15 , X 16 ), mentor’s role ( X 20 , X 21 ), and faculty management ( X 27 , X 28 , X 29 , X 30 , X 31 ). Regarding reliability, the academic enhancement and quality development dimensions demonstrated particularly strong internal consistency, as reflected by their Cronbach’s α values of 0.9709 and 0.9772, respectively. In contrast, features from dimensions like course quality and research projects were less reliable (Cronbach’s α values of 0.9096 and 0.7924), which further justified the exclusion of certain features from these dimensions.
By replacing fixed thresholds with dynamic thresholds, this method adapts to the characteristics of the data distribution. This ensures that the selected features contribute significantly to the target variable while minimizing redundancy.
Stage 2: XGBoost Recursive Feature Elimination
The 33 features filtered in the first stage were used as input to construct an XGBoost classifier. The model’s built-in feature importance evaluation method was employed to calculate the contribution of each feature to the target based on their gain values.
The RFE process followed these steps: first, based on importance rankings, the feature with the lowest contribution was identified and removed, and the feature set was updated accordingly. Then, five-fold cross-validation was used to evaluate the model’s performance ( A c c ) after each removal. If there was no significant drop in the A c c metrics, the process continued iteratively until it stabilized. This determined the optimal subset.
Through five-fold cross-validation and parameter tuning, the optimal hyperparameters for the XGBoost model were determined as follows: a learning rate of 0.2, a max_depth of 3, and n_estimators set to 50. The second stage further refined the feature set by removing features such as X 6 , X 7 , X 19 , and X 33 , based on their contribution to model performance.
Figure 3 illustrates the relationship between the number of remaining features and the model’s performance during the RFE process, along with the importance score ranking of the 29 features. Among these, feature X 8 achieved the highest score, while feature X 49 ranked the lowest. This approach not only reduced redundancy but also improved the predictive efficiency and interpretability of the model.
The correlation heatmap and MI analysis results show that the 29 retained features exhibit low inter-feature correlations, indicating that the redundancy was effectively eliminated while preserving those with a significant discriminatory power for Y . Furthermore, the bar chart provides a visual representation of the information gain contributed by each feature, further validating the explanatory power of the selected features in relation to the target. These results demonstrate that the proposed method removed irrelevant and redundant features, ensuring the robustness of the final feature set.
To validate the effectiveness of the two-stage algorithm proposed in this study, experiments were conducted using three classifiers: ANN, GBDT, and SVM. Four datasets were selected for comparison: datasets processed by the ReliefF algorithm (Z. Huang et al., 2018), the PSO algorithm (P. Wang & Wang, 2024), the proposed two-stage algorithm, and the original dataset without feature selection. During the experiments, five-fold cross-validation and parameter optimization were employed to obtain the optimal hyperparameter configurations. The detailed settings are shown in Table A1 (see Appendix A).
By conducting comparative experiments on the four datasets, the results for the three models—ANN, GBDT, and SVM—on metrics such as A c c , F 1 , A U C , and the feature dimensions were obtained. The detailed results are presented in Table 4.
To visually demonstrate the performance differences of the various feature selection algorithms across models, a comparison chart was created, as shown in Figure 4.
From the perspective of classifiers, the ANN model performs best in terms of A c c and A U C , effectively capturing the non-linear relationships within the data. SVM follows, offering stable precision across different datasets. Although GBDT lags slightly behind the other two models, it stands out in terms of F 1 , indicating better performance on imbalanced datasets. This is because F 1 , as a harmonic mean of P and R , more accurately reflects the classifier’s performance on minority class samples.
From the perspective of feature selection algorithms, the proposed two-stage algorithm achieves the best performance across all classifiers, particularly in ANN and SVM, where it achieves higher A c c and A U C values. This further validates the proposed algorithm’s effectiveness in removing redundancy. In comparison, ReliefF retains too many features, which somewhat improves accuracy but fails to balance maximizing model performance and minimizing feature count. Meanwhile, although PSO selects fewer features, its performance in terms of A c c is slightly inferior, highlighting the trade-off between feature quantity and quality in selection processes. Using the original features directly yields the worst results, indicating the presence of redundancy and irrelevance in the raw data, underscoring the necessity of feature selection. Additionally, using only Stage 2 for feature selection resulted in a higher feature dimension and lower model accuracy compared to the proposed two-stage algorithm.
Additionally, Table A2 (see Appendix A) presents the prior probability for variable X 3 under different states of the target variable. These data preliminarily reveal the distribution characteristics of the variable, including its frequency, distribution proportions across each state, central tendency, and dispersion. These insights provide valuable reference points for subsequent causal relationship analyses and model optimization.

5. Educational Satisfaction Evaluation System

5.1. Tree Augmented Naive Bayesian Network

Based on the selected features, NB and TAN were further employed in this study to model and uncover the causal relationships within educational satisfaction, using overall satisfaction as the target variable to analyze the influence of satisfaction-related features under both frameworks. The dataset was split into training and testing sets in a 6:4 ratio, with 5341 samples used for training and 3562 samples used for testing.
First, overall satisfaction was set as the sole parent node. From Y , 29 directed edges were established, each pointing to one of the 29 attribute variables, thereby forming the structure of an NB network. The model learned the conditional probability distribution of each attribute variable using the training set and subsequently evaluated its performance on the testing set. The confusion matrix is shown in Table A3 (see Appendix A), providing insights into the model’s reliability and classification accuracy.
When the posterior probability threshold was set to 0.5, the overall accuracy of the NB model reached 68.48%. Figure 5 illustrates the NB network structure and its corresponding ROC curves, with the model’s A U C value reaching 88.1%. This indicates a reasonable level of overall classification performance.
However, NB assumes that all attribute variables are independent, an assumption that does not always hold true in practical satisfaction evaluations. To address this limitation, we introduced the TAN model, which fully accounts for mutual information among attribute variables. Compared to NB, the TAN algorithm incorporates associative edges, allowing it to model dependencies more effectively.
Figure 6 illustrates the TAN network structure along with its corresponding ROC curves. The A U C value reached 91.01%, significantly outperforming the NB.
In the TAN network, the overall satisfaction Y remains the sole parent node. However, beyond the directed edges extending from Y to the 29 attribute variables, TAN introduces associative edges between them, which reflect the correlations, resulting in a more complex network structure. For instance, the network topology reveals that X 32 (subject specialization) influences X 35 (independent academic research ability) and X 40 (multi-disciplinary knowledge), indicating that the value of X 32 affects the conditional probabilities of X 35 and X 40 . Such dependency relationships uncover the latent interactions between variables, providing more intuitive support for causal analysis.
Table 5 presents the confusion matrix for the TAN model. When the posterior probability threshold was set to 0.5, the model achieved an overall A c c of 78.64%. Specifically, the A c c for predicting the highest satisfaction category ( Y = 5) reached 86.33%, while the A c c for the lowest category ( Y = 1) improved to 60%.
The introduction of the TAN model not only achieved a higher classification accuracy but also offered deeper insights into the underlying causal relationships among variables. Compared to the NB model, TAN showed significant improvements in overall accuracy and reliability, further validating the positive impact of incorporating variable interdependencies on predictive performance.

5.2. Causal Model Comparison Experiment

To validate the effectiveness of Bayesian networks in predicting educational satisfaction, three commonly used machine learning algorithms—ANN, GBDT, and SVM—were applied to the same dataset, and their performance was compared with the NB and TAN models. The dataset division and the proportion of the training and testing sets followed the same configuration as the previous process.
Table 6 presents the performance of the five models on the test set. The TAN demonstrated superior performance across all evaluation metrics ( A c c , F 1 , and A U C ), significantly outperforming the other models. This result further confirms the overall advantage, indicating that the TAN model not only achieves a high classification accuracy but also effectively captures the latent dependencies among variables.
To provide a more intuitive comparison of the classification performance of each model, the experimental results are visualized in Figure 7.
The analysis reveals that the TAN model outperformed all other models across all metrics. From the perspectives of feature selection and causal analysis, the superior performance of the TAN lies in its ability to account for correlations between features, enabling it to capture potential dependencies among variables more effectively than other models. Among traditional models, ANN achieves high A c c and A U C values but exhibits lower F 1 , indicating limited capability in handling multi-class balance issues. While SVM’s is close to TAN’s, its overall classification A c c is slightly inferior. Additionally, GBDT excels in F 1 , reflecting its strength in handling imbalanced data, but its low A U C value highlights its limited ability to differentiate between positive and negative samples.
Figure 8 further illustrates the ROC curves of the four models. It is evident that the TAN model’s curve consistently surpasses those of the other models, particularly in the low false positive rate region, demonstrating its significant advantage in identifying positive samples. In contrast, the GBDT model’s curve is noticeably lower than the others. The ROC curves of SVM and ANN are relatively close but still exhibit a clear gap compared to TAN, highlighting TAN’s superior performance.
In summary, the TAN model not only demonstrates exceptional performance but also effectively leverages the correlations among variables. This advantage enables this model to stand out in the task of predicting satisfaction, showcasing its robust capability in multi-class classification problems.

5.3. Importance Ranking of Influencing Factors

This experiment combined causal inference through Bayesian networks with the Birnbaum importance metric to quantitatively rank the key factors influencing satisfaction. By using Y = 5 (indicating high satisfaction) as the reliability metric for prediction, the prior and posterior probabilities of the attribute variable under different states were calculated to quantitatively analyze the impact of each feature on the target. The results not only reveal the contribution of each feature to changes in satisfaction but also provide scientific evidence for improving education quality and optimizing enrollment strategies.
The study assessed the impact of 29 features and ranked them based on Birnbaum importance. Table 7 presents the top six features, along with their corresponding prior probabilities, posterior probabilities, and importance values. The experimental results indicate that academic resilience ( X 39 ), academic aspirations ( X 37 ), dedication and service spirit ( X 48 ), creative ability ( X 34 ), academic standards ( X 38 ), and independent academic research ability ( X 35 ) are the top six critical factors influencing changes in satisfaction.
By analyzing the top-ranked features, it was found that the dimensions of academic enhancement, quality development, and mentor’s role contribute most significantly to educational satisfaction.
Academic enhancement is the core objective of graduate education, including indicators such as academic resilience ( X 39 ), academic aspirations ( X 37 ), and academic standards ( X 38 ). These top-ranked features highlight the fact that students prioritize the development of their academic capabilities, reflecting a strong desire to achieve breakthroughs in academia through graduate education. Notably, academic resilience ranked first, with an importance score of 0.39115, emphasizing its critical role in shaping satisfaction.
Quality development, represented by dedication and service spirit ( X 48 ) and creative ability ( X 34 ), also emerged as being highly significant. X 48 ranked third with an importance score of 0.38979, further highlighting the critical role of postgraduate education in nurturing students’ overall competencies, especially in areas such as social responsibility. Mentor’s role is prominently reflected in the cultivation of academic interest ( X 18 ), underscoring the pivotal role of mentors in guiding students to conduct independent academic research and improve logical thinking skills.
As shown in Figure 9, the importance levels of influencing factors under Y = 5 (high satisfaction) and Y = 4 (moderate satisfaction) states were compared. Significant differences were observed in the impact of various features on the target variable, especially under high satisfaction states, where certain features demonstrated particularly strong contributions. Taking X 39 as an example, its posterior probability exhibited fluctuations across different states, indicating high sensitivity. When X 39 was in state 5, the posterior probability of Y = 5 reached 0.74419, which was substantially higher than other states.
The state transition trends reveal that for most key features, as the state transitions from low to high scores, the posterior probability of the target variable shows a significant upward trend. For instance, the posterior probability of X 39 gradually increases from state 1 to state 5, reflecting that achieving a high level of academic resilience is critical for enhancing satisfaction.

5.4. Optimal Strategy Identification for Satisfaction Factors

Compared to merely predicting satisfaction levels, optimizing the state combinations of key indicators offers a more effective approach to enhancing satisfaction. However, traditional methods, while providing general directions, often lack specific and actionable recommendations. To address this gap, this study introduces a combination search-based optimization strategy to identify the optimal state combinations for improving educational satisfaction, offering evidence-based guidance for educational improvements.
This research involved a comprehensive analysis of all possible state combinations of 29 key indicators, and the posterior probability of each combination was calculated for Y = 5. With each indicator containing five possible states, the theoretical search space comprises 5 29 possible state combinations. Given this massive computational space, an efficient algorithm was employed to calculate and filter systematically.
The optimization goal was set as a posterior probability of no less than 95% ( P Y = 5 ≥ 95%), and combinations meeting this criterion were ranked in descending order.
Table 8 presents the recommended state combinations for Y = 5 (partial results). Among all qualifying combinations, 722 state combinations met the criterion. For clarity, the top eight combinations with the highest posterior probabilities are shown. This method ensures that actionable insights are derived from the data, bridging the gap between theoretical modeling and practical implementation. By identifying precise strategies for improving satisfaction, it provides valuable guidance for improving postgraduate education quality and enhancing overall satisfaction levels.
Based on the recommendations in Table 8, the optimal combination to achieve the highest satisfaction level corresponds to a specific state configuration, which results in a posterior probability of 99.961%. This demonstrates that optimizing the 29 key indicators can significantly elevate educational satisfaction to near-perfect levels.
Further analysis reveals the following key insights:
(1)
Characteristics of high posterior probability combinations: It is evident that most features in the recommended combinations are in high-level-score states. These states reflect students’ prioritization of excellence in academic support, which highlights the critical areas where education administrators should focus their improvement efforts.
(2)
Priority of key indicators: In the top eight recommended combinations, certain indicators exhibit consistent high states (e.g., X 8 and X 37 remain in state 5), while others, such as X 17 , show more variability in state configurations. This suggests that certain indicators play a more significant role in achieving high satisfaction levels. For instance, consistent high states in indicators underscore their pivotal role in driving satisfaction.
(3)
Practicality of optimization strategies: These recommended combinations provide actionable insights for educational administrators. For example, maintaining the state of academic resilience ( X 39 ) at a high level is critical, while indicators related to the mentor’s role and quality development (e.g., X 8 , X 37 ) also require focused attention.
These findings offer a clear framework for enhancing the educational experience for postgraduates. By focusing on maintaining high-level states in critical indicators and prioritizing resources toward these areas, administrators can systematically improve student satisfaction, ultimately contributing to the overall quality of postgraduate education.

6. Discussion

In this study, we introduce an innovative research methodology integrating Bayesian networks with feature importance analysis to quantitatively assess how behavioral and contextual factors interact to shape postgraduate education satisfaction. This approach identifies key indicators and clarifies their specific impact on satisfaction. The research is based on rigorous data from 8903 valid questionnaires, employing statistical validation methods, including CFA, to ensure reliability and scientific rigor. A posterior probability target combination search strategy, with a threshold of at least 90%, identifies the optimal combination of key indicators, offering actionable optimization suggestions.
Theoretically and practically, this research fills the gap in studying postgraduate education satisfaction by applying causal modeling and feature quantification, offering a systematic analytical framework for multi-indicator, multi-state complex systems. By integrating Bayesian networks with Birnbaum’s importance theory, it comprehensively analyzes the impact of various indicators on satisfaction. The findings highlight academic enhancement, quality improvement, and the mentor’s role as core dimensions, with key behavioral indicators such as academic resilience and academic aspirations.
From an educational practice perspective, postgraduates’ core expectations center on the improvement of academic abilities and comprehensive qualities. The primary goal of postgraduate education is to cultivate students’ academic innovation and research capabilities. Academic resilience directly influences students’ evaluations. The importance of academic aspirations and commitment to service suggests that students value the realization of intrinsic goals alongside their broader responsibilities to society. The significant impact of the mentor’s role highlights the critical position of faculty guidance, serving as transmitters of academic knowledge and key providers of career development.
Compared to previous studies, this research enriches the methods of quantitative analysis for educational satisfaction. Traditional studies often rely on regression analysis or analytic hierarchy processes, focusing primarily on the direct effects of individual indicators while overlooking the interdependencies among them. This study addresses this gap by employing TAN to comprehensively model the behavioral and cognitive mechanisms underlying student satisfaction, revealing how students’ engagement patterns, perceived support, and learning motivations interact to shape their overall experience. The integration of the Birnbaum importance not only quantitatively evaluates the importance of individuals but also identifies optimized strategies through state combination searches, providing a novel pathway for precise enhancement.

7. Conclusions

Based on the findings, the following management recommendations are proposed:
(1)
Focus on academic quality development: Universities should optimize course designs and provide more opportunities for research training that encourage active learning behaviors, foster academic resilience, and enhance intrinsic motivation. This can be achieved through structured mentorship programs, stress management workshops, and resilience-building curricula, enabling students to develop sustained engagement.
(2)
Enhance the role of mentorship: Universities should prioritize the recruitment and training of mentors, emphasizing their role in shaping students’ learning behaviors, career decision-making, and psychological well-being. In particular, mentorship programs should be designed to strengthen academic resilience and provide support for students in overcoming academic challenges.
(3)
Promote comprehensive quality education: Institutions should diversify course offerings and incorporate behavioral interventions to foster collaborative learning and real-world engagement, ultimately strengthening students’ sense of social responsibility. Strengthening academic aspirations and service spirit can be facilitated through public service fellowships and career planning support.
(4)
Institutionalize support for creative and independent research: Universities should expand research assistantship opportunities and provide interdisciplinary project grants, supporting students’ development of independent research skills.
This study has certain areas for improvement, which future research can address:
(1)
Sample size and diversity: Future research can incorporate samples from diverse regions and disciplines to further validate the model’ s applicability and generalizability.
(2)
Inclusion of new features: Exploring psychological factors related to satisfaction, such as learning stress and sense of belonging, could enrich the evaluation framework.
(3)
Algorithm and model improvement: Incorporating advanced algorithms such as deep learning or more efficient combinatorial search methods could enhance the model’ s computational efficiency in handling ultra-large-scale data and improve its ability to analyze nonlinear relationships among features.
In summary, this study integrated behavioral insights with causal modeling to offer a novel approach to educational satisfaction evaluation. We identified key satisfaction drivers and provided actionable strategies to enhance educational quality. The findings serve as a valuable reference for policy management, supporting the development of targeted, behaviorally informed educational improvements.

Author Contributions

Conceptualization, S.L. and T.W.; methodology, Z.C.; software, T.W.; validation, Z.C. and S.L.; formal analysis, T.W. and S.D.; investigation, Z.C.; resources, H.Y. and S.D.; data curation, H.Y.; writing—original draft preparation, S.L.; writing—review and editing, T.W.; visualization, T.W. and S.D.; supervision, Z.C.; project administration, H.Y.; funding acquisition, S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Degree and Postgraduate Education Research Project of Shaanxi Province [SXGERC2023020], the Practice and Innovation Funds for Graduate Students of Northwestern Polytechnical University [PF202505], and the Distinguished Young Scholar Program of Shaanxi Province [2023-JQ-JC-10].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The experimental data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Optimal hyperparameter settings for each algorithm.
Table A1. Optimal hyperparameter settings for each algorithm.
AlgorithmParameterSetting Value
GBDTn_estimators150
max_depth3
learning_rate0.01
SVM C 0.1
gammascale
kernellinear
ANNhidden_layer_sizes50,50
activationrelu
solveradam
PSOpopulation_size50
max_iterations0.01
ReliefF k 20
Table A2. Modeling prior probability table ( X 3 ).
Table A2. Modeling prior probability table ( X 3 ).
Feature VariableState Y StateFrequencyPercentage
X 3 112632.50%
22414.20%
3332.68%
4591.38%
5411.30%
2156.25%
21810.65%
3786.34%
4591.38%
560.19%
311620%
27443.79%
356145.61%
464415.05%
5732.32%
411721.25%
24325.44%
341033.33%
4228753.46%
533410.62%
511620%
2105.92%
314812.03%
4122928.73%
5269285.57%
Table A3. Confusion matrix for NB model.
Table A3. Confusion matrix for NB model.
1 (54)2 (112)3 (594)4 (1563)5 (1239)
Confusion Matrix
1 (32)138344
2 (68)9322250
3 (492)185629211313
4 (1711)7152681162259
5 (1259)719279963
Reliability
1 (32)24.07%7.14%0.51%0.26%0.32%
2 (68)16.67%28.57%3.70%0.32%0.00%
3 (492)33.33%50.00%49.16%7.23%1.05%
4 (1711)12.96%13.39%45.12%74.34%20.90%
5 (1259)12.96%0.89%1.52%17.85%77.72%
Accuracy
1 (32)40.63%25.00%9.38%12.50%12.50%
2 (68)13.24%47.06%32.35%7.35%0.00%
3 (492)3.66%11.38%59.35%22.97%2.64%
4 (1711)0.41%0.88%15.66%67.91%15.14%
5 (1259)0.56%0.08%0.71%22.16%76.49%

References

  1. Al-Dmour, Y. (2024). Beyond academia: Investigating indoor environmental quality and its impact on postgraduate student satisfaction. Results in Engineering, 24, 103190. [Google Scholar] [CrossRef]
  2. Allanqawi, T., Alkadhimi, A., & Fleming, P. S. (2023). Postgraduate orthodontic education: An international perspective on content and satisfaction levels. Journal of the World Federation of Orthodontists, 12(6), 239–244. [Google Scholar] [CrossRef] [PubMed]
  3. Avramkova, I. S., Anufrieva, N. I., Kamyanets, A. V., Kuznetsova, E. O., & Scherbakova, A. I. (2021). Monitoring studies of university graduates’ satisfaction with the quality of education: The main approaches. Revista on line de Política e Gestão Educacional, 25, 638–653. [Google Scholar] [CrossRef]
  4. Belash, O., Popov, M., Ryzhov, N., Ryaskov, Y., Shaposhnikov, S., & Shestopalov, M. (2015). Research on university education quality assurance: Methodology and results of stakeholders’ satisfaction monitoring. Procedia-Social and Behavioral Sciences, 214, 344–358. [Google Scholar] [CrossRef]
  5. Bush, C. T., & Lowery, B. (2016). Postgraduate nurse practitioner education: Impact on job satisfaction. The Journal for Nurse Practitioners, 12(4), 226–234. [Google Scholar] [CrossRef]
  6. Chen, J., Li, B., Wang, J., Zhao, Y., Yao, L., & Xiong, Y. (2020). Knowledge graph enhanced third-party library recommendation for mobile application development. IEEE Access, 8, 42436–42446. [Google Scholar] [CrossRef]
  7. Chen, Q., Meng, Z., Liu, X., Jin, Q., & Su, R. (2018). Decision variants for the automatic determination of optimal feature subset in RF-RFE. Genes, 9(6), 301. [Google Scholar] [CrossRef]
  8. Crede, E., & Borrego, M. (2014). Understanding retention in US graduate programs by student nationality. Studies in Higher Education, 39(9), 1599–1616. [Google Scholar] [CrossRef]
  9. Czarnecki, K., & Litwiński, M. (2024). Cross-country relationships between student funding policies and student labour. Higher Education Policy, 37(2), 366–391. [Google Scholar] [CrossRef]
  10. Dang, C., Valdebenito, M. A., Faes, M. G., Song, J., Wei, P., & Beer, M. (2023). Structural reliability analysis by line sampling: A Bayesian active learning treatment. Structural Safety, 104, 102351. [Google Scholar] [CrossRef]
  11. Erichsen, E. A., Bolliger, D. U., & Halupa, C. (2014). Student satisfaction with graduate supervision in doctoral programs primarily delivered in distance education settings. Studies in Higher Education, 39(2), 321–338. [Google Scholar] [CrossRef]
  12. Gu, J., He, C., & Liu, H. (2017). Supervisory styles and graduate student creativity: The mediating roles of creative self-efficacy and intrinsic motivation. Studies in Higher Education, 42(4), 721–742. [Google Scholar] [CrossRef]
  13. Guo, S., & Li, M. (2018, October 9–11). Research on satisfaction of postgraduate education based on structural equation model. 2018 3rd International Conference on Politics, Economics and Law (ICPEL 2018), Weihai, China. [Google Scholar]
  14. Hazzam, J., Wilkins, S., Southall, C., & Ibrahim, B. (2024). The influence of LinkedIn group community on postgraduate student experience, satisfaction and grades. Computers & Education, 216, 105052. [Google Scholar]
  15. Hu, L., Liu, Y., Lohse, N., Tang, R., Lv, J., Peng, C., & Evans, S. (2017). Sequencing the features to minimise the non-cutting energy consumption in machining considering the change of spindle rotation speed. Energy, 139, 935–946. [Google Scholar] [CrossRef]
  16. Huang, R., Zhang, S., Bai, X., & Xu, C. (2014). Multi-level structuralized model-based definition model based on machining features for manufacturing reuse of mechanical parts. The International Journal of Advanced Manufacturing Technology, 75, 1035–1048. [Google Scholar] [CrossRef]
  17. Huang, Z., Yang, C., Zhou, X., & Huang, T. (2018). A hybrid feature selection method based on binary state transition algorithm and ReliefF. IEEE Journal of Biomedical and Health Informatics, 23(5), 1888–1898. [Google Scholar] [CrossRef]
  18. Muijs, D., & Bokhove, C. (2017). Postgraduate student satisfaction: A multilevel analysis of PTES data. British Educational Research Journal, 43(5), 904–930. [Google Scholar] [CrossRef]
  19. Pérez Fuentes, C. A., Alvarez Maestre, A. J., Cardona Rivas, A. M., Aguilar Barreto, A. J., & Sanabria Alarcón, R. K. (2023). Job satisfaction and teacher education: Correlational study in postgraduate graduates in education. Education Sciences, 13(2), 198. [Google Scholar] [CrossRef]
  20. Pineda, A. P. M. (2013). Total quality management in educational institutions: Influences on customer satisfaction. Asian Journal of Management Sciences and Education, 2(3), 31–46. [Google Scholar]
  21. Ramia, G., Mitchell, E., Morris, A., Wilson, S., Hastings, C., & Davies, J. (2022). Explaining government policy inaction on international student housing in Australia: The perspectives of stakeholders. Higher Education Policy, 37(1), 21–39. [Google Scholar] [CrossRef]
  22. Sharif, S., Shekarchizadeh, A., Djauhari, M. A., & Rasli, A. (2012). Correlation network analysis of international postgraduate students’ satisfaction in top malaysian universities: A robust approach. Modern Applied Science, 6(12), 91. [Google Scholar] [CrossRef]
  23. Sheng, W., Fan, Z., & Weng, S. (2024). Enhancing student satisfaction in educational management: A Bayesian analysis of influential factors and improvement strategies. Journal of the Knowledge Economy, 1–38. [Google Scholar] [CrossRef]
  24. Si, S., Liu, M., Jiang, Z., Jin, T., & Cai, Z. (2019). System reliability allocation and optimization based on generalized Birnbaum importance measure. IEEE Transactions on Reliability, 68(3), 831–843. [Google Scholar] [CrossRef]
  25. Sneyers, E., & De Witte, K. (2017). The effect of an academic dismissal policy on dropout, graduation rates and student satisfaction. Evidence from the Netherlands. Studies in Higher Education, 42(2), 354–389. [Google Scholar] [CrossRef]
  26. Song, J., Cui, Y., Wei, P., Valdebenito, M. A., & Zhang, W. (2024). Constrained Bayesian optimization algorithms for estimating design points in structural reliability analysis. Reliability Engineering & System Safety, 241, 109613. [Google Scholar]
  27. Stewart, B., Speldewinde, P., & Ford, B. (2018). Influence of improved teaching practices on student satisfaction ratings for two undergraduate units at an Australian university. Assessment & Evaluation in Higher Education, 43(4), 598–611. [Google Scholar]
  28. Tanveer, M., Verma, S., Sharma, R., Goel, T., & Suganthan, P. N. (2023, June 18–23). Weighted kernel ridge regression based randomized network for alzheimer’s disease diagnosis using susceptibility weighted images. 2023 International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia. [Google Scholar]
  29. Walker, D. J., & Palmer, E. (2011). The relationship between student understanding, satisfaction and performance in an Australian engineering programme. Assessment & Evaluation in Higher Education, 36(2), 157–170. [Google Scholar]
  30. Wang, M., & Wu, J. (2019). Identification of plastic properties of metal materials using spherical indentation experiment and Bayesian model updating approach. International Journal of Mechanical Sciences, 151, 733–745. [Google Scholar] [CrossRef]
  31. Wang, P., & Wang, Z. (2024). Evaluating the effectiveness of the binary PSO method in feature selection to improve the detection of android botnets. International Journal of Advanced Computer Science & Applications, 15(10), 605. [Google Scholar]
  32. Wang, Y., Yu, S., Ma, N., Wang, J., Hu, Z., Liu, Z., & He, J. (2020). Prediction of product design decision making: An investigation of eye movements and EEG features. Advanced Engineering Informatics, 45, 101095. [Google Scholar] [CrossRef]
  33. Weber, P., Medina-Oliva, G., Simon, C., & Iung, B. (2012). Overview on Bayesian networks applications for dependability, risk analysis and maintenance areas. Engineering Applications of Artificial Intelligence, 25(4), 671–682. [Google Scholar] [CrossRef]
  34. Xiao, J., Anwer, N., Li, W., Eynard, B., & Zheng, C. (2022). Dynamic Bayesian network-based disassembly sequencing optimization for electric vehicle battery. CIRP Journal of Manufacturing Science and Technology, 38, 824–835. [Google Scholar] [CrossRef]
  35. Zhang, H., Zhang, S., Zhang, Y., Liang, J., & Wang, Z. (2022). Machining feature recognition based on a novel multi-task deep learning network. Robotics and Computer-Integrated Manufacturing, 77, 102369. [Google Scholar] [CrossRef]
  36. Zhang, J., Xu, Z., Li, Y., & Jiang, S. (2015). Framework for the integration of assembly modeling and simulation based on assembly feature pair. The International Journal of Advanced Manufacturing Technology, 78, 765–780. [Google Scholar] [CrossRef]
  37. Zhang, X., Liu, J., Zhang, C., Shao, D., & Cai, Z. (2023). Innovation performance prediction of university student teams based on bayesian networks. Sustainability, 15(3), 2335. [Google Scholar] [CrossRef]
  38. Zhang, Y., Zhang, Z., Chen, X., & Xie, H. (2024). Enhancing postgraduate education quality: A comprehensive analysis of factors influencing students’ satisfaction in university-business cooperation. Innovations in Education and Teaching International, 1–18. [Google Scholar] [CrossRef]
Figure 1. NB and TAN structures.
Figure 1. NB and TAN structures.
Behavsci 15 00559 g001
Figure 2. The experimental flow.
Figure 2. The experimental flow.
Behavsci 15 00559 g002
Figure 3. XGBoost RFE iteration trends and feature importance ranking.
Figure 3. XGBoost RFE iteration trends and feature importance ranking.
Behavsci 15 00559 g003
Figure 4. Comparison of feature selection algorithms in different models.
Figure 4. Comparison of feature selection algorithms in different models.
Behavsci 15 00559 g004
Figure 5. NB network structure and ROC curves.
Figure 5. NB network structure and ROC curves.
Behavsci 15 00559 g005
Figure 6. TAN network structure and ROC curves.
Figure 6. TAN network structure and ROC curves.
Behavsci 15 00559 g006
Figure 7. Comparison chart of different metrics across different classifiers.
Figure 7. Comparison chart of different metrics across different classifiers.
Behavsci 15 00559 g007
Figure 8. ROC curves for four classifiers.
Figure 8. ROC curves for four classifiers.
Behavsci 15 00559 g008aBehavsci 15 00559 g008b
Figure 9. Importance of influencing factors in Y = 5 vs. Y = 4 states.
Figure 9. Importance of influencing factors in Y = 5 vs. Y = 4 states.
Behavsci 15 00559 g009
Table 1. TAN modeling process.
Table 1. TAN modeling process.
Input Dataset: Acquire the training dataset D .
1. Calculate Conditional Mutual Information: I P D A i ; A j C , where i j .
2. Construct a Complete Undirected Graph: Each feature A 1 , A 2 , , A n corresponds to a node, and the weight of the edge between any two features is determined by their conditional MI.
3. Build the Maximum Weight Spanning Tree.
4. Select a Root Attribute and Orient the Tree.
5. Add the Class Node: Add a node C to the directed tree, and introduce an arc to each A i .
6. Construct the TAN Model.
Table 2. Postgraduate education satisfaction indicator system.
Table 2. Postgraduate education satisfaction indicator system.
DimensionIndicatorVariable
Course QualityTeaching Quality X 1
Course Difficulty X 2
Enhancement of Ideology and Morality X 3
Enriching Humanistic Qualities X 4
Strengthen Professional Knowledge X 5
Understand the Frontiers of Science X 6
Learning Research Methods X 7
Research ProjectsDifficulty of Research Tasks X 8
Number of Research Projects X 9
Enhancement of Research Capacity X 10
Mentor GuidancePolitical Quality X 11
Teacher Ethics X 12
Mentoring Ability X 13
Mentoring Frequency X 14
Academic Level X 15
Practical Ability X 16
Mentor’s RoleCultivate Ideal Beliefs X 17
Stimulate Academic Interest X 18
Enhancement of Research Ability X 19
Correcting the Attitude of Scholarship X 20
Comply with Academic Standards X 21
Clarify Career Planning X 22
Faculty
Management
Faculty Service Evaluation X 23
Evaluation of Faculty Atmosphere X 24
Scholarship System X 25
Three-assistant Positions X 26
Library X 27
Cafeteria X 28
Accommodation X 29
Mental Health Counseling X 30
Career Guidance and Services X 31
Academic
Enhancement
Subject Specialization X 32
Professional Skills X 33
Creative Ability X 34
Independent Academic Research Ability X 35
Academic Writing Ability X 36
Academic Aspirations X 37
Academic Standards X 38
Academic Resilience X 39
Quality
Development
Multi-disciplinary Knowledge X 40
Willingness to Serve the Country X 41
Interpersonal Skills X 42
Public Speaking Skills X 43
Organizational and Leadership Skills X 44
Time Management Skills X 45
Teamwork Ability X 46
Ability to Understand National Conditions X 47
Dedication and Service Spirit X 48
International Exchange Ability X 49
Overall EvaluationOverall Educational Satisfaction Y
Note: X 11 refers to the mentor’s ability to provide ideological and ethical leadership. X 20 indicates guidance on ethical research. X 26 refers to roles such as teaching, research, and administrative assistants, offering financial support and practical experience.
Table 3. Reliability and validity testing results of the scale.
Table 3. Reliability and validity testing results of the scale.
DimensionVariableCronbach’s α CRAVEKMOBartlett χ 2 p -Value
Course Quality X 1 X 7 0.90960.8620.6600.9021314.001.01 × 10−280
Research Projects X 8 X 10 0.79240.6790.5040.5143716.640
Mentor Guidance X 11 X 16 0.97950.8570.8910.927176.822.560 × 10−36
Mentor’s Role X 17 X 22 0.98000.8570.8930.923225.589.491 × 10−47
Faculty Management X 23 X 31 0.92810.8950.6110.9082411.890
Academic Enhancement X 32 X 39 0.97090.8880.8150.949548.942.38 × 10−114
Quality Development X 40 X 49 0.97720.9090.8130.965367.601.111 × 10−73
Overall Scale X 1 X 49 0.97550.9780.5190.98223,698.650
Table 4. Comparative experimental results of feature selection algorithms.
Table 4. Comparative experimental results of feature selection algorithms.
DatasetModel A c c F 1 A U C Feature Dimensions
Phase IIANN0.7790620.6018170.91884329
GBDT0.6926320.6984610.874274
SVM0.7786140.6682410.909957
ReliefFANN0.7675460.5929020.88282040
GBDT0.6877600.6796120.863432
SVM0.7631610.6553320.899135
PSOANN0.7571480.5672430.89112820
GBDT0.6786640.6678870.811622
SVM0.7709150.6580750.887956
OriginalANN0.7616510.5520870.87139349
GBDT0.6571980.6591400.862437
SVM0.7595120.6416840.877895
Table 5. Confusion matrix for TAN model.
Table 5. Confusion matrix for TAN model.
1 (15)2 (29)3 (390)4 (1825)5 (1303)
Confusion Matrix
1 (36)99846
2 (67)21241120
3 (498)4727319618
4 (1732)01671446218
5 (1229)0011671061
Reliability
1 (36)60.00%31.03%2.05%0.22%0.46%
2 (67)13.33%41.38%10.51%0.66%0.00%
3 (498)26.67%24.14%70.00%10.74%1.38%
4 (1732)0.00%3.45%17.18%79.23%16.73%
5 (1229)0.00%0.00%0.26%9.15%81.43%
Accuracy
1 (36)25.00%25.00%22.22%11.11%16.67%
2 (67)2.99%17.91%61.19%17.91%0.00%
3 (498)0.80%1.41%54.82%39.36%3.61%
4 (1732)0.00%0.06%3.87%83.49%12.59%
5 (1229)0.00%0.00%0.08%13.59%86.33%
Table 6. Experimental results comparing different classifiers.
Table 6. Experimental results comparing different classifiers.
ClassifierAccF1AUC
TAN0.7863560.7787320.910053
NB0.6847620.5317850.881002
ANN0.7829870.5704260.892807
GBDT0.6965190.7013280.756751
SVM0.7681080.7567800.908908
Table 7. Importance ranking of influencing factors (top six).
Table 7. Importance ranking of influencing factors (top six).
VariableStatePrior ProbabilityPosterior ProbabilityImportanceRanking
X 39 10.008310.027030.391151
20.028530.03150
30.142650.02913
40.443330.15201
50.377180.74419
X 37 10.018200.024690.390472
20.035270.02548
30.157810.04484
40.429970.16588
50.358760.76268
X 48 10.008090.027780.389793
20.036840.03963
30.146130.03228
40.438500.15471
50.370440.75349
X 34 10.007530.044780.385794
20.038300.01760
30.178370.05542
40.438390.18012
50.337410.78096
X 38 10.005620.060000.384795
20.027180.02479
30.137260.03110
40.451650.14822
50.378300.74317
X 35 10.005170.043480.383776
20.025270.02222
30.145570.04861
40.459280.15847
50.364710.74777
Table 8. Recommended table of indicator combinations under Y = 5 ( P 95%).
Table 8. Recommended table of indicator combinations under Y = 5 ( P 95%).
P ( Y = 5 )99.961%99.960%99.959%99.946%99.929%99.929%99.929%99.919%
X 3 23341333
X 4 24122415
X 5 53345241
X 8 45555555
X 14 35335452
X 17 24421533
X 18 21222115
X 22 55555555
X 23 55335552
X 24 55225555
X 25 51155555
X 26 25352141
X 32 53223253
X 34 34541114
X 35 12422241
X 36 55352545
X 37 55555555
X 38 55214422
X 39 33551545
X 40 13321514
X 41 12223531
X 42 21454133
X 43 14131444
X 44 12143133
X 45 43321131
X 46 52212544
X 47 52151414
X 48 21241134
X 49 23422252
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, S.; Wang, T.; Yin, H.; Ding, S.; Cai, Z. Behavioral Analysis of Postgraduate Education Satisfaction: Unveiling Key Influencing Factors with Bayesian Networks and Feature Importance. Behav. Sci. 2025, 15, 559. https://doi.org/10.3390/bs15040559

AMA Style

Li S, Wang T, Yin H, Ding S, Cai Z. Behavioral Analysis of Postgraduate Education Satisfaction: Unveiling Key Influencing Factors with Bayesian Networks and Feature Importance. Behavioral Sciences. 2025; 15(4):559. https://doi.org/10.3390/bs15040559

Chicago/Turabian Style

Li, Sheng, Ting Wang, Hanqing Yin, Shuai Ding, and Zhiqiang Cai. 2025. "Behavioral Analysis of Postgraduate Education Satisfaction: Unveiling Key Influencing Factors with Bayesian Networks and Feature Importance" Behavioral Sciences 15, no. 4: 559. https://doi.org/10.3390/bs15040559

APA Style

Li, S., Wang, T., Yin, H., Ding, S., & Cai, Z. (2025). Behavioral Analysis of Postgraduate Education Satisfaction: Unveiling Key Influencing Factors with Bayesian Networks and Feature Importance. Behavioral Sciences, 15(4), 559. https://doi.org/10.3390/bs15040559

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop