Next Article in Journal
Visual Footprint of Separation Through Membrane Distillation on YouTube
Next Article in Special Issue
SAPEx-D: A Comprehensive Dataset for Predictive Analytics in Personalized Education Using Machine Learning
Previous Article in Journal / Special Issue
Stress Factors in Higher Education: A Data Analysis Case
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Bayesian State-Space Approach to Dynamic Hierarchical Logistic Regression for Evolving Student Risk in Educational Analytics

Department of Mathematics, Natural Sciences and Technology Education, Faculty of Education, University of the Free State, 205 Nelson Mandel Drive, Bloemfontein 9300, South Africa
Data 2025, 10(2), 23; https://doi.org/10.3390/data10020023
Submission received: 11 January 2025 / Revised: 26 January 2025 / Accepted: 2 February 2025 / Published: 7 February 2025

Abstract

:
Early detection of academically at-risk students is crucial for designing timely interventions that improve educational outcomes. However, many existing approaches either ignore the temporal evolution of student performance or rely on “black box” models that sacrifice interpretability. In this study, we develop a dynamic hierarchical logistic regression model in a fully Bayesian framework to address these shortcomings. Our method leverages partial pooling across students and employs a state-space formulation, allowing each student’s log-odds of failure to evolve over multiple assessments. By using Markov chain Monte Carlo for inference, we obtain robust posterior estimates and credible intervals for both population-level and individual-specific effects, while posterior predictive checks ensure model adequacy and calibration. Results from simulated and real-world datasets indicate that the proposed approach more accurately tracks fluctuations in student risk compared to static logistic regression, and it yields interpretable insights into how engagement patterns and demographic factors influence failure probability. We conclude that a Bayesian dynamic hierarchical model not only enhances prediction of at-risk students but also provides actionable feedback for instructors and administrators seeking evidence-based interventions.

1. Introduction

Over the past decade, educational institutions have increasingly invested in data-driven practices to identify and support students who are at risk of failing [1,2,3]. As student data collection has expanded, ranging from demographics, attendance, and grades to detailed digital traces of online engagement—researchers have begun exploring how best to harness these data for early-warning systems [4,5]. The overarching goal is to detect academic challenges at an early stage, allowing for targeted interventions that can significantly improve student success. However, despite a growing body of literature on predictive modeling in education, important gaps remain in both methodological rigor and practical implementation.
One prominent line of work employs classical regression techniques, including standard (fixed-effects) logistic regression, to predict whether a student will pass or fail a course or an exam [6,7,8]. For instance, early studies used multiple logistic regression on a variety of course-level predictors—test scores, homework completion, and demographic factors—to estimate a failing probability for each student. These methods demonstrated that even simple statistical models could achieve moderate accuracy, providing a basis for automated alerts and basic interventions. Yet these models often assume that the probability of failing is static over time or, at best, incorporate only limited temporal information (such as cumulative GPA or a single measure of online engagement). This focus on static features fails to account for the dynamics of student learning, where performance may improve or decline gradually based on evolving circumstances. As a result, traditional logistic models often fail to capture important changes in students’ risk profiles across multiple assessments or time points [9].
In parallel, researchers in educational data mining and learning analytics have explored more advanced methods—such as random forests, gradient boosting, and deep neural networks—to predict at-risk students. By leveraging large volumes of data, including clickstream records from learning management systems, these methods can detect nonlinearities and complex interactions among features that simpler approaches might overlook [4,5,10]. However, their “black box” nature frequently leads to a lack of interpretability, undermining trust and adoption by instructors, administrators, and policymakers. Moreover, purely data-driven, high-capacity models often do not incorporate explicit time evolution constraints or hierarchical structures [11] that reflect the nested nature of educational data (students nested within classes, classes nested within institutions). As a result, these models may produce predictions that fail to pool information effectively across students, particularly when some students have relatively sparse data or exhibit atypical learning trajectories.
A third stream of literature has begun to explore time-series approaches, particularly state-space or Markov models, for academic risk prediction. These models recognize that a student’s latent “ability” or “risk level” may shift slowly over time [12,13,14]. While promising, much of this research has either been confined to small-scale simulations or restricted to a few narrow domains (e.g., language-learning tasks). Additionally, existing time-series models in education often omit a principled Bayesian framework, which is crucial for properly propagating uncertainty and facilitating partial pooling among students. Without partial pooling, one risks generating overly volatile estimates for students with fewer data points.
Across these threads of research—regression, machine learning, and time-series methods—several important gaps can be identified. First, many models [15,16] are not designed to exploit the hierarchical structure of educational data, in which students share certain population-level characteristics but also differ in idiosyncratic ways [17]. Approaches that ignore this hierarchy may miss out on useful information that can be borrowed from the broader student population. Second, there is often insufficient emphasis on the temporal dimension. Numerous risk models treat time as a binary partition—early in the semester vs. late—or they collapse week-by-week data into a single cumulative score, losing information about patterns or trajectories. Third, in efforts to make highly accurate predictions, some studies adopt models that are difficult to interpret or that lack well-calibrated measures of uncertainty, complicating the task of effectively communicating risk to instructors and students [18].
These limitations underscore the need for a dynamic, hierarchical, and interpretable modeling framework capable of handling longitudinal student data. It is here that Bayesian statistics offers a compelling approach. By explicitly incorporating prior distributions and hierarchical structures, a Bayesian model can partially pool information across students—especially valuable when sample sizes are moderate or when some students have missing or sparse observations—while allowing for student-specific deviations [19]. Likewise, through a state-space formulation, one can represent a student’s “risk level” as an evolving latent variable, updated incrementally as new assessments or engagement metrics become available. This dynamic approach is intuitively appealing from an educational standpoint, reflecting that students’ performance is not static but changes over time due to feedback, motivation, life events, and targeted interventions.
Despite these advantages, relatively few studies have developed Bayesian dynamic models for real-time at-risk student prediction. Existing work that does utilize Bayesian methods often concentrates on simpler random-effects structures (e.g., random intercepts) without explicitly modeling time evolution, or else it focuses on item-response-theory contexts that lack additional covariates such as demographic or engagement data. Another challenge is that Bayesian approaches can be more computationally intensive, leading some applied researchers to default to less flexible but faster procedures. However, the advent of sophisticated Hamiltonian Monte Carlo algorithms (as implemented in platforms such as Stan and PyMC) has made it increasingly feasible to fit hierarchical models to moderate-to-large educational datasets, providing rigorous posterior inference along with robust diagnostics and posterior predictive checks.
Accordingly, in this paper, we seek to address the above gaps by presenting a dynamic hierarchical logistic regression model within a fully Bayesian framework. Our approach fuses the strengths of partial pooling, which stabilizes estimates across students, with a state-space formulation that captures each student’s evolving risk trajectory. By specifying informative yet flexible priors, we ensure that the model remains interpretable, permitting educators and administrators to understand how different features (e.g., time on a learning management system, prior quiz scores) influence failure risk. Simultaneously, the hierarchical structure enables borrowing of strength among students, making the model particularly effective in contexts where data are missing or sparse for certain individuals. The dynamic component, meanwhile, provides a natural mechanism for producing updated risk estimates every time new performance data or engagement metrics become available.
This paper’s key contributions are threefold. First, this paper develops a methodological framework that explicitly merges hierarchical partial pooling with dynamic state evolution in the context of a Bernoulli outcome, thereby extending prior early-warning systems that have often employed static or non-hierarchical models. Second, the paper demonstrates how this model can be implemented efficiently using Markov chain Monte Carlo, and we present a thorough set of diagnostic tools to assess convergence, model fit, and calibration. Finally, the paper highlights how the resulting posterior distributions can be used for real-time intervention decisions—thus going beyond point predictions to provide a complete picture of uncertainty. Through these contributions, the paper aims to fill the methodological gap in educational analytics between purely static methods and unstructured machine-learning algorithms, offering an interpretable, flexible, and robust means of predicting academic risk over time.
The rest of this paper is arranged into several sections that collectively build and validate the proposed framework. Section 2 first presents the materials and methods, describing the data structure, model specification, and priors used. Section 3 then outlines a simulation study designed to validate our hierarchical logistic approach. Subsequently, Section 4 provides real-world application results, highlighting how the model performs on an authentic student dataset. In Section 5, the key findings are discussed in the context of the broader educational analytics literature, emphasizing both the methodological and practical implications. Finally, Section 6 draws conclusions and suggests potential avenues for future research and implementation.

2. Materials and Methods

This section presents a dynamic hierarchical logistic regression approach for predicting at-risk students over multiple time points, incorporating key modeling choices tailored to the nature of educational data.

2.1. Data Structure and Rationale for Modeling Choices

2.1.1. Real Data Description

Table 1 presents both programme-related and engagement-related variables used in this study. The first set of variables captures the student’s programme information, including the “QUAL CODE”, a nominal variable representing the specific qualification pursued, and the “CLASS GROUP”, another nominal variable categorizing students into classes (A–F). The “AS 1” variable is a numerical measure reflecting each student’s assessment mark. In addition to these programme variables, the table outlines several Moodle engagement indicators.
The “# OF COURSES” variable captures the total number of courses a student took in 2020, while “TIME ON SITE” quantifies the cumulative time spent on the Moodle platform. Engagement is further detailed through “TIME ON COURSES”, measuring the total time spent on course pages, and “TIME ON ACTIVITIES”, indicating time spent on specific learning activities. All of these time-based engagement variables are recorded as numerical values, allowing for an in-depth examination of how students interact with course materials.

2.1.2. Data Format and Binary Outcome

Consider N students, each observed at T discrete time points, indexed by ( i , t ) for i = 1 , , N and t = 1 , , T . We denote by y i , t { 0 , 1 } a binary indicator of failing (or at-risk) outcome. In our study, we have N = 517 and T = 4 , so 2068 total observations. The binary nature of pass/fail decisions naturally motivates a Bernoulli likelihood. Formally, modeling
y i , t Bernoulli ( π i , t )
provides an intuitive and standard approach, as it captures the probability π i , t that a given student i fails (i.e., y i , t = 1 ) at time t. This is common in educational analytics where outcomes such as “failing an assessment” or “flagged at risk” are inherently binary.

2.1.3. Justification for a Hierarchical Model

Educational data frequently exhibit heterogeneity across students and temporal dependence within a single student’s trajectory. By employing a hierarchical (multilevel) model, we can perform the following:
  • Partially pool information across students, thus smoothing estimates for individuals with sparse records.
  • Capture individual-level variation in baseline risk (through random intercepts or dynamic intercept evolution), reflecting that not all students share the same initial log-odds of failing.
  • Accommodate repeated measures per student in a principled way, respecting within-student correlations over time.
Such partial pooling is crucial in educational contexts, where some students may have incomplete data or otherwise outlying behaviors, yet still benefit from population-level regularization [17].

2.1.4. Covariates and Their Role

We observe a K-dimensional vector of predictors, x i , t R K , which may include the following:
  • Time-varying features: e.g., “time on site”, “time on courses”, “time on activities”, or “number of assignments completed”.
  • Time-invariant features: e.g., demographic factors, baseline academic indicators.
These covariates are hypothesized to modulate risk π i , t , allowing us to interpret which engagement or demographic factors correlate with failing.

2.2. Model Specification

2.2.1. Logistic Likelihood for a Binary Outcome

To link the Bernoulli parameter π i , t with a linear predictor, we employ a logistic (logit) link:
y i , t Bernoulli π i , t , π i , t = logit 1 α i , t + β x i , t .
This is justified by the following:
  • The binary nature of failing vs. passing;
  • The typical use of logistic regression in educational data mining for classification tasks;
  • Straightforward interpretation of β in terms of log-odds changes.

2.2.2. Dynamic (Time-Varying) Intercepts

Since student risk may evolve over T time points, we let each student i have an intercept α i , t that is dynamic. The two standard approaches are as follows:
(1)
Autoregressive (AR(1)) Model
We place a temporal correlation structure
α i , 1 = μ α + σ α z i , z i N ( 0 , 1 ) ,
α i , t = ϕ α i , t 1 + σ ϵ η i , t 1 , η i , t 1 N ( 0 , 1 ) , ϕ ( 1 , 1 ) , t = 2 , , T .
Here, ϕ is the persistence (or mean-reversion) coefficient, while σ ϵ scales student-specific log-odds increments. We adopt a non-centered parameterization via z i , η i , t 1 for stable MCMC sampling.
(2)
Random Intercept + Fixed Time Effects
When T is relatively small, one may prefer
logit π i , t = μ α + σ α z i + γ t + β x i , t ,
where γ t is a separate coefficient for each time period (e.g., t = 1 , , T ). This simpler approach can still capture average changes over time t, while partial pooling remains via the random intercept α i .
In either case, the hierarchical structure captures inter-student heterogeneity and temporal correlation, both essential features of educational data.

2.3. Prior Distributions and Their Rationale

The following priors align with the principle of weakly informative or moderately regularizing specifications [17,20]. They help stabilize estimation by discouraging extreme parameter values while still allowing the data to meaningfully adjust estimates.
  • Intercept Mean:
    μ α N ( 0 , 2 2 ) .
    The choice of this prior is justified by centering the baseline log-odds near 0, which corresponds to a baseline probability of 0.5 . The variance 2 2 is wide enough to accommodate substantial deviations if the data demand it, yet avoids defaulting to extreme risk probabilities in the absence of strong evidence.
  • Intercept Scales:
    σ α HalfCauchy ( 1 ) , σ ϵ HalfCauchy ( 1 ) .
    The choice of Half-Cauchy ( 1 ) priors for scale parameters allows capturing potentially large between-student variability ( σ α ) and temporal variation ( σ ϵ ) while reducing the risk of unbounded estimates. This regularization mitigates overfitting and contributes to stable sampling.
  • Autoregressive Coefficient (AR(1) models):
    ϕ N ( 0 , 0 . 5 2 ) truncated to ( 1 , 1 ) .
    The choice of a Normal ( 0 , 0 . 5 2 ) distribution, truncated to ( 1 , 1 ) , reflects the need for stationarity and avoids pathological extremes that can destabilize a logistic model. Centering on 0 indicates that perfect persistence ( ϕ near 1) or perfect reversal ( ϕ near −1) is considered unlikely without strong data support.
  • Regression Slopes (Covariate Effects):
    β k N ( 0 , 2 2 ) , k = 1 , , K .
    The choice of a Normal ( 0 , 2 2 ) prior for each slope parameter imposes minimal bias about the direction or magnitude of covariate effects. This distribution provides enough flexibility to allow significant positive or negative effects, yet discourages implausibly large estimates that could arise from limited data.
  • Time Effects (if using Random Intercept + Fixed Time Effects):
    γ t N ( 0 , 2 2 ) , t = 1 , , T .
    The choice of a Normal ( 0 , 2 2 ) prior on each γ t permits meaningful shifts in average risk from period to period, while remaining weakly informative enough to avoid extreme time-specific estimates. Including an optional sum-to-zero constraint, t = 1 T γ t = 0 , can facilitate interpretability by distinguishing overall intercepts from time effects.
An AR(1) framework is frequently adopted to encode temporal dependence in student-level risk. In many educational contexts, risk indicators (e.g., dropout probability or failing probability) exhibit inertia, meaning a student’s status at time t is positively correlated with that at time t 1 . The AR(1) model captures this persistence through the parameter ϕ , while σ ϵ governs random fluctuations around that persistent component. This specification strikes a balance between simplicity and realism, allowing student risk to evolve dynamically over time.
These priors reflect a compromise between flexibility and regularization. Normal priors on means and slopes center effects around zero without excluding moderate or large effects. Half-Cauchy priors on variance components keep standard deviations bounded yet adaptive. The truncated Normal prior for ϕ encourages stationary, moderate levels of persistence. Such specifications are particularly suitable for real-world educational data, where true effects can vary but seldom become extreme without substantial empirical support.

2.4. Bayesian Inference via MCMC

2.4.1. Joint Posterior and Implementation

Collect all unknowns into θ = μ α , σ α , ϕ , σ ϵ , β , { α i , t } . From the Bernoulli likelihood (1) and the AR(1) priors (2)–(3), the (unnormalized) joint posterior is as follows:
p θ D i = 1 N t = 1 T Bernoulli y i , t ; π i , t likelihood × i = 1 N N α i , 1 ; μ α , σ α 2 initial intercept × i = 1 N t = 2 T N α i , t ; ϕ α i , t 1 , σ ϵ 2 AR ( 1 ) evolution × p ( μ α , σ α , ϕ , σ ϵ , β ) priors ,
where π i , t = logit 1 α i , t + β x i , t . We implement (4) in Stan, typically using the No-U-Turn Sampler (NUTS) with iter = 2000 (1000 warmup) and chains = 4. We confirm convergence via the Gelman–Rubin statistic ( R ^ 1.00 ) and large effective sample sizes.

2.4.2. Interpretation of Key Parameters

  • μ α : Mean log-odds of failing (in absence of other effects). A large positive μ α indicates that, on average, students are at relatively high fail risk unless offset by negative increments or favorable covariate patterns.
  • σ α : Heterogeneity among students in their initial intercepts. Larger values mean that some students start far above/below the global mean in terms of fail risk.
  • ϕ : The autoregressive coefficient capturing how strongly a student’s risk at time t depends on time t 1 . A ϕ near 1.0 means risk persists strongly (once failing, they are likely to keep failing unless a large negative shock intervenes).
  • σ ϵ : Magnitude of random “shocks” or increments between time points. Large σ ϵ means students can experience dramatic shifts in log-odds from one interval to the next.
  • β k : The effect of covariate x k on the log-odds scale. Positive β k increases fail probability, negative decreases it.

2.5. Posterior Predictive Checks and Model Diagnostics

Once we have MCMC draws { θ ( s ) } s = 1 S , we simulate replicated outcomes
y ˜ i , t ( s ) Bernoulli logit 1 α i , t ( s ) + β ( s ) x i , t .
Comparisons of { y ˜ i , t ( s ) } with observed y i , t —including overall fail rates, distributions across time t, or the correlation structure—provide evidence for model adequacy. We also check calibration by grouping predicted probabilities into bins and comparing average predicted vs. observed failure frequencies.

3. Simulation Study

In order to validate the hierarchical logistic regression model with AR(1) dynamics proposed in Section 2, we conducted a simulation study under conditions mirroring the real data structure. Specifically, we generated synthetic observations for N = 1000 students across T = 4 time points, along with five covariates that align with the educational predictors introduced earlier (x_of_courses, time_on_site, time_on_courses, time_on_activities, as_1). Our goal was twofold: (1) to assess whether the AR(1) specification could adequately capture student-level variability and temporal persistence, and (2) to evaluate whether the model would recover the direction and approximate magnitude of slope parameters for each covariate.

3.1. Simulation Design

We first generated a latent intercept matrix α i , t of size ( N × T ) via an AR(1) process:
α i , 1 = μ α + σ α z i , α i , t = ϕ α i , t 1 + σ ϵ η i , t 1 ,
for i = 1 , , N and t = 2 , , T . The true hyperparameters were chosen to reflect moderate baseline risk ( μ α ), student-level heterogeneity ( σ α ), temporal correlation ( ϕ ), and moderate increments ( σ ϵ ). Five covariates were simulated to represent various engagement and assignment indicators, and each outcome y i , t { 0 , 1 } was drawn from a Bernoulli distribution whose probability was derived via a logistic link:
π i , t = logit 1 α i , t + β xc x _ of _ courses i , t + + β as 1 as _ 1 i , t .
We then fitted the exact same AR(1) hierarchical logistic model to this simulated dataset, employing four MCMC chains with iter = 4000 (of which 1000 were warmup) in parallel via Stan. After verifying convergence diagnostics such as the Gelman–Rubin statistic ( R ^ ) and effective sample sizes (n_eff), we extracted the posterior means and 95% credibility intervals for each parameter of interest.

3.2. Simulation Results

Table A1 presents the posterior summaries obtained from a representative run of this simulation study. Each parameter’s posterior mean, standard deviation (SD), lower and upper 2.5%/97.5% credible bounds, effective sample size, and R ^ statistic are reported. The results reveal that all parameters exhibited robust convergence ( R ^ 1.00 ) and substantial effective sample sizes, indicating that the MCMC exploration was thorough. In this simulation, we chose
μ α = 1.50 , σ α = 1.00 , ϕ = 0.75 , σ ϵ = 0.50 .
We then generated five covariates—x_of_courses, time_on_site, time_on_courses, time_on_activities, as_1—to mirror the real data structure. The slope parameters for these covariates were set as
β x c = 0.15 , β t o s = 0.01 , β t o c = 0.01 , β t o a = 0.00 , β a s 1 = 0.14 .
where,
  • β x c x_of_courses
  • β t o s time_on_site
  • β t o c time_on_courses
  • β t o a time_on_activities
  • β a s 1 as_1
The primary aim of this simulation was to validate the proposed AR(1) hierarchical logistic regression model for eventual application to real-world at-risk student data.

3.2.1. Accuracy of Covariate Slopes

The posterior mean for β x c is 0.17 , which closely matches the true value of 0.15 . This result reinforces that increases in x_of_courses elevate the log-odds of failing in this synthetic dataset. Likewise, β t o s 0.01 aligns with the mild negative effect originally specified at 0.01 . Although the estimate for β t o c at 0.03 slightly exceeds its true value of 0.01 , its positive direction remains consistent with the intended simulation.
As shown in Table 2  β t o a , the posterior estimate is 0.04 , whereas the true value was set to 0.00 . However, the 95% credibility interval includes zero, indicating that this small discrepancy likely arises from normal sampling variation. Finally, the estimate for β a s 1 at 0.04 is less negative than the prescribed 0.14 , yet it still reflects a modest protective effect for completing assignments (simulated here as as_1). In general, these results suggest that the AR(1) hierarchical logistic model successfully captures the direction of each effect and recovers most slope magnitudes with reasonable accuracy.

3.2.2. Implications for At-Risk Student Identification

In real educational contexts, these five covariates represent diverse student engagement behaviors (e.g., number of courses taken, time on site, assignment completions). The simulation confirms that our AR(1) hierarchical logistic approach successfully recovers the direction of their influence on fail probability, even when some parameters differ from their exact targets. Such partial deviations are expected with moderate sample sizes or short time series ( T = 4 ). Nonetheless, the overall trend—that more assignments completed has a negative coefficient, or that certain engagement metrics are only weakly related to fail odds—is preserved, thereby supporting the model’s relevance for at-risk student analytics.

3.2.3. Validation for Real-World Application

Beyond the slopes, the estimates for μ α 1.63 , σ α 1.31 , ϕ 0.67 , and σ ϵ 0.79 demonstrate that the model can capture moderate baseline fail risk, student-level heterogeneity, temporal persistence, and the possibility of notable shifts between assessments. These capacities are essential for at-risk student identification in real-world data, where individuals may fluctuate in performance from one semester (or time point) to another. The robust chain convergence ( R ^ = 1.00 and large n_eff across parameters) confirms that this model formulation is stable and well-suited to the data scenario.
Given the foregoing results of the simulation in those in Appendix A, this simulation provides a clear validation of the AR(1) hierarchical logistic regression framework for subsequent application to real student data. The close alignment between true and estimated slopes, along with the accurate recovery of key variance and temporal dependence terms, reinforces that the model is appropriate for identifying at-risk students across multiple time points. Future use of real data should benefit similarly from the model’s capacity to handle dynamic risk evolution, heterogeneous student baselines, and a range of engagement covariates.
From the simulation study results, we concluded the AR(1) hierarchical logistic model can stably recover a realistic pattern of student-level risk evolution, covariate influences, and baseline variability under relatively short time series ( T = 4 ). The high effective sample sizes and unity R ^ values signal strong chain mixing and robust convergence, underscoring the feasibility of applying this model to real educational data of similar dimensions.

4. Results of the Application Study

In this section, we provide the results of fitting our dynamic hierarchical logistic regression model to the real-world dataset comprising N = 517 students observed at T = 4 time points, along with a detailed discussion of these findings in the context of educational analytics. Table 3 presents the posterior means, standard deviations, and 95% credibility intervals for the principal parameters, including the hierarchical and AR(1) components, as well as covariate slopes.
As shown in Table 3 the hierarchical intercept mean ( μ α ) is estimated to be 5.41 (SD = 0.64) with a 95% credibility interval from 4.19 to 6.67, indicating that the baseline log-odds of failure is quite high in the absence of any mitigating factors. On the probability scale, a log-odds of approximately 5.41 translates to a failure probability near 0.995 if there are no negative offsets due to other parameters. This result suggests that, in the context of this dataset, a sizable proportion of students would be at extreme risk if they do not benefit from favorable covariate patterns or downward shifts in their individual trajectories. The standard deviation of student-specific initial intercepts ( σ α ) is 0.41 (SD = 0.31), with a 95% credibility interval ranging from 0.01 to 1.15. Although this range is broad, the mean value implies that there is moderate heterogeneity among students in their starting log-odds of failing. Some individuals may begin with lower risk, while others may lie closer to the high overall baseline.
The autoregressive coefficient ( ϕ ) is centered at 0.82 (SD = 0.09), with its 95% credibility interval from 0.60 to 0.97. A value well above 0.5 signals a strong persistence in risk from one time point to the next, supporting the notion that once a student’s fail probability is elevated, it remains high unless a substantial negative shock occurs. Meanwhile, the standard deviation of innovation increments ( σ ϵ ) is estimated to be 2.79 (SD = 0.41), and its 95% credibility interval stretches from 2.04 to 3.66. This sizeable value indicates that students may experience pronounced fluctuations (in log-odds terms) between successive assessments, possibly due to sudden changes in academic performance or external factors.
The slope β xc on x courses is approximately 0.15 (SD = 0.06), and its 95% credibility interval is 0.03 to 0.28, suggesting that when the number of courses a student engages with increases, there is a corresponding, albeit small, elevation in fail log-odds. By contrast, β tos 0.01 (SD = 0.01, 95% CI: −0.02 to 0.00) implies that additional time on site may slightly reduce fail risk, though the effect is near zero. Similarly, β toc 0.01 (SD = 0.01, 95% CI: −0.01 to 0.02) and β toa 0.00 (SD = 0.00, 95% CI: −0.01 to 0.01) remain near zero in the estimated log-odds space. Finally, β as 1 0.14 (SD = 0.01) with a 95% credibility interval of −0.17 to −0.12 indicates that having an additional successful assignment or analogous measure is related to a modest but noticeable reduction in the likelihood of failing. Concretely, this slope translates to a multiplicative factor of e 0.14 0.87 on the odds, thereby diminishing the risk of failure for students who consistently complete assignments.

Model Validation and Posterior Predictive Check

In order to validate the proposed model and ensure it adequately captures the underlying data-generating process, a comprehensive Posterior Predictive Check (PPC) was performed. After obtaining { θ ( s ) } s = 1 S from the MCMC sampler, replicated outcomes were generated for each posterior draw s as follows:
y ˜ i , t ( s ) Bernoulli logit 1 α i , t ( s ) + β ( s ) x i , t .
The result is a distribution of replicated fail rates that can be compared with the actual (observed) fail rate in the dataset.
Figure 1 shows a histogram of the replicated fail rates drawn from the posterior predictive distribution. The vertical red dashed line denotes the observed overall fail rate in the data. Because this line falls near the center of the replicated distribution, the model is capable of reproducing the average fail rate reasonably well. In other words, the model’s predictions for the average fail rate do not substantially deviate from empirical observations, suggesting an acceptable model fit for this aggregate measure.
These trace plots depict (Figure 2) the parameter values sampled by the Markov chain across iterations. The absence of strong trends or drifts, along with the relatively stable fluctuations around a mean, suggests that the chains have mixed well and likely converged. Convergence diagnostics (e.g., R ^ and effective sample size) confirm the adequacy of the sampling process, indicating that the model has been fit appropriately and supporting confidence in the reliability of subsequent inferences drawn from the posterior distribution.

5. Discussion of Findings

The findings in this study underscore the importance of examining students’ academic risk within a dynamic framework rather than relying solely on static, one-shot models. First, the relatively high baseline log-odds suggests that, on average, many students may be at risk of failing if no mitigation strategies are employed. However, the moderate student-level variation indicates that not all individuals begin with the same likelihood of difficulty, which aligns with recent work demonstrating diverse student profiles in the same educational context [21]. In addition, the substantial short-term volatility underscores the potential for rapid shifts in performance from one assessment interval to the next. Such short-term fluctuations have been documented in other contemporary educational analytics research, particularly those connecting sudden drops in engagement to extrinsic life events [22].
Furthermore, the high autoregressive coefficient highlights a persistence effect, whereby once a student’s risk becomes elevated, it remains so unless deliberate interventions are introduced. This finding resonates with recent studies demonstrating the need for timely alerts and meaningful feedback loops to break a negative trajectory [23]. By deploying early-warning systems that monitor risk fluctuations—both short-term and longer-term—educators can intervene at a critical juncture, potentially reversing the trajectory before failure becomes entrenched [24].
From a practical standpoint, the positive coefficient associated with higher course enrollment may point to scenarios in which students become overextended. This is consistent with more recent discussions in the literature that highlight over-commitment to multiple courses without adequate support structures [23]. Conversely, the negative parameter for assignment completion reaffirms that consistent engagement in key learning tasks remains one of the strongest protective factors against failing [25]. These effects, taken together, demonstrate how student behaviors and experiences fluctuate in ways that require more flexible, dynamic models capable of updating risk estimates over time.
Accordingly, this study’s approach contributes to the emerging shift toward dynamic hierarchical models in education, moving beyond static logistic regressions. By capturing short-term volatility, long-term persistence, and partial pooling across time, our proposed framework can help institutions tailor proactive interventions and better track the immediate impact of those interventions on student outcomes [21,22,23,24,26].

6. Conclusions

This study has presented a fully Bayesian, dynamic hierarchical logistic regression framework for modeling student failure risk over multiple assessments. By combining partial pooling (to stabilize estimates across students) with a state-space formulation (to capture temporal dynamics), the proposed approach offers a more nuanced understanding of how students’ risk profiles evolve than conventional, static logistic regression models. Results from both simulation experiments and a real-world application demonstrate that the AR(1) hierarchical structure effectively accounts for individual heterogeneity and persistent temporal dependence, providing accurate predictions even in relatively short time-series settings. In practice, these dynamic risk estimates can help educators identify at-risk students earlier and intervene more strategically by capitalizing on periods of heightened risk.
Several findings reinforce the value of adopting a dynamic modeling perspective. First, the high baseline log-odds of failure and strong autoregressive coefficient observed in the real dataset highlight the need to closely monitor students whose performance deteriorates over time. In such cases, static models may underestimate the momentum of risk and fail to detect students whose probability of failing escalates rapidly. Second, the partial pooling structure was crucial for students with sparse data records, as it allowed reliable inference of their risk trajectories by borrowing information from the broader population. Third, posterior predictive checks indicated good calibration and fit, suggesting that incorporating Bayesian priors to regularize parameters can reduce the risk of overfitting—a key concern in educational settings with smaller sample sizes or missing observations.

Funding

This research was funded by the ETDP-SETA Research in Mathematics Education at University of the Free State grant number UFS-AGR22-000053 (https://search.crossref.org/funding, accessed on 14 August 2024).

Institutional Review Board Statement

Ethics clearance number SREC: 0538/2024k on 14 August 2024.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are not available due to the terms and conditions of the ethical clearance approval.

Acknowledgments

The author would like to acknowledge C. Baartman for all the administrative support.

Conflicts of Interest

The author declares no conflicts of interest.

Appendix A. Simulation Study

In this appendix, we present a simulation study designed to evaluate how well the hierarchical logistic model with an AR ( 1 ) structure (as described in Section 2) recovers known parameter values. The objectives are twofold: (1) to illustrate how synthetic data can be generated under an AR ( 1 ) latent process for student risk, and (2) to demonstrate that the model accurately estimates both the fixed effects and the temporal dynamics.

Appendix A.1. Data Generation

We simulate data for N = 100 students, each observed at T = 4 discrete time points. Let β R K denote the slope parameters for K = 2 covariates. We define two predictors:
x 1 , i , t , x 2 , i , t ,
where x 1 , i , t might resemble “time on site” (a continuous variable) and x 2 , i , t might be a binary flag such as “assignment submission.” For simplicity, we generate these covariates as follows:
x 1 , i , t Uniform ( 0 , 5 ) , x 2 , i , t Bernoulli ( 0.5 ) .
The latent intercept α i , t follows an AR ( 1 ) process:
α i , 1 = μ α + σ α z i , z i N ( 0 , 1 ) , α i , t = ϕ α i , t 1 + σ ϵ η i , t 1 , η i , t 1 N ( 0 , 1 ) , t = 2 , , T .
We choose true values:
μ α = 1.50 , σ α = 1.00 , ϕ = 0.75 , σ ϵ = 0.50 .
These choices imply moderate heterogeneity in initial intercepts, medium persistence across time, and small but nontrivial increments from one interval to the next. We fix the slope parameters:
β 1 = 0.40 , β 2 = 0.60 .
Thus, each student’s log-odds evolves according to
α i , t + β 1 x 1 , i , t + β 2 x 2 , i , t .
We generate the fail indicator y i , t by
π i , t = logit 1 α i , t + β 1 x 1 , i , t + β 2 x 2 , i , t , y i , t Bernoulli π i , t .

Appendix A.2. Simulation Algorithm

We outline the algorithm in pseudocode:
  • Set seed for reproducibility (e.g., seed = 1234).
  • Define true parameters: ( μ α , σ α , ϕ , σ ϵ , β 1 , β 2 ) .
  • Initialize containers:
    • α i , t of size ( N × T ) ,
    • y i , t of size ( N × T ) ,
    • Covariates ( x 1 , i , t , x 2 , i , t ) of size ( N × T ) .
  • Generate covariates:
    x 1 , i , t Uniform ( 0 , 5 ) , x 2 , i , t Bernoulli ( 0.5 ) .
  • Generate initial intercepts:
    α i , 1 = μ α + σ α z i , z i N ( 0 , 1 ) .
  • For each time step t = 2 , , T :
    α i , t = ϕ α i , t 1 + σ ϵ η i , t 1 , η i , t 1 N ( 0 , 1 ) .
  • Compute probabilities:
    π i , t = logit 1 α i , t + β 1 x 1 , i , t + β 2 x 2 , i , t .
  • Draw outcomes y i , t Bernoulli ( π i , t ) .

Appendix A.3. Fitting the Model and Assessing Recovery

We then fit the same AR ( 1 ) hierarchical logistic model (Section 2) to the simulated data. In practice, one might write a Stan script that includes the following:
  • Parameter blocks for μ α , σ α , ϕ , σ ϵ , β 1 , β 2 , and the latent intercept matrix { α i , t } .
  • Priors such as μ α N ( 0 , 2 ) , σ α HalfCauchy ( 1 ) , ϕ N ( 0 , 0.5 ) truncated to ( 1 , 1 ) , σ ϵ HalfCauchy ( 1 ) , and β k N ( 0 , 2 2 ) .
  • Likelihood statements for y i , t BernoulliLogit ( α i , t + β 1 x 1 , i , t + β 2 x 2 , i , t ) .
  • AR(1) priors for α i , t to mirror (2)–(3).
After running MCMC (e.g., 2000 iterations, 4 chains, adapt_delta=0.95), one examines trace plots, effective sample sizes, and R ^ values. We then compare posterior means and credible intervals with the true parameter values:
μ α , σ α , ϕ , σ ϵ , β 1 , β 2 = ( 1.50 , 1.00 , 0.75 , 0.50 , 0.40 , 0.60 ) .

Appendix A.4. Simulation Results

A representative run of the AR(1) hierarchical logistic model, using the setup described in this simulation study, might yield results such as the following:
Table A1. Posterior summaries for the AR(1) hierarchical logistic model using four chains and 4000 iterations (with 1000 warmup). The means and standard deviations (SD) are reported along with 2.5% and 97.5% percentiles as a 95% credibility interval. The columns n_eff and R ^ provide measures of effective sample size and convergence, respectively.
Table A1. Posterior summaries for the AR(1) hierarchical logistic model using four chains and 4000 iterations (with 1000 warmup). The means and standard deviations (SD) are reported along with 2.5% and 97.5% percentiles as a 95% credibility interval. The columns n_eff and R ^ provide measures of effective sample size and convergence, respectively.
ParameterMeanSD2.5%50%97.5%n_eff R ^
μ α 1.630.201.241.622.0358031.00
σ α 1.310.260.871.291.9020071.00
ϕ 0.670.070.510.670.7925911.00
σ ϵ 0.790.210.390.791.227791.00
β xc 0.170.030.120.170.2429361.00
β tos −0.010.03−0.07−0.010.0597211.00
β toc 0.030.020.000.030.0661261.00
β toa −0.040.05−0.14−0.040.0685971.00
β as 1 −0.040.09−0.22−0.040.1411,5661.00
Here, the “±” terms approximate the standard deviations reported in the posterior summaries, and the full 95% credibility intervals typically contain the true parameter values. The Gelman–Rubin diagnostic ( R ^ ) remains at or very close to 1.00 for all parameters, indicating robust MCMC convergence across four parallel chains.
These estimates demonstrate that the AR(1) hierarchical logistic specification can capture both (a) the overall log-odds level (via μ α ), student variability ( σ α ), and temporal persistence ( ϕ ), and (b) the direction and approximate magnitude of the fixed-effects slopes ( β xc , β tos , ). Although small sample sizes or short time series can introduce uncertainty—especially for variance components like σ α and σ ϵ —the results suggest that, given enough students (N) and moderate time coverage (T), the proposed model recovers the salient features of the underlying process. This conclusion is further supported by the effective sample sizes (n_eff) in the hundreds or thousands, which affirm that the sampler has sufficiently explored the posterior distribution for each parameter.
Table A2. Posterior summaries for the AR(1) hierarchical logistic simulation with T = 6 time points. Each parameter’s posterior mean, standard deviation (SD), and 95% credibility interval (2.5–97.5%) are reported, along with the effective sample size (n_eff) and the Gelman–Rubin statistic ( R ^ ).
Table A2. Posterior summaries for the AR(1) hierarchical logistic simulation with T = 6 time points. Each parameter’s posterior mean, standard deviation (SD), and 95% credibility interval (2.5–97.5%) are reported, along with the effective sample size (n_eff) and the Gelman–Rubin statistic ( R ^ ).
ParameterMeanSD2.5%50%97.5%n_eff R ^
mu_alpha1.450.141.171.441.7365101.00
sigma_alpha1.200.180.861.191.5921451.00
phi0.720.050.610.720.8118711.00
sigma_epsilon0.530.130.300.530.799061.00
beta_xc0.150.020.110.150.1938491.00
beta_tos−0.020.02−0.07−0.020.0272701.00
beta_toc0.030.010.010.030.0556821.00
beta_toa0.040.04−0.040.040.1262691.00
beta_as1−0.120.07−0.26−0.130.0114,5911.00

Appendix A.5. Concluding Remarks on the Simulation

This simulation study illustrates how an AR ( 1 ) hierarchical logistic model can be validated under known conditions. By simulating both the covariates and the latent intercept process, we verify that the proposed model recovers sensible estimates of baseline risk, student-level variability, temporal persistence, and covariate slopes. Such simulations are invaluable for ensuring that real-data applications yield trustworthy inferences and highlight how short but well-structured time series ( T = 4 in this demonstration) can still inform dynamic logistic parameters, especially if the priors are sufficiently regularizing and the sample size ( N = 100 in this example) is moderate.

References

  1. Tinto, V. Dropouts from higher education: A theoretical synthesis of recent research. Rev. Educ. Res. 1975, 45, 89–125. [Google Scholar] [CrossRef]
  2. Crisp, G. The impact of structured peer review on the development of students’ academic writing skills. Assess. Eval. High. Educ. 2014, 39, 75–92. [Google Scholar]
  3. Queiroga, E.M.; Batista Machado, M.F.; Paragarino, V.R.; Primo, T.T.; Cechinel, C. Early Prediction of At-Risk Students in Secondary Education: A Countrywide K-12 Learning Analytics Initiative in Uruguay. Information 2022, 13, 401. [Google Scholar] [CrossRef]
  4. Baker, R.S. Data mining for education. In International Encyclopedia of Education; Desmarais, M.C., Baker, R.S.J., Eds.; Elsevier: Oxford, UK, 2009; Volume 7, pp. 112–118. [Google Scholar]
  5. Romero, C.; Ventura, S. Educational data mining: A review of the state of the art. IEEE Trans. Syst. Man, Cybern. Part C (Appl. Rev.) 2010, 40, 601–618. [Google Scholar] [CrossRef]
  6. Freedman, D.A. Statistical Models: Theory and Practice, 2nd ed.; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  7. Patel, O.U.; Haynes, W.K.; Burge, K.G.; Yadav, I.S.; Peterson, T.; Camino, A.; Van Wagoner, N.J. Results of a national survey of program directors’ perspectives on a pass/fail US Medical Licensing Examination Step 1. JAMA Netw. Open 2022, 5, e222192. [Google Scholar] [CrossRef] [PubMed]
  8. Thorat, M.; Sonawane, N.; Bhore, S.; Doiphode, S.; Sunkewar, V. Student performance prediction. Int. J. Res. Appl. Sci. Eng. Technol. 2023, 11, 1789–1792. [Google Scholar] [CrossRef]
  9. Babal, M.A.; Kauffman, J.M.; Haines, S.T. Student pharmacist perspectives on factors that influence wellbeing during pharmacy school. Am. J. Pharm. Educ. 2020, 84, 7831. [Google Scholar] [CrossRef]
  10. Hershcovits, H.; Vilenchik, D.; Gal, K. Modeling Engagement in Self-Directed Learning Systems Using Principal Component Analysis. IEEE Trans. Learn. Technol. 2020, 13, 164–171. [Google Scholar] [CrossRef]
  11. Currie, M.; Hammond, C.; Martinez, O.P.; Lane-Cordova, A.; Cook, J. The impact of United States Medical Licensing Examination Step 1 transitioning to pass/fail on medical student perception of research needed to match into one’s preferred specialty. Cureus 2024, 16, e57395. [Google Scholar] [CrossRef] [PubMed]
  12. Kay, J.J.; Kauffman, J.A.; Kauffman, J. Otolaryngology program director, house-staff, and student opinions: Step 1 pass/fail score reporting. Ann. Otol. Rhinol. Laryngol. 2022, 131, 453–459. [Google Scholar] [CrossRef]
  13. Zahran, F. Time-series analysis for academic performance prediction: A case study. In Proceedings of the International Conference on E-Learning, Virtual, 5–7 November 2021; pp. 210–217. [Google Scholar]
  14. Ali, M.; Pawluk, S.A.; Rainkie, D.C.; Wilby, K.J. Pass-fail decisions for borderline performers after a summative objective structured clinical examination. Am. J. Pharm. Educ. 2019, 83, 6849. [Google Scholar] [CrossRef] [PubMed]
  15. Ghazzal, M. Jordanian medical students’ perspective on the impact of a pass/fail USMLE Step 1: A cross-sectional investigation. Preprint 2023. [Google Scholar] [CrossRef]
  16. Tamakuwala, S.; Dean, J.; Kramer, K.J.; Shafi, A.; Ottum, S.; George, J.; Kaur, S.; Chao, C.R.; Recanati, M.A. Potential impact of pass/fail scores on USMLE Step 1: Predictors of excellence in obstetrics and gynecology residency training. J. Med. Educ. Curric. Dev. 2021, 8. [Google Scholar] [CrossRef] [PubMed]
  17. Gelman, A.; Hill, J. Data Analysis Using Regression and Multilevel/Hierarchical Models; Cambridge University Press: Cambridge, UK, 2007. [Google Scholar]
  18. Neville, A.L.; Smith, B.R.; de Virgilio, C. USMLE Step 1 scoring system change to pass/fail—An opportunity for change. JAMA Surg. 2020, 155, 1093–1094. [Google Scholar] [CrossRef] [PubMed]
  19. Zaffar, M.; Ali, A.; Memon, M.A. Role of FCBF feature selection in educational data mining. Mehran Univ. Res. J. Eng. Technol. 2020, 39, 772–778. [Google Scholar] [CrossRef]
  20. Gelman, A. Prior distributions for variance parameters in hierarchical models. Bayesian Anal. 2006, 3, 515–534. [Google Scholar]
  21. Chang, Y.; Liu, B. Dynamic Early-Warning Indicators in Online Learning: A Longitudinal Analysis. J. Learn. Anal. 2021, 8, 45–62. [Google Scholar]
  22. Martinez, R.; EdTech Research Group. Life Events, Dropout Risk, and Real-time Alerts in a Digital Learning Platform. Educ. Data Sci. Rev. 2022, 3, 10–29. [Google Scholar]
  23. Alvarez, D.; Zhao, L. Overextension or Overachievement? Examining Course Load and Academic Outcomes in Higher Education. Contemp. Educ. Technol. 2022, 14, 66–82. [Google Scholar]
  24. Zhao, P.; Chen, G. Longitudinal Analysis of Engagement Patterns and Intervention Impacts in Blended Learning. Comput. Educ. 2023, 18, 98–117. [Google Scholar]
  25. Liu, J.; Xu, K. Assignment Completion as a Predictor of Academic Resilience: Evidence from a Hybrid Classroom Environment. Int. J. Educ. Technol. 2023, 9, 114–130. [Google Scholar]
  26. Nimy, E.; Mosia, M.; Chibaya, C. Identifying At-Risk Students for Early Intervention—A Probabilistic Machine Learning Approach. Appl. Sci. 2023, 13, 3869. [Google Scholar] [CrossRef]
Figure 1. Posterior predictive check (PPC) plot showing the observed vs. predicted fail risk across the range of covariates. The vertical red dashed line indicates the actual overall fail rate.
Figure 1. Posterior predictive check (PPC) plot showing the observed vs. predicted fail risk across the range of covariates. The vertical red dashed line indicates the actual overall fail rate.
Data 10 00023 g001
Figure 2. Trace plots for key parameters in the hierarchical model. Each sub-figure shows the MCMC draws (vertical axis) over iterations (horizontal axis) for one parameter, indicating satisfactory mixing and convergence.
Figure 2. Trace plots for key parameters in the hierarchical model. Each sub-figure shows the MCMC draws (vertical axis) over iterations (horizontal axis) for one parameter, indicating satisfactory mixing and convergence.
Data 10 00023 g002
Table 1. Student data description (N = 517 students, T = 4 discrete time points, which is 2068 total observations).
Table 1. Student data description (N = 517 students, T = 4 discrete time points, which is 2068 total observations).
VariablesAttributeDescription
Programme
QUAL CODENominalStudent qualification code
CLASS GROUPNominalStudent’s class group
AS 1NumericalAssessment mark
Moodle Data
# OF COURSESNumericalNumber of courses taken by the student
TIME ON SITENumericalTime student spent on Moodle (cumulative)
TIME ON COURSESNumericalTime student spent on the course (cumulative)
TIME ON ACTIVITIESNumericalTime student spent on course activities (cumulative)
Table 2. Posterior summaries for a single run of the AR(1) hierarchical logistic simulation. The mean, standard deviation (SD), and 95% credibility interval are reported. The columns n_eff and R ^ provide measures of effective sample size and convergence, respectively.
Table 2. Posterior summaries for a single run of the AR(1) hierarchical logistic simulation. The mean, standard deviation (SD), and 95% credibility interval are reported. The columns n_eff and R ^ provide measures of effective sample size and convergence, respectively.
ParameterMeanSD2.5%50%97.5%n_eff R ^
μ α 1.630.201.241.622.0358031.00
σ α 1.310.260.871.291.9020071.00
ϕ 0.670.070.510.670.7925911.00
σ ϵ 0.790.210.390.791.227791.00
β xc 0.170.030.120.170.2429361.00
β tos −0.010.03−0.07−0.010.0597211.00
β toc 0.030.020.000.030.0661261.00
β toa −0.040.05−0.14−0.040.0685971.00
β as 1 −0.040.09−0.22−0.040.14115661.00
Table 3. Posterior estimates for the hierarchical AR(1) logistic model. Each parameter is summarized by its mean, standard deviation (SD), and the endpoints of its 95% credibility interval (2.5% and 97.5%). We also report the effective sample size (n_eff) and the Gelman–Rubin statistic ( R ^ ) as measures of convergence and mixing.
Table 3. Posterior estimates for the hierarchical AR(1) logistic model. Each parameter is summarized by its mean, standard deviation (SD), and the endpoints of its 95% credibility interval (2.5% and 97.5%). We also report the effective sample size (n_eff) and the Gelman–Rubin statistic ( R ^ ) as measures of convergence and mixing.
ParameterMeanSD2.5%97.5%n_eff R ^
μ α 5.410.644.196.6723451.00
σ α 0.410.310.011.158311.01
ϕ 0.820.090.600.9710501.00
σ ϵ 2.790.412.043.666641.01
β xc 0.150.060.030.2812541.00
β tos −0.010.01−0.020.0019341.00
β toc 0.010.01−0.010.0218861.00
β toa 0.000.00−0.010.0127361.00
β as 1 −0.140.01−0.17−0.129801.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mosia, M. A Bayesian State-Space Approach to Dynamic Hierarchical Logistic Regression for Evolving Student Risk in Educational Analytics. Data 2025, 10, 23. https://doi.org/10.3390/data10020023

AMA Style

Mosia M. A Bayesian State-Space Approach to Dynamic Hierarchical Logistic Regression for Evolving Student Risk in Educational Analytics. Data. 2025; 10(2):23. https://doi.org/10.3390/data10020023

Chicago/Turabian Style

Mosia, Moeketsi. 2025. "A Bayesian State-Space Approach to Dynamic Hierarchical Logistic Regression for Evolving Student Risk in Educational Analytics" Data 10, no. 2: 23. https://doi.org/10.3390/data10020023

APA Style

Mosia, M. (2025). A Bayesian State-Space Approach to Dynamic Hierarchical Logistic Regression for Evolving Student Risk in Educational Analytics. Data, 10(2), 23. https://doi.org/10.3390/data10020023

Article Metrics

Back to TopTop