Next Article in Journal
Estimator’s Properties of Specific Time-Dependent Multivariate Time Series
Next Article in Special Issue
Spatial Modeling of Auto Insurance Loss Metrics to Uncover Impact of COVID-19 Pandemic
Previous Article in Journal
Development of the Methodology and Mathematical Tools for Strategic Decision-Making Process Modeling of the Digital Business Ecosystem
Previous Article in Special Issue
Interval Estimation for the Two-Parameter Exponential Distribution Based on Upper Record Value Data Using Bayesian Approaches
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Bayesian Nonparametric Regression via Stationary Smoothness Priors

Economics Department, Purdue University, West Lafayette, IN 47907, USA
Mathematics 2025, 13(7), 1162; https://doi.org/10.3390/math13071162
Submission received: 26 February 2025 / Revised: 25 March 2025 / Accepted: 27 March 2025 / Published: 31 March 2025
(This article belongs to the Special Issue Bayesian Statistics and Causal Inference)

Abstract

:
A procedure for Bayesian nonparametric regression is described that automatically adjusts the degree of smoothing as the curvature of the underlying function changes. Relative to previous work adopting a similar approach that either employs a single global smoothing parameter or assumes that the smoothing process follows a random walk, the model considered here permits adaptive smoothing and imposes stationarity in the autoregressive smoothing process. An efficient Markov Chain Monte Carlo (MCMC) scheme for model estimation is fully described for this stationary case, and the performance of the method is illustrated in several generated data experiments. An application is also provided, analyzing the relationship between behavioral problems in students and academic achievement. Point estimates from the nonparametric methods suggest (a) expected achievement declines monotonically with a behavioral problems index (BPI) score and (b) the rate of decline is relatively flat at the left tail of the BPI distribution and then becomes sharply more negative.
MSC:
62G08

1. Introduction

The literature on nonparametric Bayesian estimation of regression relationships is substantial, as a variety of approaches have been advanced to successfully recover shapes and properties of unknown functions in the presence of observational data. One approach to Bayesian nonparametric regression, for example, employs basis functions or splines to flexibly model regression relationships (e.g., Smith and Kohn [1], Kohn et al. [2], and Chib et al. [3]). These approaches typically require the researcher to choose a functional form for the basis and the potential locations of a series of knots, and as estimation proceeds, the data are used to determine if terms in the series representation are retained or discarded. Other studies employ Gaussian processes (GPs) for this purpose (e.g., Rasmussen [4] and Dixon et al. [5] (in particular, Chapter 3 of Dixon et al. [5] provides considerable treatment of the GP approach to Bayesian nonparametric regression)). GP models, which require the researcher to choose a mean function and covariance kernel, place a prior distribution over the function space, and plausible functional realizations from the associated stochastic process are identified given the data that are observed. Still another set of studies estimate relationships nonparametrically by treating each point on the regression surface as a parameter to estimate, and then by representing the nonparametric relationship in traditional regression form. Carefully constructed priors are employed in this setting that impose similarity among neighboring points on the regression surface, thereby smoothing out what might otherwise be an erratic collection of regression coefficients. This paper follows in the path of this “smoothness prior” approach and extends it by allowing for local adaptability in the degree of smoothing, by imposing stationarity in the smoothing parameter process, and by allowing the data to update some features of how the smoothing parameters evolve.
To generally describe the approach taken, the case where our covariate, say x, is discrete-valued serves as a useful explanatory starting point. When seeking to estimate a regression function f ( x ) with such data, one would likely begin by creating a series of dummy/indicator variables which exhaust all possible x outcomes. The resulting collection of estimated dummy variable coefficients when proceeding this way, however, may produce a rather erratic estimate of the regression function, particularly when there are few observations within some of the discrete cells and/or the error variance is large. To help remedy this problem, one could adopt an informative prior expressing that adjacent dummy variable coefficients should be similar in value, essentially borrowing informational strength across the discrete cells. Expressions of such local “similarity” can be achieved, for example, by constructing a prior that centers differences of the dummy variable coefficients over zero. The extent of shrinkage/smoothing that takes place is then determined by how tightly those differences are centered over zero.
Studies such as Koop and Poirier [6], Koop and Tobias [7], Chib and Jeliazkov [8], and Chan and Tobias [9], for example, employ this smoothness prior approach. Their specifications contain a single variance parameter in the construction of the smoothness prior, which is akin to a constant global bandwidth parameter in frequentist kernel-based nonparametric regression. While these methods generally perform quite well, they are nonetheless somewhat limited by the unique smoothing parameter: the same degree of smoothing is employed over regions of the support where the function is flat or linear, and over regions where the function may be rapidly changing. Tobias and Bond [10] extend this single-parameter model, allow for adaptability in the degree of smoothing, and structure their prior so that the smoothing parameters follow a random walk. Although the adaptability afforded by such an algorithm may offer real benefit in empirical work, the assumed random walk process is nonetheless somewhat limiting as it imposes, for example, that prior variances grow as one moves across the support of x.
In this paper, we describe an alternative to the random walk model which may offer improved performance in practice. We allow the data to partially inform the way smoothness evolves by assuming the smoothing parameters follow an AR(1) process rather than a random walk. A truncated prior for the autoregressive coefficient is employed to impose stationarity of the smoothing parameter process. This elaborated structure introduces several new parameters to the model (e.g., a stationary mean and an autoregressive coefficient), and an efficient Markov Chain Monte Carlo (MCMC) scheme is derived and employed for estimating this generalized model. We show in several generated experiments that the corresponding AR(1) smoothing process can adapt more quickly than the random walk to functional curvature changes when they occur. Moreover, our experience suggests that the AR(1) structure tends to be computationally more stable and perhaps less sensitive to changes in prior hyperparameters. Offsetting this, the random walk model contains fewer parameters than the AR(1), as it imposes that the autoregressive coefficient equals unity, which both simplifies the resulting estimation algorithm and could potentially lead to increased precision. In some of the conducted experiments, support for the random walk specification is revealed, and in those cases the mass of the posterior distribution of the autoregressive coefficient is seen to concentrate near the unit root. In other cases, we find evidence against the random walk, and in those cases, the posterior distribution of the autoregressive coefficient concentrates in the interior of the support and away from the unit root.
The outline of this paper is as follows. The next section outlines our model and describes the prior specifications used to conduct nonparametric estimation and smoothing. Section 3 fully describes an MCMC scheme for model estimation, while generated data experiments are conducted in Section 4. Section 5 provides an application investigating the impact of behavioral problems in adolescents on achievement scores in mathematics and two measures of reading acumen. In this application, we find some evidence of nonlinearities, which are then further supported by simpler parametric models whose specifications are guided by our initial exploratory nonparametric findings. Specifically, additional behavioral problems are found to have a relatively small impact on achievement when the extent of behavioral issues is currently small, while increasing the number of behavioral problems has clearly negative impacts on achievement when the number of behavioral issues is already moderate or large. Finally, this paper concludes with a summary in Section 6.

2. The Model

Before describing our model, we establish a few notational conventions. Univariate random variables (i.e., x) are denoted with lowercase letters or symbols, while vector-valued quantities (i.e., x ) are presented in bold lowercase font. Matrices (i.e., X ) are denoted with bold capital letters or Greek symbols, as appropriate.
In this paper, we will introduce a new framework for estimating the conditional mean function in a univariate nonparametric regression model. Before describing our approach, we note that many extensions of this base model are easily handled. For example, the posterior simulator can be easily adapted to include additional controls in a partially linear framework: y i = f ( x i ) + z i β + u i . Non-normality of the disturbance vector can also be accommodated through scale or finite mixtures (see, e.g., Chan et al. [11], chapter 15, and associated references). The strategy for nonparametric estimation described below can also be included as a component of a larger treatment response model when seeking to identify the causal effect of a covariate or set of covariates on an outcome of interest (e.g., Kline and Tobias [12]). We fix ideas here, however, on the basic univariate nonparametric regression model in order to focus attention on the contributions of our estimation method:
y i = f ( x i ) + ϵ i , i = 1 , 2 , , n , ϵ | X N ( 0 , σ 2 I n )
where N ( μ , Σ ) denotes a normal/Gaussian distribution with mean μ and covariance matrix Σ . The covariate x takes on k x n distinct values. We let x 1 * , x 2 * , , x k x * denote the realized distinct ordered values of the random variable x so that x j * > x j 1 * j = 2 , 3 , k x . Furthermore, we let Δ j denote the distance between consecutive x * values, i.e., Δ j x j * x j 1 * .
Our approach treats each f ( x j * ) as a parameter to estimate. Smoothness will be achieved by specifying a prior expressing local similarity among neighboring f ( x * ) values, and the function is recovered by piecing together the collection of smoothed coefficients. We begin this process by constructing a ( 1 × k x ) label vector d i , for each observation i, as follows:
d i = [ 1 ( x i = x 1 * ) 1 ( x i = x 2 * ) 1 ( x i = x k x * ) ] ,
where 1 ( · ) denotes a standard indicator function. The vector d i contains zeros everywhere except for a one in the position that corresponds to the observed value of x i . We let D denote the stacked n × k x label matrix for the full sample and π denote the associated ( k x × 1 ) vector of (sorted) dummy variable coefficients:
D = d 1 d 2 d n , π = π 1 π 2 π k x .
With this construction, we can write our nonparametric regression problem in standard and familiar regression form as
y = D π + ϵ .
In the following section, we describe how this method is operationalized through the construction of smoothness priors on π . The priors employed will also be adaptive, automatically adjusting the degree of smoothing as the curvature of the function f changes.

Priors and Regression Function Smoothing

We smooth the collection of π coefficients via a prior which imposes that neighboring pointwise slopes of the regression function should be similar in value. Such priors have been employed previously by, for example, Koop and Poirier [6], Koop and Tobias [7], Chib and Jeliazkov [8], Chib et al. [13], and Bond and Tobias [10], among others. The prior employed here will be different, however, as it does not employ a single global smoothing parameter. Instead, it associates every cell with its own smoothing parameter to permit local adaptability, while imposing stationarity in the smoothing parameter distribution.
We begin by defining the ( k x × k x ) second-order differencing matrix:
H π = 1 0 0 0 0 0 0 0 1 0 0 0 0 0 Δ 2 1 ( Δ 3 1 + Δ 2 1 ) Δ 3 1 0 0 0 0 0 Δ 3 1 ( Δ 4 1 + Δ 3 1 ) Δ 4 1 0 0 0 0 0 0 0 Δ k x 1 1 ( Δ k x 1 + Δ k x 1 1 ) Δ k x 1 .
The matrix H π is constructed so that the quantity H π π will select off the first two elements of the dummy variable coefficient vector ( π 1 and π 2 ) and then select the first differences of adjacent pointwise slopes of our regression function. These slope-first differences will take the form Δ j 1 ( π j π j 1 ) Δ j 1 1 ( π j 1 π j 2 ) , j = 3 , 4 , , k x .
We achieve smoothing by imposing a belief that adjacent slopes should be similar in value. We accomplish this through the prior
H π π | π 0 , η N X π π 0 , V η ,
where
X π = I 2 0 ( k x 2 ) × 2 , π 0 = π 01 π 02 , and V η = exp ( η 1 ) 0 0 0 0 exp ( η 2 ) 0 0 0 0 0 exp ( η k x ) .
The η parameters are log variances/log volatilities, which govern the degree of smoothing of the regression function and summarize how that smoothing might change across the support of x. To explain the role and interpretation of the η parameters in greater detail, first note that the prior in (1) centers first differences of slopes of our regression function over zero. As exp ( η i ) 0 , j > 2 , the prior imposes that adjacent slope differences are essentially equal to zero, thus constraining the regression function to be linear. More generally, however, the exp ( η j ) can (and typically will) depart from zero. In such a case, the prior expresses preference for local similarity in neighboring function values (thereby achieving “smooth” results) yet does not strictly impose linearity. In regions where the true function behaves according to a linear or near-linear relationship, we would hope to associate those f ( x * ) with small η values. In contrast, when the function is changing rapidly, a larger η would base results on relatively small local neighborhoods, thus allowing for improved detection of local changes in functional curvature. The model as presented offers the potential for adaptability and the automatic determination of the appropriate neighborhood size/degree of smoothing, as η can change with the local properties of the underlying function. We provide some examples of the algorithm’s success in adapting in this way in the experiments of Section 4. It is worth noting that many studies pursuing nonparametric estimation through priors like these employ a constant smoothing parameter, specifying V η to be diagonal with a common η 2 across all diagonal elements. Our model nests this as a special case and associates each distinct x j * with its own smoothing parameter. To operationalize the approach, some additional structure is required which describes how exp ( η ) evolves. We will return to complete this description shortly.
An additional hierarchical prior is placed over π 0 of the form
π 0 N μ ̲ π 0 , V ̲ π 0 ,
where the underline notation is used to denote terminal hyperparameters selected by the researcher. In specifying our model this way, the prior covariance matrix V ̲ π 0 can be selected with large diagonal elements to let the data speak regarding the first two points on our regression curve. Specifically, if small values of η are generally desired in order to smooth the curve, while the elements of η are also correlated in their evolution, the resulting prior for the initial conditions may be undesirably concentrated around the prior mean μ ̲ π 0 . The hierarchical specification in (1) and (3) combined with diffuse choices for V ̲ π 0 can help mitigate this problem and allow the data to better inform π 1 and π 2 .
As discussed earlier in this section, the variances { exp ( η j ) } control the degree of smoothing of our regression function. It seems reasonable to assume that the η j parameters, like those of the regression function itself, should be locally similar. While Tobias and Bond [10] express this similarity by assuming η smoothly evolves as a random walk, this (potentially undesirably) implies that the prior variance of η must increase as one moves into the right tail of the x distribution (Tobias and Bond [10] mention that a stationary version of the process produced similar results, but did not present an algorithm for such estimation. We formally describe such an algorithm in Section 3). When adopting such a specification, the imposed unit root may also imply that η adapts to curvature changes in f ( x ) rather slowly, while alternate specifications may permit more rapid adaptability. We provide examples that both favor and oppose the random walk structure in the generated data experiments of the following section.
In what follows, we assume that the elements of η follow an autoregressive or AR(1) process:
η 2 = μ + ρ ( η 1 μ ) + u 2 η 3 = μ + ρ ( η 2 μ ) + u 3 η k x = μ + ρ ( η ( k x 1 ) μ ) + u k x
where u j i i d N ( 0 , τ 2 ) , j = 2 , 3 , , k x . The parameter ρ measures the degree of persistence in the series, and the process is stationary provided | ρ | < 1 . We will employ priors that guarantee stationarity by truncating the support of ρ accordingly, although results from the nonstationary random walk model may be approximately reproduced when the posterior for ρ concentrates on values near, but strictly less than, unity. The smoothing process is initialized by assuming that η 1 is drawn from the stationary distribution of the process, i.e., η 1 N ( μ , τ 2 / ( 1 ρ 2 ) ) . We refer to this model in what follows as the “stationary” case or the “estimated- ρ ” case. This is in contrast to the random walk model, which imposes that ρ = 1 (and thus eliminates μ ) and adds a prior for the initial condition of the form η 1 N ( μ ̲ , V ̲ ) . Other higher-order autoregressive processes for the smoothing parameters could also be entertained, although we restrict attention to the baseline AR(1) model here (in the experiments and application of the following sections, we show that, while learning takes place regarding the persistence parameter ρ , it is not strongly identified by the data. This suggests that further generalizations of the smoothing parameter process may prove quite difficult to implement in practice).
To further describe the implied joint prior for η , we first define a few objects. We define the ( k x × k x ) matrix G ρ , ( k x × 1 ) vector g ρ , and diagonal ( k x × k x ) covariance matrix Ω = Ω ( τ 2 , ρ ) as follows:
G ρ = 1 0 0 0 0 ρ 1 0 0 0 0 ρ 1 0 0 0 0 0 ρ 1 , g ρ = ( 1 ρ 2 ) ( 1 ρ ) ( 1 ρ ) ( 1 ρ ) , Ω = τ 2 ( 1 ρ 2 ) 0 0 0 0 τ 2 0 0 0 0 τ 2 0 0 0 0 τ 2 .
With these definitions in hand, we can express the joint prior for η as
G ρ ( η ι k x μ ) | ρ , τ 2 N 0 , Ω ,
where ι z denotes a z × 1 vector of ones. The model specification is complete upon choosing priors for μ , ρ , τ 2 , and σ 2 . In what follows, we make the following choices:
μ N ( μ ̲ , V ̲ μ ) , τ 2 I G ( a ̲ τ , b ̲ τ ) , and σ 2 I G ( a ̲ σ , b ̲ σ ) ,
with I G ( · , · ) denoting an inverse gamma distribution (see, e.g., Chan et al. [11], page 442). In addition, we consider two different prior specifications for ρ . The first is a discrete prior with J points of support (denoted { ρ ̲ j } j = 1 J , | ρ ̲ j | < 1 j ):
P r ( ρ = ρ ̲ j ) = p ̲ j , ρ { ρ ̲ 1 , ρ ̲ 2 , , ρ ̲ J } , j = 1 J p ̲ j = 1 , p ̲ j ( 0 , 1 ) j .
In addition, we also consider a continuous prior of the form
ρ T N ( 1 , 1 ) ( μ ̲ ρ , V ̲ ρ ) ,
where T N ( a , b ) ( μ , σ 2 ) denotes a normal distribution with mean μ and variance σ 2 that is truncated to the interval ( a , b ) (see, e.g., Chan et al. [11] pages 132–134). We observe that the truncation in the prior for ρ imposes that the smoothing process is stationary.

3. Posterior Simulation

In this section, we seek to derive a Markov Chain Monte Carlo (MCMC) scheme to estimate our nonparametric regression model. To this end, we begin by noting that the joint posterior distribution of our model parameters is given up to proportionality as
p ( π , π 0 , η , ρ , σ 2 , τ 2 , μ | y ) ϕ y ; D π , σ 2 I n p ( ρ ) p ( σ 2 ) p ( τ 2 ) p ( μ ) p ( π 0 ) p ( η | μ , ρ , τ 2 ) p ( π | π 0 , η ) ,
with p ( · ) serving as generic notation for a density or mass function. In (7), ϕ ( x ; μ , Σ ) denotes a multivariate normal density function for the random vector x with mean vector μ and covariance matrix Σ . The specific forms of the priors appearing in (7) were given in the previous section, and p ( ρ ) may represent either the continuous or discrete variant of the prior for the autoregressive coefficient ρ .
Below we fully describe all steps of our Markov Chain Monte Carlo sampling scheme, which involves iterative simulation from the conditional posterior distributions of the model parameters. With just one exception (the sampling of ρ in the continuous case), each step only requires simulation from standard distributions. When required, a well-mixing Metropolis–Hastings (MH) algorithm is suggested for the sampling of ρ , and thus the method should be generally accessible by practitioners.
The approach that follows can be interpreted as an approximate (though highly accurate) MCMC procedure. The approximate nature of our results owes to a Gaussian mixture approximation that will be used to facilitate drawing the smoothing parameters/log volatilities η . Specifically, we use a 10-component Gaussian mixture approximation of the log of a χ 2 ( 1 ) random variable, as in Omori et al. [14], to aid in the sampling of η . This will require the introduction of a vector of component indicators, denoted s , and the sampling of these indicators will conclude one sweep of the sampler. As discussed in Del Negro and Primiceri [15] in the context of stochastic volatility models, the ordering of the sampling steps is important in this context.
To describe the ordering issue formally, we will conceptually group a subset of parameters together as follows: γ { π , π 0 , σ 2 , μ , ρ , τ 2 } . We will also make use of the notation γ x to denote all elements of γ other than x. Each element of γ will be sampled from its exact conditional posterior distribution, as derived from (7). Our sampling scheme will draw, in order, from [ η | s , γ , y ] , [ γ | η , y ] , and finally [ s | η , γ , y ] . Sampling from the last two of these amounts to drawing from the joint conditional posterior distribution [ s , γ | η , y ] given that elements of γ will be drawn from their exact conditional posteriors, which do not involve any approximations (and thus s ). Importantly, the sampling of γ in this scheme needs to take place immediately before s . Before describing each of the required eight steps, we note the structural similarity of the nonparametric regression problem considered here with traditional stochastic volatility models. The reader is invited to see, for example, Chan [16] for related discussions. As described in Kim, Shephard, and Chib [17], one can make adjustments for the mixture approximation error via, for example, importance sampling reweighting of the produced simulations. This is seldom performed in practice, however, as the approximation is highly accurate and such adjustments tend to produce no meaningful differences in results.
  • Step 1: η | s , γ , y :
In equation form, the prior in (1) can be expressed as
ψ = X π π 0 + ν ,
where ψ H π π and ν is multivariate Gaussian with mean zero and diagonal covariance matrix V η containing { exp ( η i ) } i = 1 k x stacked across the diagonal. Rearranging this equation by squaring both sides and taking the natural logarithm, we can express this prior as
log ( ψ X π π 0 ) 2 = η + ζ ,
where each element of the ( k x × 1 ) vector ζ are i i d draws from a log χ 2 ( 1 ) distribution and ⊙ denotes the Hadamard or element-wise product. This representation is beneficial, as η enters the right-hand side as essentially a vector of intercept parameters, thus opening the door to standard sampling possibilities. However, ζ is not Gaussian, which introduces an unfortunate computational complication.
To address this complication, we replace the log χ 2 ( 1 ) with a highly accurate Gaussian mixture approximation. Specifically, we use a ten-component mixture approximation, as described in Omori et al. [14]:
ζ i s = 1 10 p s N ( m s , ω s 2 ) .
The component probabilities p s , means m s , and variances ω s 2 are fixed (rather than estimated) and appear in Table 1 of Omori et al. [14]. A mixture approximation-based analysis thus proceeds by introducing a set of component indicators s = { s i } i = 1 k x such that
ζ ˜ i | s i N ( m s i , ω s i 2 ) , Pr ( s i = j ) = p j , j = 1 , 2 , , 10 .
Conditioned on the component indicators s , we have
log ( [ ψ X π π 0 ] 2 ) = m s + η + ζ ˜ , ζ ˜ | s N ( 0 , Γ s ) ,
where Γ s diag { ω s i 2 } .
The prior for η in (4) can be expressed as
η N ι k x μ , G ρ 1 Ω ( G ρ 1 ) .
Combining (9) and (10), one can show
η | γ , s , y N D η d η , D η ,
where
D η = Γ s 1 + G ρ Ω 1 G ρ 1 , d η = Γ s 1 log ( ψ X π π 0 2 ) m s + G ρ Ω 1 G ρ ι k x μ .
  • Step 2: π | γ π , η , y
When conducting this step, we choose to sample ψ = H π π . Once ψ is sampled, the desired smoothed regression coefficients can be calculated at each iteration by simply inverting the mapping, i.e., calculating π = H π 1 ψ .
Our regression equation can be reparameterized in terms of ψ by noting
y = D θ + ϵ = D H π 1 H π π + ϵ = D ˜ ψ + ϵ ,
where D ˜ D H π 1 .
It follows that the conditional posterior distribution of ψ is Gaussian:
ψ | γ ψ , η , y N ( D ψ d ψ , D ψ ) ,
where
D ψ D ˜ D ˜ / σ 2 + V η 1 1 , d ψ D ˜ y / σ 2 + V η 1 X π π 0 .
  • Step 3: π 0 | γ π 0 , η , y
From (7), the conditional posterior distribution of the initial conditions π 0 is proportional to the product of the prior H π π N ( X π π 0 , V η ) and the prior for π 0 . This implies
π 0 | γ π 0 , η , y N D π 0 d π 0 , D π 0 ,
where
D π 0 = X π V η 1 X π + V ̲ π 0 1 1 , d π 0 = X π V η 1 ψ + V ̲ π 0 1 μ ̲ π 0 .
  • Step 4: σ 2 | γ σ 2 , η , y
From (7), it is evident that the conditional posterior distribution of the error variance σ 2 follows an inverse gamma distribution:
σ 2 | γ σ 2 , η , y I G a ̲ σ + n 2 , b ̲ σ 1 + 1 2 ( y D π ) ( y D π ) 1 .
  • Step 5: μ | γ μ , η , y
We begin by writing the prior for η in (4) as
G ρ η = μ 1 ( 1 ρ ) ( 1 ρ ) + u 1 u 2 u k x ,
where V a r ( u 1 ) = τ 2 / ( 1 ρ 2 ) and V a r ( u j ) = τ 2 j 2 . The right-hand-side vector [ 1 ( 1 ρ ) ( 1 ρ ) ] acts as a covariate vector (with known values given ρ ), and μ is its corresponding parameter. One can then show
μ | γ μ , η , y N D μ d μ , D μ
where
D μ = τ 2 ( 1 ρ 2 ) + ( k x 1 ) ( 1 ρ ) 2 + V ̲ μ 1 , d μ = g ρ G ρ η / τ 2 + V ̲ μ 1 μ ̲ 0 .
  • Step 6: τ 2 | γ τ 2 , η , y
Defining
G ˜ ρ = 1 ρ 2 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 G ρ ,
the prior in (4) implies
G ˜ ρ ( η ι k x μ ) N 0 , τ 2 I k x .
It follows that
τ 2 | γ τ 2 , η , y I G a ̲ τ + k x 2 , b ̲ τ 1 + 1 2 ( η ι k x μ ) G ˜ ρ G ˜ ρ ( η ι k x μ ) 1 .
  • Step 7: ρ | γ ρ , η , y
For ρ , the conditional posterior distribution is proportional to
ρ | γ ρ , η , y p ( ρ ) h ( ρ ) i = 2 k x ϕ η i μ ; ρ ( η i 1 μ ) , τ 2 ,
where the function h ( ρ ) is defined as
h ( ρ ) = ( 1 ρ 2 ) 1 / 2 exp 1 2 τ 2 ( η 1 μ ) 2 ( 1 ρ 2 ) .
  • Case 1: Continuous Prior
In the continuous case, p ( ρ ) is given in (6). The function h ( ρ ) introduces a non-conjugacy, as the resulting conditional posterior does not take the form of a standard distribution. We will thus use a Metropolis–Hastings (M-H) step to conduct the sampling.
If we were to ignore the contribution of the initial condition through the term h ( ρ ) , the conditional posterior distribution would be proportional to
ρ | γ ρ , η , y T N ( 1 , 1 ) ( D ρ d ρ , D ρ )
where
D ρ = i = 2 k x ( η i 1 μ ) 2 / τ 2 + V ̲ ρ 1 1 , d ρ = i = 2 k x ( η i μ ) ( η i 1 μ ) / τ 2 + V ̲ ρ 1 μ ̲ ρ .
Using (14) as a proposal distribution in an M-H step, we sample a candidate ρ c from (14) given that our chain is currently at ρ = ρ t . The M-H acceptance probability thus reduces to h ( ρ c ) / h ( ρ t ) . Specifically, we draw u * U ( 0 , 1 ) and accept the candidate ρ c provided
u * < 1 ρ c 2 1 ρ t 2 1 / 2 exp 1 2 τ 2 ( η 1 μ ) 2 ( ρ c 2 ρ t 2 ) .
Otherwise, we set the new value of the chain, ρ t + 1 , equal to ρ t .
  • Case 2: Discrete Prior
If a discrete prior for ρ is specified, as in (5), the posterior is also discrete with support points { ρ ̲ j } j = 1 J . Furthermore, the posterior probabilities are given up to proportionality as
P r ( ρ = ρ ̲ j | γ ρ , η , y ) p ̲ j h ( ρ ̲ j ) i = 2 k x ϕ η i μ ; ρ ̲ j ( η i 1 μ ) , τ 2 .
The expression on the right-hand side of the equation above can be computed for all j = 1 , 2 , , J and then normalized by dividing by the sum of all such values. A draw from the corresponding discrete distribution can then be obtained via the method of inversion.
  • Step 8: s | γ , η , y
The conditional posterior distribution of each component indicator is proportional to
P r ( s i = j | γ , η , y ) p j ω j 1 ϕ log ( [ ψ i x i π π 0 c ] 2 ) m j η i ω j , j = 1 , 2 , , 10 ,   and   i = 1 , 2 , , k x ,
with ϕ ( · ) denoting the standard normal density function. For a given i, the above can be calculated for all j = 1 , 2 , , 10 and then normalized to obtain a valid discrete distribution. The process is then repeated for all i.

4. Generated Data Experiments

In this section, we perform brief generated data experiments to investigate the performance of the stationary/estimated- ρ adaptive smoothing procedure outlined in Section 2 and Section 3. We seek to illustrate the method’s ability to recover the shape of an unknown regression function and to contrast the behavior of smoothing parameters under this method with those obtained under other assumed processes, including the random walk.
To this end, we begin by generating 1000 observations from a regression model with interesting changes in curvature. Specifically, we first sample x i i i d U ( 2 , 1.5 ) with U ( a , b ) denoting a uniform distribution on the interval ( a , b ) . To induce some degree of discreteness, consistent with our application of the following section, we round each x i value to the second decimal place. We then sample ϵ i i i d N ( 0 , 1 ) and generate y as follows:
y i = 1 + 0.5 x i + 2 exp 100 ( x i + 1 ) 2 + 2 ( 3 / 5 ) π 1 / 4 exp 25 x i 2 2 ( 1 25 x i 2 ) + σ ϵ i ,
with the error standard deviation σ = 0.25 . For our priors, we set μ ̲ π 0 = [ 2 2 ] , V ̲ π 0 = 1.0 × 10 5 I 2 , a σ = 3 , b σ = 5 / 2 , μ ̲ = 10 , V ̲ μ = 16 , a τ = 3 , and b τ = 0.5 . For the prior for ρ , we consider the discrete prior variant, with ρ { 0.5 , 0.51 , 0.52 , , 0.99 } . This support is selected to express preference for at least a moderate degree of positive persistence in the smoothing parameter process. Cases with negative ρ may also be considered, although positive persistence is seemingly a natural position to take when modeling regression functions that are believed to be smooth. In experiments not reported here, cases with negative ρ tended to produce immediate and large changes in η values and difficult-to-interpret oscillatory smoothing patterns as one departs from the stationary mean μ .
To construct the prior probabilities p ̲ j for each element of this discrete vector, we first calculate the ordinate of the standard normal density at [ ( ρ ̲ j μ ) / σ ] 2 where μ = 0.8 and σ = 0.1 . We perform this for each of the discrete ρ ̲ j cells. The resulting collection of values is then normalized by dividing by the sum of all such ordinates to produce a proper prior over this discrete vector of points. Of course, a variety of other methods for assigning/constructing the prior probabilities can also be employed.
Figure 1 provides a summary of the results from this generated data experiment. The top panel of Figure 1 provides a scatterplot of the raw data together with the stationary nonparametric estimates under the employed discrete prior for ρ . As shown in the figure, the method performs very well and accurately recovers the shape of and changes in the true regression function. The bottom panel of Figure 1 plots (in red) posterior means of the smoothing parameters η across each x * value. To gain a sense of how these results might differ from, or be similar to, smoothing results from other approaches (like the random walk), we estimate the model again under two different assumptions regarding the evolution of η . In the first model, we simply fix ρ = 0.3 throughout. In the second model, we consider the random walk specification. Parameters common across approaches are assigned the same priors, with hyperparameters of those priors given previously. Posterior means of the smoothing parameters/log volatilities η are presented for each of these two additional specifications in the lower panel of Figure 1.
The lower panel reveals a few interesting patterns and highlights important differences across approaches. First, results under all three priors show evidence of adaptability: values of η are relatively large where the function exhibits the most rapid changes in curvature. Conversely, when f ( x ) locally behaves as a linear function, η values are relatively small, shrinking results toward linearity over the corresponding region as a consequence. Second, we see differences in the degree of adjustment across approaches with different treatments of the AR parameter ρ . The smoothing parameters under the random walk model evolve slowly, exhibiting small local changes in η values, while results with estimated- ρ and those fixing ρ = 0.3 adjust quickly by comparison. Near the functional spike at x = 1 , for example, all approaches admit the largest value of η , while the ρ = 0.3 results then decline sharply to either side, returning rather quickly to values near the stationary mean μ . The approach with estimated- ρ lies somewhere in between the random walk and ρ = 0.3 cases: it resembles overall patterns produced by the random walk model, but adjusts the values of η more quickly as one moves away from the peaks in the regression function where curvature changes are most pronounced.
Figure 2 documents how learning takes place in the stationary model regarding the AR coefficient ρ under two different discrete priors. The left panel plots the prior (top left) for the analysis used to generate Figure 1 and the corresponding marginal posterior (bottom left). The prior employed is rather informative regarding the degree of persistence, but learning clearly occurs, as the posterior becomes left-skewed and shifts toward values of ρ near one. In the right panels of Figure 2, we repeat the analysis, this time increasing the generating σ value from 0.1 to 1, thus creating a nearly uniform prior for ρ across the discrete cells from 0.5 to 0.99 (top right). The posterior (bottom right) again differs considerably from the prior, and places even more of its mass near the unit root, suggesting preference for the random walk specification (or a close approximation to it) for this experiment. Mean absolute error (MAE), for example (letting f ( x ) denote the true value of the function in the experiment and f ^ ( x ) the estimated value (i.e., the posterior mean), we calculate the MAE as k x 1 i = 1 k x | f ^ ( x i * ) f ( x i * ) | ), was found to be 0.029 and 0.030 for the random walk and estimated- ρ specifications, respectively, and 0.037 when fixing ρ = 0.3 . The movement of the posterior of ρ toward the unit root is consistent with this overall preference for the random walk model in this experiment.
We consider an additional set of experiments, this time generating data from a linear model with f ( x ) = 1 + 0.5 x . With the exception of ρ , we continue to employ the same priors as those just used in the nonlinear experiment. For ρ , we again employ a discrete prior, this time over the support points ρ { 0.3 , 0.31 , 0.32 , , 0.99 } , and with discrete prior probabilities p ̲ j generated in a similar manner, with μ = 0.8 and σ = 0.2 . Results from this experiment are provided in Figure 3.
In the top left panel of Figure 3, we again see that the method performs admirably in recovering the simple linear regression function. In the bottom left panel, we compare posterior means of the smoothing parameters η for both the random walk and estimated- ρ cases. Here, we see that the random walk specification again slowly evolves, gradually moving toward smaller values of η , which further shrink the vector π toward linearity. Values of η are largest for small x in the random walk specification and then decline, owing to the influence of the N ( 10 , 16 ) prior for the initial smoothing parameter η 1 . This prior is chosen to match the prior employed for its counterpart, the stationary mean μ , in the estimated- ρ case. The relatively slowly adapting random walk specification takes some time to move away from the influence of this initial condition, whereas the estimated- ρ model immediately jumps to the stationary mean μ and remains there across the support of x.
Posterior means of η under the estimated- ρ model are uniformly smaller than those of the random walk model and thus more strongly impose linearity—an appropriate and desirable outcome for this example. The right panels of Figure 3 provide additional results, as they plot prior (top) and posterior (bottom) probabilities of ρ over its discrete support. Here, we again see that learning has taken place, as the marginal prior and posterior distributions are clearly different. Specifically, the posterior is observed to move toward interior ρ values and to place little mass around ρ 1 , unlike our previous nonlinear example of Figure 1 where the posterior moved toward the unit root. Finally, we note that the MAE for the estimated- ρ model was 0.005, while that of the random walk specification was 0.012. Taken together, we see that the linear and nonlinear experiments considered here provide examples where learning about ρ clearly occurs. This learning leads to posterior shifts in opposite directions in our two experiments, although both movements are in directions leading to improved fits.

5. Application

We illustrate our methods using data from the National Longitudinal Survey of Youth (NLSY) 1979. The NLSY is a nationally representative sample of young men and women, all born between 1957 and 1964, containing a wealth of demographic, academic, and labor market data on its participants. We use these data to better understand how behavioral problems among adolescents affect their academic achievement, the latter measured through a series of test scores within the NLSY. Specifically, the nonparametric methods derived in Section 2 and Section 3 are employed to flexibly explore the relationship between behavioral problem severity and academic achievement. Before discussing the results of this investigation, we first describe the variables analyzed in more detail.
The NLSY contains scores from the Peabody Individual Achievement Test (PIAT) in both mathematics and reading. The specific variables we extract and separately analyze are the total standard score in mathematics (MATHZ2000) (these parenthetical designations give the variable names in the NLSY), total standard score in reading recognition (RECOGZ2000), and total standard score in reading comprehension (COMPZ2000). Test scores range from 65 to 135 for each of the three tests in our sample, with sample means (and standard deviations) equal to 106.8 (13.5), 109.0 (14.2), and 105.0 (13.5) for math, reading recognition, and reading comprehension, respectively.
The NLSY also contains a supplemental mother’s file, in which mothers of the NLSY participants are asked to respond to a series of questions regarding their children. Included among responses in the mother’s file are a set of 28 questions surrounding specific behaviors that their children may or may not have demonstrated over the preceding three months. These questions touch on characteristics such as antisocial behavior, anxiousness/depression, headstrongness, hyperactivity, immature dependency, and peer conflict/social withdrawal (see https://nlsinfo.org/content/cohorts/nlsy79-children/topical-guide/assessments/behavior-problems-index-bpi (accessed on February 2025) for additional details). In answering these questions, mothers categorize the stated behaviors as occurring “often”, “sometimes”, or “not” (true). An overall index (BPIZX2000) is then constructed in the NLSY (which we denote as the BPI) by aggregating assigned scores across questions, with higher values of the index indicative of higher degrees of behavioral issues. In our final sample, the BPI ranges from a low of 74 to a high of 145, creating k x = 72 distinct values when constructing D as in Section 2. Of these 72 distinct values, nine cells are not represented in the sample, three contain exactly one observation, and no cells contain more than 44 observations. When cells are not empty, the average number of observations associated with each distinct BPI value is approximately 16.3. A seeming advantage of our approach is that our smoothness prior allows us to handle cells that are either infrequently represented or not represented at all; the prior borrows information from neighboring outcomes in forming posterior predictions at those values. Our final sample analyzed consists of n = 1076 observations, and each test score outcome is considered in a separate analysis. For our priors, we employ the continuous prior for ρ , setting μ ̲ ρ = 0.8 and V ̲ ρ = 0 . 15 2 . As discussed in Section 3, posterior simulation thus employs a Metropolis–Hastings step for sampling ρ , and associated acceptance rates were very high, equaling 0.91, 0.91, and 0.89 for mathematics, reading recognition, and reading comprehension, respectively. For the remaining hyperparameters, we set a σ = 3 , b σ = 0.005 , μ ̲ 0 = 3 , V ̲ μ 0 = 4 , a τ = 3 , and b τ = 0.5 . Each posterior simulator is run for 82,500 iterations, with the first 2500 discarded as the burn-in period.
The results of our analysis are summarized in Figure 4. The upper left panel of the figure provides plots of the nonparametric conditional mean estimates E ( Score | BPI ) ^ = E ( π | y ) for all three tests (solid) along with estimates from simple linear models (dashed). The remaining three panels of the figure provide more detailed results for each test score outcome. These panels plot nonparametric estimates (black, solid), linear estimates (black, dashed), posterior regions associated with the nonparametric estimates (light blue, shaded) (these are calculated as the posterior mean of each element of π ± 1.96 times the corresponding posterior standard deviation), and dummy variable coefficient estimates (red, dashed). The latter are obtained from OLS regressions of the given test score on a series of indicator variables for all BPI scores observed in the sample.
The upper left panel of the figure reveals several interesting results. First, the nonparametric estimates are suggestive of a pattern of nonlinearity that is consistent across all tests. Specifically, each conditional mean function is somewhat flat, though declining, for low values of the BPI, and then they turn more sharply negative once the BPI exceeds a value of approximately 90. Despite this near initial flatness, posterior means are found to be monotonically decreasing in the BPI across the full BPI support for all of our tests. This clearly suggests that behavioral issues have a globally negative effect on student achievement. When comparing the nonparametric results against linear estimates, another consistent pattern emerges: linear point estimates tend to overstate expected achievement at both tails of the BPI distribution, and this is particularly true at high values of the BPI for the reading recognition test score. Though not reported in the figure, we also note that the marginal posterior distributions of the autoregressive coefficient ρ are similar across tests, with posterior means (and standard deviations) equal to 0.78 (0.132), 0.78 (0.134), and 0.79 (0.139) for mathematics, reading recognition, and reading comprehension, respectively.
The remaining three panels of the figure reveal the value of smoothing for this application. Dummy variable coefficient estimates—which may be interpreted as the starting point for the smoothed analysis—are highly variable, making it difficult to clearly identify overall patterns in the data. The nonparametric estimates clearly cut through the middle of this erratic collection and produce results that are easily interpretable by comparison.
Figure 4 also reveals that estimates from the linear specifications do fall within our nonparametric posterior regions for all tests, as there is considerable estimation uncertainty at the right tail of the BPI distribution. We therefore consider an additional model, guided by the exploratory nonparametric analysis, to further investigate the potential existence of nonlinearities in the test score/BPI relationships. Specifically, we estimate a linear spline with a single unknown changepoint. Such an analysis will allow us to determine the most likely location for a change in curvature (slope) if one is present, along with values of the slopes themselves both before and after the change. The model we take to the data is of the form
y i = β 0 + β 1 B P I i + β 2 ( B P I i λ ) + + ϵ i ,
for each of our three test score outcomes. In Equation (16), λ denotes our single changepoint, the notation x + max { x , 0 } and ϵ i i i d N ( 0 , σ 2 ) . It is clear that the spline function permits a change in slope at the value λ : the slope of our line is β 1 when the BPI λ and β 1 + β 2 when the BPI > λ . We also place a uniform prior across integer-valued λ in the set { 79 , 80 , 81 , , 130 } , thus allowing at least a modest number of observations in each of the two slope regimes. For our priors, we choose
β 0 β 1 β 2 N 100 0 0 , 100 0 0 0 100 0 0 0 100 a n d σ 2 I G ( 3 , 0.005 ) .
The Gibbs sampler is employed to fit the model under these priors, running each posterior simulator for 110,000 iterations and discarding the first 10,000 as the burn-in period. Chan et al [11], Exercise 12.10, provide detailed steps for posterior simulation in a model of this form, and we follow those steps here.
If our earlier interpretations of the initial “flatness” of the regression functions followed by a more precipitous decline are correct, we should see those patterns echoed in the changepoint specification. Specifically, we expect to find (a) a revision of our uniform prior for the changepoint λ to a posterior for λ with a mode near BPI = 90 and (b) evidence that the slope of our regression function is more negative after the changepoint then before, suggesting that increments to the BPI have more substantial negative impacts on student achievement as one moves into the right tail of the BPI distribution.
Our estimation results provide confirmation of these patterns. Figure 5 plots posterior frequencies of the changepoint λ for each of our three test score outcomes. Interestingly, the posterior mode of the changepoint in each case is exactly 91, consistent with Figure 4, which shows a relatively flat function until the BPI ≈ 90, followed by a negative decline.
Table 1 provides posterior means and standard deviations for the regression coefficients β and variance parameter σ 2 in the changepoint model. As shown in the table, β 1 —which is interpreted as the pre-changepoint slope—has a posterior mean near zero and a correspondingly large posterior standard deviation. This, again, is consistent with an initial period of relative flatness in the relationship between the BPI and test scores. The regression coefficient β 2 , however, which is interpreted as the change in slope after λ , has consistently negative posterior means across test scores and places most of its mass over negative values: the posterior probability that β 2 < 0 is 0.93, 0.98, and 0.92 across math, reading recognition, and reading comprehension scores, respectively. These results provide additional confirmation that further movements into the right tail of the behavioral problems index distribution produce the largest negative impacts on student achievement.
We conclude with a comparison of our Bayesian nonparametric results and those obtained via frequentist kernel regression. Specifically, we additionally employ the Nadaraya–Watson (Nadaraya [18] and Watson [19]) or constant kernel estimator to estimate our regression relationships nonparametrically:
E ( S c o r e | B P I = B P I 0 ) ^ = i = 1 n S c o r e i K B P I i B P I 0 h i = 1 n K B P I i B P I 0 h ,
where a Gaussian kernel K is selected and the bandwidth/global smoothing parameter h is set equal to five. The results are presented in Figure 6, with analysis again conducted separately for each test score outcome.
Table 1. Coefficient posterior means (and standard deviations below) from changepoint model.
Table 1. Coefficient posterior means (and standard deviations below) from changepoint model.
MathReading Rec.Reading Comp.
β 0 108.08108.62106.16
(8.72)(9.20)(9.02)
β 1 0.0080.0290.014
(0.106)(0.113)(0.110)
β 2 −0.217−0.323−0.204
(0.182)(0.183)(0.197)
σ 2 178.27195.84176.92
(7.88)(8.65)(7.82)
As shown in the figure, point estimates from the Bayesian and Nadaraya–Watson approaches are similar, but depart at the right tail of the BPI distribution, particularly for the mathematics and reading comprehension test score outcomes. These departures both illustrate the potential value of adaptive smoothing and reveal limitations when using a constant global smoothing parameter. In our sample, only 1.46% ( n = 15 ) of observations contain a BPI in excess of 130, and thus considerable uncertainty exists when estimating the function over this region. A constant bandwidth h = 5 is chosen within our kernel regression to suitably smooth the functions over small/moderate BPI values, while this same level of smoothing also implies that just a few noisy data points are used to estimate the functions at the largest BPI values. This often implies a rather erratic (and seemingly implausible) kernel-based point estimate when the BPI > 130. In contrast, the Bayesian adaptive procedure imposes a higher degree of smoothing over this region and leads to a continuation of the overall downward trend present in the figures for smaller values of the BPI.

6. Conclusions

A new scheme for nonparametric Bayesian estimations of regression functions was presented. An efficient Gibbs sampling algorithm was developed that involves sampling from standard distributions, permits local adaptability in the degree of smoothing, and imposes stationarity in the smoothing parameter process. The algorithm was shown to perform well in generated data experiments, and its value was illustrated in an application relating a behavioral problems index (BPI) to student achievement. Analysis of those data showed that the test score/BPI relationship is relatively flat/slightly negative when behavioral problems are low, while marginal increments to the BPI have a clearly negative effect on student achievement in math and reading as one moves into the right tail of the BPI distribution.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study can be accessed from https://www.nlsinfo.org/investigator/pages/login (accessed on October 2024).

Acknowledgments

The author is grateful for the research assistance provided by Yaling Qi in acquiring this data.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Smith, M.; Kohn, R. Nonparametric Regression Using Bayesian Variable Selection. J. Econom. 1996, 75, 317–343. [Google Scholar] [CrossRef]
  2. Kohn, R.; Smith, M.; Chan, D. Nonparametric Regression Using Linear Combinations of Basis Functions. Stat. Comput. 2001, 11, 313–322. [Google Scholar] [CrossRef]
  3. Chib, S.; Greenberg, E.; Simoni, A. Nonparametric Bayes Analysis of the Sharp and Fuzzy Regression Discontinuity Designs. Econom. Theory 2023, 39, 481–533. [Google Scholar]
  4. Rasmussen, C.E.; Williams, C.K.I. Gaussian Processes for Machine Learning; MIT Press: Cambridge, MA, USA, 2006. [Google Scholar]
  5. Dixon, M.; Halperin, I.; Bilokon, P. Machine Learning in Finance; Springer: Berlin/Heidelberg, Germany, 2020. [Google Scholar]
  6. Koop, G.; Poirier, D. Bayesian Variants of Some Classical Semiparametric Regression Techniques. J. Econom. 2004, 123, 259–282. [Google Scholar] [CrossRef]
  7. Koop, G.; Tobias, J. Semiparametric Bayesian Inference in Smooth Coefficient Models. J. Econom. 2006, 134, 283–315. [Google Scholar] [CrossRef]
  8. Chib, S.; Jeliazkov, I. Inference in Semiparametric Dynamic Models for Binary Longitudinal Data. J. Am. Stat. Assoc. 2006, 101, 685–700. [Google Scholar] [CrossRef]
  9. Chan, J.; Tobias, J. An Alternate Parameterization for Bayesian Nonparametric/Semiparametric Regression. In Advances in Econometrics: Topics in Identification, Limited Dependent Variables, Partial Observability, Experimentation, and Flexible Modeling; Emerald Insight: Leeds, UK, 2019; Volume 40B, pp. 47–64. [Google Scholar]
  10. Tobias, J.; Bond, T. Semiparametric Bayesian Estimation in an Ordinal Probit Model with Application to Life Satisfaction Across Countries, Age and Gender. J. Econom. 2025. in process. [Google Scholar] [CrossRef]
  11. Chan, J.; Koop, G.; Poirier, D.; Tobias, J. Bayesian Econometric Methods, 2nd ed.; Cambridge University Press: Cambridge, UK, 2019; p. 484. [Google Scholar]
  12. Kline, B.; Tobias, J. The Wages of BMI: Bayesian Analysis of a Skewed Treatment Response Model with Nonparametric Endogeneity. J. Appl. Econom. 2008, 23, 767–793. [Google Scholar] [CrossRef]
  13. Chib, S.; Greenberg, E.; Jeliazkov, I. Estimation of Semiparametric Models in the Presence of Endogeneity and Sample Selection. J. Comput. Graph. Stat. 2009, 18, 321–348. [Google Scholar]
  14. Omori, Y.; Chib, S.; Shephard, N.; Nakajima, J. Stochastic Volatility with Leverage: Fast and Efficient Likelihood Inference. J. Econom. 2007, 140, 425–449. [Google Scholar] [CrossRef]
  15. Del Negro, M.; Primiceri, G. Time-Varying Structural Vector Autoregressions and Monetary Policy: A Corrigendum. Rev. Econ. Stud. 2015, 1342–1345. [Google Scholar] [CrossRef]
  16. Chan, J.C. Comparing Stochastic Volatility Specifications for Large Bayesian VARs. J. Econom. 2023, 235, 1419–1446. [Google Scholar]
  17. Kim, S.; Shephard, N.; Chib, S. Stochastic Volatility: Likelihood Inference and Comparison with ARCH Models. Rev. Econ. Stud. 1998, 65, 361–393. [Google Scholar]
  18. Nadaraya, E.A. On Estimating Regression. Theory Probab. Its Appl. 1964, 9, 141–142. [Google Scholar] [CrossRef]
  19. Watson, G.S. Smooth Regression Analysis. Sankhya Ser. A 1964, 26, 359–372. [Google Scholar]
Figure 1. Regression function and smoothing parameter posterior means: nonlinear example.
Figure 1. Regression function and smoothing parameter posterior means: nonlinear example.
Mathematics 13 01162 g001
Figure 2. Prior and posterior distributions of ρ : μ = 0.8 , σ = 0.1 prior (left panels) and μ = 0.8 , σ = 1 prior (right panels).
Figure 2. Prior and posterior distributions of ρ : μ = 0.8 , σ = 0.1 prior (left panels) and μ = 0.8 , σ = 1 prior (right panels).
Mathematics 13 01162 g002
Figure 3. Regression function, smoothing parameter, and AR coefficient results: linear example.
Figure 3. Regression function, smoothing parameter, and AR coefficient results: linear example.
Mathematics 13 01162 g003
Figure 4. Estimation results from nonparametric regressions of achievement on BPI.
Figure 4. Estimation results from nonparametric regressions of achievement on BPI.
Mathematics 13 01162 g004
Figure 5. Posterior frequencies of changepoint λ .
Figure 5. Posterior frequencies of changepoint λ .
Mathematics 13 01162 g005
Figure 6. Bayesian and Nadaraya–Watson estimates of achievement/BPI relationships.
Figure 6. Bayesian and Nadaraya–Watson estimates of achievement/BPI relationships.
Mathematics 13 01162 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tobias, J.L. Adaptive Bayesian Nonparametric Regression via Stationary Smoothness Priors. Mathematics 2025, 13, 1162. https://doi.org/10.3390/math13071162

AMA Style

Tobias JL. Adaptive Bayesian Nonparametric Regression via Stationary Smoothness Priors. Mathematics. 2025; 13(7):1162. https://doi.org/10.3390/math13071162

Chicago/Turabian Style

Tobias, Justin L. 2025. "Adaptive Bayesian Nonparametric Regression via Stationary Smoothness Priors" Mathematics 13, no. 7: 1162. https://doi.org/10.3390/math13071162

APA Style

Tobias, J. L. (2025). Adaptive Bayesian Nonparametric Regression via Stationary Smoothness Priors. Mathematics, 13(7), 1162. https://doi.org/10.3390/math13071162

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop