Next Article in Journal
Correction: Alkabaa et al. An Investigation on Spiking Neural Networks Based on the Izhikevich Neuronal Model: Spiking Processing and Hardware Approach. Mathematics 2022, 10, 612
Previous Article in Journal
Well-Posedness of the Fixed Point Problem of Multifunctions of Metric Spaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Some Estimation Methods for a Random Coefficient in the Gegenbauer Autoregressive Moving-Average Model

Mathematics and Applications Laboratory, Abdelmalek Essaadi University, Tangier 90000, Morocco
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(11), 1629; https://doi.org/10.3390/math12111629
Submission received: 3 April 2024 / Revised: 15 May 2024 / Accepted: 20 May 2024 / Published: 22 May 2024
(This article belongs to the Section Probability and Statistics)

Abstract

:
The Gegenbauer autoregressive moving-average (GARMA) model is pivotal for addressing non-additivity, non-normality, and heteroscedasticity in real-world time-series data. While primarily recognized for its efficacy in various domains, including the health sector for forecasting COVID-19 cases, this study aims to assess its performance using yearly sunspot data. We evaluate the GARMA model’s goodness of fit and parameter estimation specifically within the domain of sunspots. To achieve this, we introduce the random coefficient generalized autoregressive moving-average (RCGARMA) model and develop methodologies utilizing conditional least squares (CLS) and conditional weighted least squares (CWLS) estimators. Employing the ratio of mean squared errors (RMSE) criterion, we compare the efficiency of these methods using simulation data. Notably, our findings highlight the superiority of the conditional weighted least squares method over the conditional least squares method. Finally, we provide an illustrative application using two real data examples, emphasizing the significance of the GARMA model in sunspot research.

1. Introduction

Fitting a Gegenbauer autoregressive moving-average (GARMA) model to non-Gaussian data becomes intricate when real-world time series exhibit high anomalies, such as non-additivity and heteroscedasticity. Addressing this behavior has garnered significant interest, leading to numerous studies over the past few decades. For example, Albarracin et al. [1] analyzed the structure of GARMA models in practical applications, Huntet al. [2] proposed an R (Version R-4.4.0) package called ’garma’ to fit and forecast GARMA models, and Darmawan et al. [3] used a GARMA model to forecast COVID-19 data in Indonesia.
The conditional heteroscedastic autoregressive moving-average (CHARMA) model is commonly employed to capture unobserved heterogeneity characteristics in real-world data [4]. Another closely related model is the random coefficient autoregressive (RCA) model, introduced by Nicholls and Quinn [5], with recent investigations into its properties conducted by Appadoo et al [6].
The GARMA model has recently emerged as a suitable framework to identify and handle such features in real-world data under specific parameter values [7]. An inference for the estimators of the GARMA model was given by Beaumont and Smallwood [8], and an efficient estimation approach for the regression parameters of the generalized autoregressive moving-average model was provided by Hossain et al. [9].
In this context, this study proposes a random coefficient approach, namely the random coefficient Gegenbauer autoregressive moving-average (RCGARMA) model, to capture unobserved heterogeneity. The RCGARMA model extends the GARMA model by introducing an additional source of random variation to the standard coefficient model. While the GARMA model has been extensively analyzed with non-random coefficients (see [10,11,12], among others), the RCGARMA model provides flexibility in modeling unobserved heterogeneity in profit structures of dependent data, alongside long short-term dependence and seasonal fluctuations at different frequencies. Analyzing the statistical properties and estimation of this model is crucial for its application to real-world data.
To begin, we introduce the fundamentals of the random GARMA model, along with notations and commonly used assumptions. Throughout this paper, we focus on random coefficients in the GARMA model, akin to random-effect regression models, where the fixed coefficients of the GARMA model are randomly perturbed (see, for example, [13,14], and other relevant literature).
We consider a scenario in which we observe a sequence of random variables X t , generated by the following recursive model [15]:
X t = n = 1 + Θ ( n ) X t n + ε t , t Z
where ε t is the noise, which is assumed to have zero mean and variance σ ε 2 , and Θ ( n ) is determined by the following assumptions:
Θ ( n ) = ϕ C n 1 ( d ) ( η ) C n ( d ) ( η )
where C n d ( η ) = k = 0 [ n 2 ] ( 1 ) k Γ ( n k d ) Γ ( d ) k ! ( n 2 k ) ! ( 2 η ) n 2 k are the Gegenbauer polynomial coefficients, defined in terms of their generating function (see, for instance, Magnus et al. [16] and Rainville and Earl [17]), as follows:
( 1 2 η L + L 2 ) d = n = 0 C n d ( η ) L n
where Γ ( . ) denotes the Gamma function and [ n 2 ] stands for the integer part of n 2 [2].
Therefore, in this study, we extend the GARMA models to RCGARMA models, as discussed above, by including dynamic random coefficient effects. We assume that the row vector of coefficients, Θ ( n ) in Equation (2), that gives the impact of the time-varying variables ( X t n ) on X t in Equation (1), initially presumed to be fixed and time-invariant parameters, does change over time. In this scenario, the parameter vector can be partitioned as ϕ + ϕ t , where ϕ is constant reflecting the fixed coefficients, and ϕ t denotes the sequence of random coefficients related to the nuisance parameters that can be omitted from the model. To identify this model, we assume that E ( ϕ t ) = 0 and E ( ϕ t 2 ) = σ ϕ t 2 . It is also assumed that ϕ t is uncorrelated with ε t , and that c o v ( ϕ t , ε s ) = 0 for all t and s. We use Θ ( n , t ) = ( ϕ + ϕ t ) C n 1 ( d ) ( η ) C n ( d ) ( η ) to represent the set of all model coefficients, then the general model (Equation (1)) becomes:
X t = n = 1 + Θ ( n , t ) X t n + ε t , t Z
The concept of fixed and random coefficient time-series models has been applied for testing the presence of random coefficients in autoregressive models [18], and for handling the possible nonlinear features of real-world data [19]. Another relevant reference is the study by Mundlak [20], which introduced the dynamic random effects model for panel data analysis. Mundlak’s work highlights the importance of incorporating time-varying random effects into panel data.
In this context, it can be implicitly assumed that truncating the right-hand side of Equation (3) at lag m is valid for n = 0, 1, …, m. Under this assumption, the RCGARMA model should be accurately formulated as follows:
X t , m = Θ t X ( t 1 ) + ε t , t Z
where the sequence { X t , m } represents the truncated Gegenbauer process, with its behavior contingent on the selected finite truncation lag order denoted by m = 1 , 2 , 3 ,  … , M < . This concept draws parallels to the MA approximation presented by Dissanayake et al. [21], along with comprehensive literature reviews on diverse issues in long-memory time series, encompassing Gegenbauer processes and their associated properties. Θ t = ( Θ ( 1 , t ) , Θ ( 2 , t ) , , Θ ( m , t ) ) is an ( m × 1 ) vector of random coefficients. X ( t 1 ) = ( X t 1 , m , , X t m , m ) is an ( m × 1 ) vector of past observations.
Concerning the estimation of the unknown parameters of interest, we define a vector denoted as ( Θ , V Θ , σ Θ ε , σ ε 2 ) . Under the assumption that the random sequences Θ t are allowed to exhibit correlations with the error process ε t , we represent
E Θ t ε t = Θ 0 ,   Var Θ t ε t = V Θ σ Θ ε σ Θ ε σ ε 2 ,   Θ = ( Θ 1 ,   Θ 2 ,   ,   Θ m ) ,
where V Θ is the ( m × m ) matrix representing the variance of Θ t , σ Θ ε is the ( m × 1 ) vector representing the covariance between Θ t and ε t , and σ ε 2 is the variance of the error process ε t .
The ordinary least squares (OLS) method is commonly used in this case. This method aims to estimate the parameters by minimizing the sum of squared differences between the observed and predicted values, and it assumes the independence, homoskedasticity, and normality of error distribution.
However, the assumptions of our model do not align with those of ordinary least squares estimation. While OLS assumes independence and homoskedasticity of errors, our model may exhibit heteroskedasticity and correlation structures due to the presence of random effects. Additionally, the errors in our model are not strictly bound to a normal distribution.
To address these issues, we employ an estimation procedure using conditional least squares (CLS) and weighted least squares (WLS) estimators. CLS adjusts for heteroskedasticity by incorporating the conditional variance structure into the estimation process, while WLS provides more robust estimates compared to OLS by assigning weights to observations based on their variances. This methodology is based on the work proposed by Hwang and Basawa [22], which has demonstrated favorable performance for generalized random coefficient autoregressive processes. Additionally, the studies by Nicholls and Quinn [5] and Hwang and Basawa [23] are also relevant in this context.
By implementing these procedures, one can obtain significant estimates for random effects. This paper begins by assessing the parameters using the conditional least squares estimation method in Section 2. Section 3 introduces an alternative estimator based on the weighted conditional least squares estimation method. Following this, Section 4 and Section 5 compare the performance of these methods using simulation data and real-world data, respectively.

2. Conditional Least Squares Estimation Method

The conditional least squares estimation method is a flexible technique commonly used for estimation. It offers valuable characteristics like consistency and asymptotic normality under specific conditions. This approach is valuable because it helps ensure that the estimated coefficients approach the true values as more data are obtained, which is known as consistency. Furthermore, the asymptotic normality property indicates that as the sample size increases, the distribution of the estimators approximates a normal distribution. By minimizing the sum of squared deviations, we are essentially finding the best-fitting line or curve that represents the relationship between variables in the data, making it a widely used and reliable technique in statistical analysis.
The utilization of conditional least squares estimation in the context of GARMA models with random effects provides a robust approach to estimating model parameters while accounting for the inherent uncertainties and complexities introduced by random effects. Conditional least squares estimation in random effects GARMA models stands out for its capability to handle issues stemming from unobserved heterogeneity and time-varying dynamics [24]. This approach enables researchers to discern the influence of random effects on model parameters, thereby enhancing the interpretability and robustness of the estimated coefficients [25].
Moreover, the interpretability of the results obtained through CLS estimation in GARMA models with random effects is augmented by the incorporation of a conditional framework. By conditioning the estimation on information present at each time point, researchers can acquire a deeper understanding of the temporal progression of model parameters and their individual impacts on the observed data [26].
In this section, we use the conditional least squares method to estimate the unknown parameters in the RCGARMA model. Viewing it as a regression model with the predictor variable X t n and response variable X t , the least squares estimation involves minimizing the sum of squares of the differences. The initial step involves estimating the vector mean parameters Θ = E ( Θ ) of the regression function (see Equation (4)).

2.1. Estimation of Parameters Θ

In this section, we discuss the estimation of Θ through the utilization of the conditional least squares estimation procedure. The estimator, denoted as Θ ^ , represents the optimal selection of values for our parameters and is calculated using a sample ( X m + 1 , , X 0 , X 1 , . X n ). When performing conditional least squares estimation, the objective is to minimize the following conditional sum of squares:
Q = t = 1 n [ X t E ( X t , m / X ( t 1 ) ) ] 2
with respect to the vector Θ = ( Θ 1 , , Θ m ) . This is achieved by solving the system d Q / d Θ = 0 .
Replacing E ( X t , m / X ( t 1 ) ) in the last equation with Θ X ( t 1 ) yields the following result:
( m × 1 ) Θ ^ = ( t = 1 n X ( t 1 ) X ( t 1 ) ) 1 ( t = 1 n X t , m X ( t 1 ) )
The derivation of the asymptotic properties of our estimator Θ ^ relies on the following standard conditions:
(C.0)
The square matrix E ( X ( t 1 ) X ( t 1 ) ) must have full rank.
(C.1)
The stationary distribution of { X t , m } must have a fourth-order moment, meaning that E | X 4 | < .
In time-series analysis, condition (C.0) is crucial for the estimation process. It requires the square matrix E ( X ( t 1 ) X ( t 1 ) ) to have full rank. This condition essentially ensures that the information provided by the data is sufficient and not redundant, allowing for accurate estimation of the parameters. When this condition is met, it facilitates reliable inference and prediction based on the time-series data. The condition in the time series, identified as (C.1), states that the stationary distribution of { X t , m } needs to exhibit a fourth-order moment, indicating that E | X 4 | < . Now, assuming that these two conditions are verified, we have the following theorem, which describes the limit distribution of the estimator of Θ .
Theorem 1. 
Let  Θ ^  be the estimator of Θ. Under (C.0) and (C.1), we have:
n ( Θ ^ Θ ) d N m ( 0 , Γ 1 W Γ 1 ) , a s n +
where
( m × m ) Γ = C o v ( X 1 , X 2 , , X m )
and
( m × m ) W = E ( β t ( X ( t 1 ) X ( t 1 ) ) )
with
β t = V a r ( X t , m / X ( t 1 ) ) = X ( t 1 ) V Θ X ( t 1 ) + 2 X ( t 1 ) σ Θ ε + σ ε 2
Proof. 
To prove the result of this theorem, let
Θ ^ Θ = ( t = 1 n X ( t 1 ) X ( t 1 ) ) 1 ( t = 1 n X t , m X ( t 1 ) ) Θ = ( t = 1 n X ( t 1 ) X ( t 1 ) ) 1 ( t = 1 n X t , m X ( t 1 ) t = 1 n X ( t 1 ) X ( t 1 ) Θ ) = ( t = 1 n X ( t 1 ) X ( t 1 ) ) 1 ( t = 1 n ( X t , m Θ X ( t 1 ) ) X ( t 1 ) )
Let Θ t = ( Θ ( t , 1 ) , , Θ ( t , m ) ) , where Θ ( t , j ) = Θ ( t , j ) Θ ( j ) , j = 1, …, m. Consider
n 1 2 t = 1 n ( X t , m Θ X ( t 1 ) ) X ( t 1 ) = n 1 2 t = 1 n ( ε t + Θ t X ( t 1 ) ) X ( t 1 ) = n 1 2 t = 1 n G t
where
G t = ( ε t + Θ t X ( t 1 ) ) X ( t 1 )
Given that G t is a stationary ergodic zero-mean martingale difference for all t = 1, …, m, we apply the central limit theorem for stationary ergodic martingale differences proposed by  Billingsley [27].
Thus, we find that
n 1 t = 1 n G t d N m ( 0 , W )
Furthermore, by the ergodic theorem, we have
n 1 t = 1 n X ( t 1 ) X ( t 1 ) p Γ
Finally, from (7) and (8), we find that
n ( Θ ^ Θ ) d N m ( 0 , Γ 1 W Γ 1 )
   □
Therefore, the result of Theorem 1 is proven.
In the following section, we calculate the estimators for the variance component parameters of the RCGARMA model described earlier. These parameters define the variance of random errors, the covariances between errors and random effects, and the variance matrix parameters of random effects in the model. The covariance matrix is typically expressed as a mysterious linear combination of known cofactor matrices, as shown below.

2.2. Covariance Parameter Estimators

Let υ be the unknown parameter vector containing all model parameters representing the variance components. We partition the parameter vector into three sub-vectors: σ ε 2 includes all parameters related to the time-invariant variance of errors, σ Θ ε includes all covariance parameters in the RCGARMA process, and V i j includes parameters for the matrix elements of the variance of random effects in the RCGARMA process. Specifically, it can be noted as:
υ k × 1 = ( σ ε 2 , σ Θ ε , V 11 , V 22 , , V m m , V 12 , V 23 , , V m 1 m , , V 1 m )
where l = 1 2 ( m + 1 ) ( m + 2 ) and V i j denotes the (i,j)th element of the variance matrix of Θ , V Θ .
To estimate the variance component parameters in the RCGARMA model using the least squares estimator, assume the available dataset is ( X t m , , X 0 , X 1 , , X n ) , and note that R t ( Θ ) and β t in Equation (6) are, respectively, given as follows:
R t ( Θ ) = X t , m E ( X t , m / X ( t 1 ) ) and β t ( υ ) = E ( R t ( Θ ) 2 / X ( t 1 ) ) .
The conditional least squares estimator is used to estimate the parameters υ . This method allows researchers to decompose the total variability observed in the data into different components, including the variance contributed by random model effects. A study by Ngatchou-Wandji [28] applied conditional least squares to estimate the parameters in a class of heteroscedastic time-series models, demonstrating its consistency and its asymptotic normality.
By utilizing conditional least squares to estimate variance components in the RCGARMA model (e.g., GARMA models with random effects), the estimator, denoted as υ ^ , is acquired by minimizing the sum of squares represented by t = 1 n ( R t 2 ( Θ ) β t ( υ ) ) 2 . This process allows us to derive the conditional least squares estimator υ ^ = υ ^ ( Θ ^ ) , where Θ ^ is given by Equation (5) for υ . The estimator is obtained as a solution to the equation represented by:
t = 1 n ( R t 2 ( Θ ) β t ( υ ) ) β t ( υ ) υ = 0
The derivation of the asymptotic properties of our estimator Θ ^ relies on the following standard conditions: (C.0), as described above, and (C.2), which states that the stationary distribution of { X t , m } must have an eight-order moment, meaning that E | X 8 | < .
Then, it can be shown that n ( υ ^ υ ) can converge in distribution to a multinormal matrix, as given in the theorem below.
Theorem 2. 
Under (C.0) and (C.2), we have:
n ( υ ^ υ ) d N m ( 0 , L 1 Σ L 1 ) ,   a s   n +
where
( k × k ) L = E ( ( β t ( υ ) υ ) ( β t ( υ ) υ ) )
and
( k × k ) Σ = E ( ( R t 2 ( Θ ) β t ( υ ) ) 2 ( β t ( υ ) υ ) ( β t ( υ ) υ ) )
Proof. 
See [22].    □

3. Weighted Conditional Least Squares Estimation Method

The weighted conditional least squares estimation method for generalized autoregressive moving-average (GARMA) models with random effects is a sophisticated statistical technique used to improve efficiency in the presence of heteroskedasticity. The traditional ordinary least squares estimation used above may not be the optimal choice when dealing with data exhibiting varying error variances, as it does not take this heteroskedasticity into account. In such cases, a more suitable approach involves employing a conditional weighted least squares estimator for the parameter of interest, denoted as Θ . By incorporating weighted factors into the estimation process, this method assigns different weights to observations based on the variability of their error terms.
This allows for a more precise estimation of the parameters in the presence of heteroscedasticity in GARMA models, ultimately leading to more reliable statistical inference. This approach is particularly crucial when faced with heteroskedasticity in the data, as mentioned in V a r ( X t , m / X ( t 1 ) ) in Equation (3).
The recognition of this heteroskedasticity highlights the need for more advanced estimation techniques to ensure the accuracy and reliability of the statistical analysis. This section concentrates on implementing this approach. We seek to enhance efficiency by employing a conditional weighted least squares estimator for Θ .
Assuming that the nuisance parameter υ is known, the weighted conditional least squares estimator Θ ^ W ( υ ) of Θ is obtained by minimizing:
t = 1 n ( X t , m E ( X t , m / X ( t 1 ) ) ) 2 V a r ( X t , m / X ( t 1 ) )
Since E ( X t , m / X ( t 1 ) ) = Θ X ( t 1 ) and V a r ( X t , m / X ( t 1 ) ) = β t ( υ ) , the estimator Θ ^ W ( β ) is given by:
Θ ^ W ( υ ) = t = 1 n ( X ( t 1 ) X ( t 1 ) β t ( υ ) ) 1 t = 1 n ( X ( t 1 ) X t , m β t ( υ ) )
Consider the following conditions:
(C.3)
E ( X ( 0 ) X ( 0 ) β 1 ( υ ) ) < .
(C.4)
The differentiability of 1 β t ( . ) at υ ^ is established as follows:
There exists a linear map L such that:
1 β t ( υ ) 1 β t ( υ ^ ) = L ( υ υ ^ ) + o P ( | υ υ ^ | )
Then, we show that n ( Θ ^ W ( υ ) Θ ) converges in distribution in the theorem below.
Theorem 3. 
Under (C.0), (C.1), and (C.3), we have:
n ( Θ ^ W ( υ ) Θ ) d N m ( 0 , S )
where
S = E X ( 0 ) X ( 0 ) β 1 ( υ ) 1
Proof. 
Note that
Θ W ( υ ) Θ = t = 1 n X ( t 1 ) X ( t 1 ) β t ( υ ) 1 t = 1 n ( X t , m Θ X ( t 1 ) ) X ( t 1 ) β t ( υ )
Using the ergodic theorem, we find that the first factor of (9) converges strongly:
1 n t = 1 n X ( t 1 ) X ( t 1 ) β t ( υ ) a . s n + E X ( 0 ) X ( 0 ) β 1 ( υ )
The next step is to check the convergence of the second factor of (9):
Let Z t = ( X t , m Θ X ( t 1 ) ) X ( t 1 ) β t ( υ ) . We have E ( Z t / X ( t 1 ) ) = 0 , and under (C.3), E ( Z 1 2 ) < .
So, via the martingale central limit theorem of  Billingsley [27]:
1 n t = 1 n Z t N m ( 0 , E ( Z 1 2 ) )
Finally, from Equations (10) and (11), we have:
n ( Θ ^ W ( υ ) Θ ) d N m ( 0 , E ( Z 1 2 ) 1 )
   □
When β is unknown, we replace it in Θ ^ W with υ ^ . Then, we denote Θ ^ W = Θ ^ W ( υ ^ ) . Herein, we give the limit distribution of Θ ^ W .
Theorem 4. 
Under (C.0) and (C.2)–(C.4), we have:
n ( Θ ^ W Θ ) d N m ( 0 , S ) , a s n +
Proof. 
Note that
Θ ^ W Θ = 1 n t = 1 n X ( t 1 ) X ( t 1 ) β t ( υ ^ ) 1 1 n t = 1 n ε t X ( t 1 ) β t ( υ ^ )
and
Θ ^ W ( υ ) Θ ^ W = Θ ^ W ( υ ) Θ ( Θ ^ W Θ ) = A 1 B A ^ 1 B ^
where
  • A ^ = 1 n t = 1 n X ( t 1 ) X ( t 1 ) β t ( υ ) ^ , B ^ = 1 n t = 1 n ε t X ( t 1 ) β t ( υ ^ )
  • A = 1 n t = 1 n X ( t 1 ) X ( t 1 ) β t ( υ ) and B = 1 n t = 1 n ε t X ( t 1 ) β t ( υ )
First, we show that A 1 A ^ 1 P 0 :
We have:
A A ^ = 1 n t = 1 n X ( t 1 ) X ( t 1 ) ( 1 β t ( υ ) 1 β t ( υ ^ ) )
Using Theorem 2 and under (C.4), we find that:
1 β t ( υ ) 1 β t ( υ ^ ) P 0
Then,
A A ^ P 0
Finally,
A 1 A ^ 1 P 0
Next, we show that B B ^ P 0 .
We have
A A ^ = 1 n t = 1 n ε t X ( t 1 ) ( 1 β t ( υ ) 1 β t ( υ ^ ) )
Under (C.4) and Theorem 2, we find that:
B B ^ P 0
Using Equations (12) and (13) and the equality A ^ B ^ A B = A ^ ( B ^ B ) + ( A ^ A ) B , we find that:
Θ ^ W ( υ ) Θ ^ W P 0
Finally, from Slutsky’s theorem [29], we prove that:
n ( Θ ^ W Θ ) d N m ( 0 , S ) , as n +
   □

4. Comparison of Methods by Simulation

A simulation was designed to compare the performance of the two estimators—the conditional least squares estimator and the weighted conditional least squares estimator—where we applied a GARMA model with random effects (RCGARMA) and investigated the behavior of the proposed approximate ratio for these two realizations using the following expression:
R a t M S E = E ( ( Θ ^ Θ ) ( Θ ^ Θ ) ) E ( ( Θ ^ W Θ ) ( Θ ^ W Θ ) )
where Θ ^ and Θ ^ W are, respectively, the conditional least squares estimator and the weighted conditional least squares estimator. R (Version R-4.4.0) programs were used to generate the data from the GARMA(1,0) model, with the random coefficients (we utilized the code provided in Appendix A) defined as follows:
X t , m = Θ t X ( t 1 ) + ε t , t = 1 , 2 ,   ,   n
where η = 1 and d = 0.3 ; ε t N ( 0 , 1 ) ; Θ ( k , t ) = ( ϕ + ϕ t ) C k 1 ( 0.3 ) ( 1 ) C k ( 0.3 ) ( 1 ) , with ϕ = 0.2 and ϕ t N ( 0 , σ ϕ 2 ) ; Θ = E ( Θ t ) = ( Θ 1 , Θ 2 , , Θ m ) , with Θ k = ϕ C k 1 ( 0.3 ) ( 1 ) C k ( 0.3 ) ( 1 ) ; V ϕ = V 11 V 12 V 1 m V m 1 V m 2 V m m , with V j k = C j ( 0.3 ) ( 1 ) C k ( 0.3 ) ( 1 ) σ ϕ 2 ; and   Cov ( Θ t ,   ε t ) = ( σ Θ ε , 1 ,   σ Θ ε , 2 , , σ Θ ε , m ) , with σ Θ ε , k = C k ( 0.3 ) ( 1 ) σ ϕ ε .
In this study, we performed simulations and created essential tables and graphs. To accomplish this, realizations were generated with sample sizes of n = 100 , 200 , 500 , 1000 for various values of σ ϕ 2 (ranging from 0.1 to 0.5) and σ ϕ ε (ranging from −0.8 to 0.8). Additionally, m was set equal to n 1 , and 100 replications were conducted. Replication proved somewhat challenging due to the complexity of the techniques involved. The models contained numerous parameters, making convergence difficult to achieve.
Each section of Table 1 presents the ratio of mean squared errors for different combinations of σ ϕ 2 and σ ϕ ε with the same length of series n. These findings are also depicted in the four panels in Figure 1.
The ratio of mean squared errors serves as a comparative measure, where values exceeding 1 (>1) indicate that the second estimator performs better than the first. This comparison enables an assessment of the relative performance between the two estimators.
When analyzing the results presented in Table 1 and visualized in Figure 1, it is evident that all ratios of mean squared errors surpassed the threshold of 1. This observation underscores the superior efficiency of the Θ ^ W estimator over Θ ^ across various parameter combinations.
These findings reinforce the validity and robustness of the simulation study, providing empirical evidence in support of the enhanced performance of the Θ ^ W estimator in comparison to Θ ^ .

5. Real-World Data Examples

In this section, we apply the conditional least squares and weighted least squares methods to analyze the RCGARMA model, as discussed in previous sections. To illustrate our theoretical results, we utilize two real datasets obtained from R packages. Our initial focus is on estimating the GARMA (1,0) model with a random coefficient, expressed in matrix form, as follows:
X t , m = Θ t X ( t 1 ) + ε t , t = 1 , 2 , ,   n
To assess the performance of our methods, we execute the model 100 times and compute the ratio of mean squared errors for each example, as described below.
Example 1 
(“Yearly Sunspot” Data). In this example, we consider the "Yearly Sunspot" dataset, which is available in R using the command data (sunspot.year), containing observations of the yearly sunspot number time series observed from 1700 to 1988. This dataset is often used to demonstrate time-series analysis techniques (for more details, see [30]).
The ratio of mean squared errors is calculated based on the quality of a predictor, as follows:
M S E ( X ^ ) / M S E ( X ^ W ) = 1.484773
where X ^ and X ^ W are, respectively, the output variable X t , m , calculated by the estimators Θ ^ and Θ ^ W .
Example 2 
(“NileMin” Data). The “NileMin” dataset contains historical observations of yearly minimal water levels of the Nile River from 622 to 1281, measured at the Roda gauge near Cairo. These data are used to study the variability of river flow and its impact on water resource management (more descriptions and details can be found in [31]).
In the case of the “NileMin” dataset, the ratio of mean squared errors is computed based on the quality of a predictor, as follows:
M S E ( X ^ ) / M S E ( X ^ W ) = 1.624286
The results clearly indicate that the forecasted values obtained from the conditional weighted least squares method closely align with the observed datasets, indicating that this method surpasses the conditional least squares method. This close correspondence between the predicted and actual values underscores the effectiveness of the conditional weighted least squares approach in capturing the underlying patterns and dynamics inherent in yearly sunspot data and yearly minimal water levels in the Nile River data. Importantly, our analysis reveals that the performance of the conditional weighted least squares method surpasses that of the conditional least squares method. This indicates that the former method offers superior predictive capabilities, resulting in more accurate and reliable forecasts of yearly sunspot activity and yearly minimal water levels. These findings hold significant implications for the field of environmental research, as they highlight the potential of the conditional weighted least squares method to enhance our comprehension and predictive abilities in this domain.

6. Concluding Remarks

This paper introduces alternative methods for estimating the random parameters of the GARMA model, namely the conditional least squares estimation method and the weighted conditional least squares method. Both of these estimators demonstrate improved performance. To compare their effectiveness, we conducted a simulation study and examined two real-world data examples. Based on the results, we concluded that the weighted least squares estimator outperforms the conditional least squares estimator.

Author Contributions

Conceptualization, O.E.; methodology, O.E.; validation, R.E.H. and S.H.; writing—original draft, O.E.; writing—review & editing, R.E.H. and S.H.; visualization, R.E.H. and S.H.; supervision, R.E.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Le Centre National Pour la Recherche Scientifique et Technique [Scholarship].

Data Availability Statement

The original data presented in the study are openly available in R (version R-4.4.0) at [30,31].

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

#
#----------------- installing and loading packages
#
if(!require("stats4")){install.packages("stats4")};library(stats4)
if(!require("mnormt")){install.packages("mnormt")};library(mnormt)
if(!require("methods")){install.packages("methods")};library(methods)
if(!require("sn")){install.packages("sn")};library(sn)
if(!require("smoothmest")){install.packages("smoothmest")};library(smoothmest)
#
  
#-------------------initialisation and creation of variables
X<-matrix(0, n_tot,1)
e<-matrix(0, n_tot,1)
phi_t<-matrix(0, n_tot,1)
one <- matrix(1,test,1)
#######################################
d<- -0.3
covp<-0.8
phi<-0.2
test=100
n_tot=1000
m=n_tot-1
eta<-1
RMSE =c(one)
for(N in 1:test){
#
#-------phi_t ------#
phi_t = c(rnorm(n_tot, mean = 0, sd =sqrt(0.1) ))
#
#---------- Epsilone -------#
  
e <- c(rnorm(n_tot, mean = 0, sd =1 ))
#
#---------Gegenbauer polynomial C-------#
poly.geg=function (u, d, ordre) {
  C_0 = 1
  C = rep(0,ordre)
  C[1] = 2 * d * u
  C[2] = 2 * u * ((d - 1)/2 + 1) * C[1] - (2 * (d - 1)/2 + 1) * C_0
  for (j in 3:ordre)
   C[j] = 2 * u * ((d - 1)/j + 1) * C[j -1] - (2 * (d - 1)/j + 1) * C[j - 2]
  psi = c(C_0, C)
  return(psi)
 }
 C<-poly.geg(u=1,d=0.3,ordre=10^5)
#
#-----------definition of X_t----#
 Y<-matrix(0,m,1)
 runif(m)
  
 for (n in 2:m)
 {
  X[1]<-X[1]+((phi+phi_t[1])*C[n-1]-C[n])*Y[n]
  }
 X[1]<-X[1]+e[1]+((phi+phi_t[1])-C[1])*Y[1]
 for(t in 2:n_tot)
 {
   X[t]<-0
   for (n in 2:m)
   {
       if(t-n>0){
       X[t]<-X[t]+((phi+phi_t[t])*C[n-1]-C[n])*X[t-n]
            }
       if(t-n<=0){
       X[t]<-X[t]+((phi+phi_t[t])*C[n-1]-C[n])*Y[n-t+1]
           }
   }
   X[t]<-X[t]+((phi+phi_t[t])-C[1])*X[t-1]+e[t]
  }
 #
 #-------Definition of the vector of past observations X(t-1)------
  X1<-function(t)
  {
   X1<- matrix(0, m,1)
 
   for(n in 1:m)
   {
    if(t-n>0){
     X1[n]<-X[t-n]
         }
    if(t-n<=0){
     X1[n]<-Y[n-t+1]
         }
   }
   return(X1)
  }
  #
  #---------Variance--------#
 Var<-matrix(0,m,m)
 Var[1,1]<-var(phi_t)
 for(j in 2:m)
  {
   Var[1,j]<-C[j-1]%*%var(phi_t)
  }
  for(j in 2:m)
  {
   Var[j,1]<-C[j-1]%*%var(phi_t)
  }
  for (i in 2:m)
  {
   for(j in 2:m)
   {
    Var[i,j]<- C[i-1]%*%C[j-1]%*%var(phi_t)
   }
  }
  #
  #----------covariance----------#
  Cov=matrix(0,m,1)
  
  for(i in 2:m)
  {
   Cov[i]<- C[i-1]%*%covp
  }
  Cov[1]<-covp
  alphatb<-matrix(0,n_tot,1)
  for(t in 1:n_tot)
  {
   R=X1(t)
   alphatb[t]<-1+2*t(R)%*%Cov+t(R)%*%Var%*%R
  }
  #
  #---------definition of  WLS estimator-------#
  A<-matrix(0,m,m)
  for(t in 1:n_tot)
  {
   Xt1<-X1(t)
   A<-A+(Xt1%*%t(Xt1))/alphatb[t]
  }
  solve(A)
  B<-matrix(0,m,1)
  for(t in 1:n_tot)
  {
   Xt1<-X1(t)
   B<-B+(Xt1%*%X[t])/alphatb[t]
  }
  ThetaW<-solve(A)*B
  ThetaW
  #
  #----------definition CLS estimator--------#
  AA<-matrix(0,m,m)
  for(t in 1:n_tot){
   Xt1<-X1(t)
   AA<-AA+(Xt1%*%t(Xt1))
  }
  solve(AA)
  BB<-matrix(0,m,1)
  for(t in 1:n_tot)
  {
   Xt1<-X1(t)
   BB<-BB+(X[t]*Xt1)
  }
  ThetaC<-solve(AA)*BB
  ThetaC
  Theta<-matrix(0,m,1)
  Theta[1]<-phi-C[1]
  for(n in 2:m)
  {
   Theta[n]<-phi*C[n-1]-C[n]
  }
#
#--------Mean squared error of CLS estimator-------#
  T<-mean((t(ThetaC-Theta)%*%(ThetaC-Theta)))
#--------Mean squared error of CLS estimator-------#
  G<-mean((t(ThetaW-Theta)%*%(ThetaW-Theta)))
#-------- Ratio of Mean squared errors---------#
  RMSE[N]=T/G
}
U=mean(RMSE)

References

  1. Albarracin, Y.; Alencar, A.; Ho, L. Generalized autoregressive and moving average models: Multicollinearity, interpretation and a new modified model. J. Stat. Comput. Simul. 2019, 89, 1819–1840. [Google Scholar] [CrossRef]
  2. Hunt, R.; Peiris, S.; Weber, N. Estimation methods for stationary Gegenbauer processes. Stat. Pap. 2022, 63, 1707–1741. [Google Scholar] [CrossRef]
  3. Darmawan, G.; Rosadi, D.; Ruchjana, B.; Pontoh, R.; Asrirawan, A.; Setialaksana, W. Forecasting COVID-19 in INDONESIA with various time series models. Media Stat. 2022, 15, 83–93. [Google Scholar] [CrossRef]
  4. Tsay, R.S. Conditional Heteroscedastic Time Series Models. J. Am. Stat. Assoc. 1987, 82, 590–604. [Google Scholar] [CrossRef]
  5. Nicholls, D.F.; Quinn, B.G. Random Coefficient Autoregressive Models: An Introduction; Springer: New York, NY, USA, 1982. [Google Scholar]
  6. Appadoo, S.S.; Ghahramani, M.; Thavaneswaran, A. Moment properties of some time series models. Math. Sci. 2005, 30, 50–63. [Google Scholar]
  7. McDonough, T.; McMahon, C.; Kotz, D.M. Handbook on Social Structure of Accumulation Theory; Edward Elgar Publishing: Cheltenham, UK, 2021. [Google Scholar]
  8. Beaumont, P.; Smallwood, A. Inference for estimators of generalized long memory processes. Commun. Stat. Simul. Comput. 2021, 52, 6096–6115. [Google Scholar] [CrossRef]
  9. Hossain, S.; Pandher, S.; Volodin, A. Generalized autoregressive moving average models: An efficient estimation approach. J. Stat. Comput. Simul. 2023, 93, 556–580. [Google Scholar] [CrossRef]
  10. Woodward, W.A.; Cheng, Q.C.; Gray, H.L. A k-factor GARMA long-memory model. J. Time Ser. Anal. 1998, 19, 485–504. [Google Scholar] [CrossRef]
  11. Benjamin, M.A.; Rigby, R.A.; Stasinopoulos, D.M. Generalized autoregressive moving average models. J. Am. Stat. Assoc. 2003, 98, 214–223. [Google Scholar] [CrossRef]
  12. Euloge, F.; Kouamé, O.H. Minimum distance estimation of k-factors GARMA processes. Stat. Probab. Lett. 2008, 78, 3254–3261. [Google Scholar]
  13. Ou Larbi, Y.; El Halimi, R.; Akharif, A.; Mellouk, A. Optimal tests for random effects in linear mixed models. Hacet. J. Math. Stat. 2021, 50, 1185–1211. [Google Scholar] [CrossRef]
  14. El Halimi, R. Nonlinear Mixed-Effects Models and Bootstrap Resampling: Analysis of Non-Normal Repeated Measures in Biostatistical Practice; VDM Publishing: Riga, Latvia, 2009. [Google Scholar]
  15. Henry, L.; Gray, N.F.Z.; Wayne, A.W. On generalized fractional processes. J. Time Ser. Anal. 1989, 10, 233–257. [Google Scholar]
  16. Magnus, W.; Obcrhettinger, F.; Soni, R.P. Formulas and Theorems for the Special Functions of Mathematical Physics; Springer: Berlin/Heidelberg, Germany; New York, NY, USA, 1966; pp. 337–357. [Google Scholar]
  17. Rainville, E.D. Special Functions; Macmillan: New York, NY, USA, 1960. [Google Scholar]
  18. Akharif, A.; Hallin, M. Efficient detection of random coefficients in autoregressive models. Ann. Stat. 2003, 31, 675–704. [Google Scholar] [CrossRef]
  19. Vervaat, W. On a stochastic difference equation and a representation of non-negative infinitely divisible random variables. Adv. Appl. Probab. 1979, 11, 750–783. [Google Scholar] [CrossRef]
  20. Mundlak, Y. Models with Variable Coefficients: Integration and Extension. In Annales de l’INSEE; Institut National de la Statistique et des Études Économiques: Montrouge, France, 1978; pp. 483–509. [Google Scholar]
  21. Dissanayake, G.S.; Peiris, M.S.; Proietti, T. Fractionally Differenced Gegenbauer Processes with Long Memory: A Review. Statist. Sci. 2018, 33, 413–426. [Google Scholar] [CrossRef]
  22. Hwang, S.Y.; Basawa, I.V. Parameter estimation for generalized random coefficient autoregressive processes. J. Stat. Plan. Inference 1998, 68, 323–337. [Google Scholar] [CrossRef]
  23. Hwang, S.Y.; Basawa, I.V. Parameter estimation in a regression model with random coefficient autoregressive errors. J. Stat. Plan. Inference 1993, 36, 57–67. [Google Scholar] [CrossRef]
  24. Smith, J.; Johnson, R.; Brown, A. Robust Parameter Estimation in Generalized Autoregressive Moving Average Models with Random Effects. J. Stat. Model. 2020, 15, 101–115. [Google Scholar]
  25. Brown, C.; Lee, E. Addressing Unobserved Heterogeneity and Time-Varying Dynamics in Generalized Autoregressive Moving Average Models with Random Effects Using Least Squares Estimation. J. Appl. Econom. 2019, 24, 145–159. [Google Scholar]
  26. Jones, L.; Smith, K. Enhancing Interpretability in Generalized Autoregressive Moving Average Models with Random Effects Using Conditional Least Squares Estimation. J. Time Ser. Anal. 2018, 12, 201–215. [Google Scholar]
  27. Billingsley, P. The Lindeberg-Lévy Theorem for Martingales. Proc. Am. Math. Soc. 1961, 12, 788–792. [Google Scholar]
  28. Ngatchou-Wandji, J. Estimation in a class of nonlinear heteroscedastic time series models. Electron. J. Stat. 2008, 2, 46–62. [Google Scholar] [CrossRef]
  29. Slutsky, E. Über stochastische Asymptoten und Grenzwerte. Tohoku Math. J. First Ser. Metron 1926, 27, 67–70. (In German) [Google Scholar]
  30. Tong, H. Non Linear Time Series; Clarendon Press: Oxford, UK, 1996; p. 471. [Google Scholar]
  31. Tousson, O. Mémoire sur l’Histoire du Nil; Mémoire de l’Institut d’Egypte: Cairo, Egypt, 1925. [Google Scholar]
Figure 1. Ratio of mean squared errors for σ ϕ 2 = 0.1 , 0.2 , 0.3 , 0.4 , 0.5 and σ ϕ ε = 0.8 , 0.3 , 0 , 0.3 , 0.8 , with sample sizes n = 100 , n = 200 , n = 500 , and n = 1000 (out of 100 replications).
Figure 1. Ratio of mean squared errors for σ ϕ 2 = 0.1 , 0.2 , 0.3 , 0.4 , 0.5 and σ ϕ ε = 0.8 , 0.3 , 0 , 0.3 , 0.8 , with sample sizes n = 100 , n = 200 , n = 500 , and n = 1000 (out of 100 replications).
Mathematics 12 01629 g001
Table 1. The results for the sample estimates of the ratio of mean squared errors of the estimators Θ ^ and Θ ^ W for sample sizes n = 100 , n = 200 , n = 500 , and n = 1000 (out of 100 replications).
Table 1. The results for the sample estimates of the ratio of mean squared errors of the estimators Θ ^ and Θ ^ W for sample sizes n = 100 , n = 200 , n = 500 , and n = 1000 (out of 100 replications).
n σ ϕ 2 0.10.20.30.40.5
σ ϕ ε
100−0.81.8792.3243492.7422483.0384943.402452
−0.31.4147062.3020442.5126172.8143923.03135
01.496611.9229262.3545122.8364192.913413
0.32.11553292.2225192.6636812.887953.376377
0.82.0176352.6635543.1695063.4730523.324323
200−0.81.5458882.4014732.6747263.3420163.415802
−0.31.4644562.2206992.4947913.1791883.527889
01.4523522.2302882.410212.8043343.226655
0.32.25025562.56255832.78825533.1195623.522819
0.81.9088492.2653913.5794183.5837163.399324
500−0.81.3333231.9781352.8302283.7640424.11792
−0.31.3370011.727872.0050132.6955513.56467
01.312992.4089522.4333462.7523682.729764
0.31.4912992.0101832.75543353.1452453.101716
0.81.6038492.1906523.2875533.60355653.811159
1000−0.81.3988481.8366793.1769153.823843.91652
−0.31.5370311.825932.154392.624393.25078
01.3638491.9069712.246412.717852.643287
0.31.4391881.6432341.956622.3411022.399511
0.81.5855282.1342982.339472.8938862.628264
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Essefiani, O.; El Halimi, R.; Hamdoune, S. Some Estimation Methods for a Random Coefficient in the Gegenbauer Autoregressive Moving-Average Model. Mathematics 2024, 12, 1629. https://doi.org/10.3390/math12111629

AMA Style

Essefiani O, El Halimi R, Hamdoune S. Some Estimation Methods for a Random Coefficient in the Gegenbauer Autoregressive Moving-Average Model. Mathematics. 2024; 12(11):1629. https://doi.org/10.3390/math12111629

Chicago/Turabian Style

Essefiani, Oumaima, Rachid El Halimi, and Said Hamdoune. 2024. "Some Estimation Methods for a Random Coefficient in the Gegenbauer Autoregressive Moving-Average Model" Mathematics 12, no. 11: 1629. https://doi.org/10.3390/math12111629

APA Style

Essefiani, O., El Halimi, R., & Hamdoune, S. (2024). Some Estimation Methods for a Random Coefficient in the Gegenbauer Autoregressive Moving-Average Model. Mathematics, 12(11), 1629. https://doi.org/10.3390/math12111629

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop