1. Introduction
Recently, the varying coefficient model has attracted much attention among econometricians and statisticians. One attractive feature of this model is its ability to capture the nonlinearity of the data without suffering from the “curse of dimensionality”. In general, it is of the form
where
are responses;
and
are associated covariates;
is a
p-dimensional vector of unknown functions;
are independent and identically distributed random errors with
and
.
Due to its flexibility, the varying coefficient model has been studied in many different contexts and has been successfully applied to nonlinear time series analysis, longitudinal and functional data analysis, panel data analysis, spatial data analysis, and time-varying models in finance. See, for example, the work of Cai
et al. [
1], Cai [
2], Cai and Li [
3], Cai
et al. [
4], Fan and Zhang [
5], Fan
et al. [
6], Fotheringham
et al. [
7], Hoover
et al. [
8], Li
et al. [
9] and Xiao [
10], among others.
In the above models, the varying coefficient model is generally estimated by the local-linear approach. Usually, the errors are assumed to be i.i.d. to start. However, in applications, heteroscedasticity is often found in residuals from both cross-sectional and time series modelling. In the context of the linear regression model, it is well known that if the errors are heteroscedastic, then the generalized least-squares (GLS) estimator is more efficient than ordinary least-squares (OLS) estimator. To the best of our knowledge, there has been no work on the problem of designing an efficient estimation method for varying coefficient models with heteroscedastic errors. In this paper, we propose an efficient estimator for varying coefficients based on the local linear approach.
The paper is structured as follows. We introduce an efficient estimator in
Section 2, and their asymptotic properties are given in
Section 3. We report the results of some Monte Carlo simulations in
Section 4.
2. Efficient Estimation
Without considering heteroscedasticity, we apply a local linear regression technique to estimate the varying coefficient functions. For each given
u, the local linear estimator
of
is the part corresponding to
of the minimizer of
where
K is a kernel function,
h is a bandwidth and
. Then we have
where
and
.
The estimator ignores the information contained in the variance matrix and it is inefficient. To overcome this, we propose a class of efficient estimators in the following.
Denote
, for the moment where we assume that
is known. Multiply both sides of model (1.1) by
, we have the following homoscedastic varying coefficient model
where
and
with
.
Applying the local linear approach to model (2.3), the efficient estimator of
is given as follows
where
,
3. Asymptotic Property
First, we make the following assumptions. Let .
Assumption 1. The errors are independent and and .
Assumption 2. The random variable U has a bounded support Π. Its density function is Lipschitz continuous and bounded away from 0 on its support.
Assumption 3. The matrixes and are non-singular for each .
Assumption 4. There is an such that and for some such that as .
Assumption 5. have continuous second derivatives in .
Assumption 6. The function is a symmetric density function with compact support and the bandwidth h satisfies and as .
For the estimator
, Cai
et al. [
1] proved the following result:
Theorem 1 Under the assumptions 1–6, the estimator
is asymptotically normal, namely,
where
,
.
For the estimator , we obtain the following result by the Theorem 1 directly.
Theorem 2 Under the assumptions 1–6, the estimator
is asymptotically normal, namely,
where
.
Denote
. By the proof of Theorem 1 in Cai
et al. [
1], we have
and
Since
, then we have
This implies that
is asymptotically more efficient than
in terms of asymptotic covariance matrix.
Remark 1. Since depends on the unknown parameters , it is infeasible. To provide a feasible efficient estimator of , we need to estimate consistently. It is not difficult to show that the resultant feasible estimator has the asymptotic property as .
Remark 2. To obtain the consistent estimator of the variance function
, it is important to model
. Several kinds of variance function have been proposed. Discussion on the parametric variance function can be found in Carroll and Ruppert [
11]. Muller and Stadtmuller [
12], Chiou and Muller [
13] and Ruppert
et al. [
14] studied nonparametric variance estimation. Muller and Zhao [
15] proposed a general semiparametric variance function model in a fixed design regression setting. Keilegom and Wang [
16] considered a general class of mean-variance regression models, in which both the mean function and the variance function were semiparametrically modeled. Zhu
et al. [
17] consider a single-index structure to study heteroscedasticity in a single-index regression model with high-dimensional predictors.
4. Simulation Studies
In this section we compare the behavior of the conventional estimator
with that of the new estimator
, given in (2.2) and (2.4), respectively, when the sample size is finite. The data are generated from the following varying coefficient model
where
,
. Firstly, we consider the following four known variance functions:
Secondly, we consider the case that the variance function is unknown. For simplicity, the variance function is assumed to have the following parametric structure,
with
. Obviously, we can build the following linear regression model
with
. In practice,
is not available, but it may be estimated by
, where
are the local linear estimates of model (4.1) without considering the heteroscedasticity structure. Applying the least squares approach to liner model (4.2) with
was replaced by
, we can obtain the estimators of
, denoted by
and
respectively. Accordingly, we get the estimator of
, as
.
To study the effect of the distribution of the error for our method, we take the following three different types of the error distribution, (1) , (2) , (3) . The Gauss kernel function and are used in our simulation studies.
We compare the proposed efficient estimator
with that of the ordinary local linear estimator
by using the estimated mean average squared error (MASE),
where
are the estimate of the coefficient
in
replications. The simulation results are presented in Table 1, and for all the scenarios we studied, the proposed efficient estimators outperform the ordinary local linear estimators.
Table 1.
Mean average squared error (MASE) index for the estimators of varying coefficients.
Table 1.
Mean average squared error (MASE) index for the estimators of varying coefficients.
Variance | Sample | | | |
---|
Function | n | | | |
---|
A | 30 | 0.4873 0.1354 | 0.5125 0.1277 | 0.5109 0.1374 |
| 50 | 0.4063 0.0805 | 0.3820 0.0804 | 0.3686 0.0805 |
| 80 | 0.3146 0.0539 | 0.3054 0.0523 | 0.2910 0.0525 |
B | 30 | 0.2018 0.1936 | 0.1872 0.1791 | 0.1835 0.1752 |
| 50 | 0.1129 0.1077 | 0.1174 0.1126 | 0.1114 0.1072 |
| 80 | 0.0728 0.0696 | 0.0734 0.0706 | 0.0733 0.0706 |
C | 30 | 0.1469 0.1302 | 0.1502 0.1269 | 0.1571 0.1304 |
| 50 | 0.1024 0.0920 | 0.1039 0.0893 | 0.1015 0.0884 |
| 80 | 0.0670 0.0679 | 0.0731 0.0660 | 0.0734 0.0645 |
D | 30 | 0.1446 0.1422 | 0.1390 0.1368 | 0.1487 0.1474 |
| 50 | 0.0841 0.0831 | 0.0844 0.0832 | 0.0819 0.0807 |
| 80 | 0.0612 0.0606 | 0.0556 0.0549 | 0.0595 0.0589 |
E | 30 | 0.3950 0.3936 | 0.4227 0.4133 | 0.3996 0.3913 |
| 50 | 0.3094 0.3039 | 0.3137 0.3068 | 0.3077 0.3015 |
| 80 | 0.2549 0.2508 | 0.2513 0.2508 | 0.2544 0.2520 |
5. Conclusions
In this paper, we focus on the estimation problem of the varying coefficient model with heteroscedastic errors. Based on the local linear method, we develop a simple approach to estimate the nonparametric coefficient functions by taking the estimated error heteroscedasticity into account. The resulting estimators are shown to have smaller asymptotic variances than the conventional local-linear estimators. The asymptotic normality of the proposed estimator is established. Furthermore, some simulation experiments are performed to evaluate the finite sample behaviors of the proposed estimators.