Next Article in Journal
Sensor Selection and State Estimation of Continuous mAb Production Processes
Previous Article in Journal
A Lightweight and Privacy-Preserving Authentication Protocol for Healthcare in an IoT Environment
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Asymptotic Normality of M-Estimator in Linear Regression Model with Asymptotically Almost Negatively Associated Errors

School of Mathematics and Statistics, Bigdata Modeling and Intelligent Computing Research Institute, Hubei University of Education, Wuhan 430205, China
Mathematics 2023, 11(18), 3858; https://doi.org/10.3390/math11183858
Submission received: 18 July 2023 / Revised: 30 August 2023 / Accepted: 4 September 2023 / Published: 9 September 2023
(This article belongs to the Section Probability and Statistics)

Abstract

:
This paper studies a linear regression model in which the errors are asymptotically almost negatively associated (AANA, in short) random variables. Firstly, the central limit theorem for AANA sequences of random variables is established. Then, we use the central limit theorem to investigate the asymptotic normality of the M-estimator for the unknown parameters. Some results for independent and negatively associated (NA, in short) random variables are extended to the case of AANA setting. Finally, a simulation is carried out to verify the asymptotic normality of the M-estimator in the model.

1. Introduction

A lot of researchers have derived the asymptotic properties of the estimators in linear regression models with independent random errors, which are not reasonable in many applications. Therefore, it is of great significance to study linear regression models with dependent errors.
Consider the following linear regression model:
Y i = x i β 0 + e i ,   ( 1 i n , n 1 ) ,
where β 0 is an unknown p × 1 regression parameter vector, x i = ( x i 1 , x i 2 , , x i p ) is a known p × 1 explanatory vector, { Y i } are the scalar response variables, and the errors { e i } are asymptotically almost negatively associated random variables.
In many situations, the assumption that random variables are independent is not suitable, so people often study dependent random variables. AANA random variables are widely used. Independent and NA random variables are special AANA random variables. The concept of AANA sequences was proposed by Chandra and Ghosal [1].
Definition 1
([1]). A random sequence { X n , n 1 } is said to be asymptotically almost negatively associated (AANA, in short) if a non-negative sequence q ( n ) 0 ( n ) exists, such that:
C o v ( f ( X n ) , g ( X n + 1 , X n + 2 , , X n + k ) ) q ( n ) ( V a r ( f ( X n ) ) V a r ( g ( X n + 1 , X n + 2 , , X n + k ) ) ) 1 / 2 ,
for all n , k 1 and for all coordinate wise non-decreasing continuous functions f and g whenever variances exist. { q ( n ) , n 1 } refers to the mixing coefficients.
Many results on the AANA sequences of random variables have been achieved. The family of AANA sequences contains NA (in particular, independent) sequences (with q ( n ) = 0 , n 1 ) and some more random sequences (see [1]). The following two examples are AANA sequences but not NA: ξ n = ( 1 + a n 2 ) 1 / 2 ( η n + a n η n + 1 ) (see [1]) and ζ n = η n + a n η n + 1 (see [2]), where η 1 , η 2 , are i . i . d . N ( 0 , 1 ) random variables, a n > 0 and a n 0 as n . Some Rosenthal-type inequalities for AANA sequences are obtained in [3]. More studies on AANA sequences of random variables can be found in [4,5,6].
M-estimation include LS and LAD estimation as special cases. In order to investigate the location parameter model, Huber [7] first introduced M-estimation. Huber [8] extended M-estimation to a generalized linear model and proposed the following definition.
Definition 2
([8]). Assume that ρ ( ) is some suitable function on R 1 . The M-estimator β ^ n of β 0 is defined by:
β ^ n = arg min β 0 R p i = 1 n ρ ( Y i x i β 0 ) .
A series of useful results using M-estimation have been derived by various researchers. Under independent errors, the consistency and asymptotic normality, the methods and theories of linear hypothesis testing, and the linear representation of the M-estimator were discussed more comprehensively in [9,10,11,12,13]. Wu [14] studied the consistency of the M-estimator in a linear model with NA errors. More studies on M-estimation can be found in [15,16,17,18].
However, we have not found studies which focus on the asymptotic normality of the M-estimator in a linear regression model whose errors are AANA in the literature. Therefore, this article considers the linear regression model with AANA errors. The main novelties of this paper are as follows. Firstly, we use moment inequalities to establish the central limit theorem for AANA sequences, which extend the corresponding results for NA sequences to AANA sequences. Secondly, we use the central limit theorem to study the asymptotic normality of the M-estimators of the unknown parameters. The results extend the corresponding ones for independent and some dependent errors. At last, we carried out a simulation to verify the validity of the results.
The remainder of this paper is organized as follows. In Section 2, the central limit theorem for AANA random sequences is derived. In Section 3, the asymptotic normality of the M-estimators of unknown parameters is investigated under some mild conditions. In Section 4, a simulated study is presented to support the results. In Section 5, we conclude the paper.
In the paper, let C be a positive constant whose values may vary at different places. i . i . d . stands for independent and identically distributed. x + = max { 0 , x } and x = max { 0 , x } . means “defined as”. is the Euclidean norm.

2. Central Limit Theorem for AANA Random Sequences

To derive the asymptotic normality of M-estimators of unknown parameters, we give an important conclusion in this section first and call it the central limit theorem for AANA random sequences.
Theorem 1.
Assume that  { X n ; n 1 }  is an identically distributed AANA random sequence with mixing coefficients  { q ( n ) , n 1 }  and the following conditions are satisfied:
  • (A1) E X 1 = 0 ,  0 < E X 1 2 < ;
  • (A2)
    lim n E   S n 2 / n = σ 2 ;
  • (A3) there exists a strictly increasing natural numbers sequence { n k } , for 0 < α 1 , such that:
    k = 1 ( m k n k ) 1 + α / 2 <   and   lim k 1 m k E S n k 1 , m k 2 = σ 2 ;
  • (A4) n = 1 q ( n ) = h < ; Then,
    S n E S n 2 d N ( 0 , 1 ) ,
    where  n 0 = 0 ,  m k = n k n k 1 ( k 1 ) ,  S m , n = j = 1 n X j + m ,  S n = S 0 , n , and  d stands for convergence in distribution.
Remark 1.
In Theorem 1, except for the extra condition ( A 4 ), other conditions are the same as that of Theorem 2.2 in Su and Chi [19].
As independent and NA sequences are special AANA sequences, we can easily obtain the following corollaries.
Corollary 1.
Let  { X n ; n 1 }  be an  i . i . d .  random sequence, as the conditions ( A 1 )–( A 3 ) are satisfied, and then conclusion (5) holds.
Corollary 2
([19]). Let { X n ; n 1 } be an identically distributed NA random sequence, as the conditions ( A 1 )–( A 3 ) are satisfied, and then conclusion (5) holds.
To prove Theorem 1, we need the following lemmas.
Lemma 1
([3]). If { X n , n 1 } is an AANA sequence whose mixing coefficients are  { q ( n ) , n 1 } , then so is  { f n ( X n ) , n 1 } , where  f 1 , f 2 , are non-increasing or non-decreasing functions.
Lemma 2
([3]). Let { X n , n 1 } be an AANA sequence of zero mean with mixing coefficients  { q ( n ) , n 1 } , then there exists a positive constant  C p depending only on  p , such that:
E ( max 1 i n | T i | p ) C p { i = 1 n E | X i | p + ( i = 1 n 1 q 2 / m ( i ) X i p ) 2 } ,
for any  n and  1 < p 2 , where  T i = j = 1 i X j ,   X i p = ( E | X i | p ) 1 / p ,   m = p / ( p 1 ) .
Lemma 3
([20]). Let { X n , n 1 } be an  i . i . d . random sequence with  E X j = 0 and  E | X j | 2 + α < for  0 α 1 and all  j 1 . Assume that  f : R R is twice derivative and  f α < . Then,
| E f ( S n / s n ) f d Φ | c f α s n 2 α j = 1 n E | X j | 2 + α ,
where  s n = ( E S n 2 ) 1 / 2 = ( j = 1 n E X j 2 ) 1 / 2 and  Φ is the standard normal distribution function.
Lemma 4
([21]). If { X n , n 1 } is an AANA sequence with zero mean and mixing coefficients  { q ( n ) , n 1 } , satisfying  σ n 2 E X n 2 < ,  n 1 , then
E S n 2 2 i = 1 n σ i 2 + 4 ( i = 1 n 1 q ( i ) σ i ) 2 ,
where  S n = i = 1 n X i .
Lemma 5.
Let  { X n , n 1 }  be an AANA sequence with mixing coefficients  { q ( n ) , n 1 } . Assume that  h 1 ( x )  and  h 2 ( y )  are absolutely continuous, bounded and real on  R  with  | h 1 ( x ) | C  and  | h 2 ( y ) | C . Then,
  | C o v [ h 1 ( X n ) , h 2 ( g ( X n + 1 , X n + 2 , , X n + k ) ) ] | C [ C o v ( X n , g ( X n + 1 , X n + 2 , , X n + k ) ) + q ( n ) X n 2 , 1 g ( X n + 1 , , X n + k ) 2 , 1 ]
for all  n , k 1  and for all coordinate wise non-decreasing continuous functions  g  whenever the variances exist, where  X 2 , 1 = 0 ( P ( | X | x ) ) 1 2 d x .
Proof. 
The proof is similar to that of the Lemma 3.5 in Zhang [22], with q ( n ) taking the place of its ρ ( X , Y ) . We omit the details here. □
Lemma 6.
Let  { X n , n 1 }  be an AANA sequence with mixing coefficients  { q ( n ) , n 1 } . If  E | X j | 2 < , then
| E j = 1 n e i λ j X j j = 1 n E e i λ j X j | C 1 j < l n | λ j λ l C o v ( X j , X l ) | + C j = 1 n 1 q ( j )
for any real  λ j ,  j = 1 , , n .
Proof. 
We proceed the proof by induction on n . The result is true for n = 1 trivially and is true for n = 2 by Lemma 5 and X 2 , 1 = 0 ( P ( | X | x ) ) 1 2 d x (see [23]). We assume that the result is true for n N . For n = N + 1 , we can suppose that for some  ε = ± 1 , δ = ± 1 and k { 1 , 2 , , N } , ε λ j 0 for 1 j k , while δ λ j 0 for k + 1 j N + 1 .
Write U = j = 1 k ε λ j X j and V = j = k + 1 N + 1 δ λ j X j . Then,
| E j = 1 n e i λ j X j j = 1 n E e i λ j X j | | E e i ε U + i δ V E e i ε U E e i δ V | + | E e i ε U | | E e i δ V j = k + 1 N + 1 E e i λ j X j | + | j = k + 1 N + 1 E e i λ j X j | | E e i ε U j = 1 k E e i λ j X j | | E e i ε U + i δ V E e i ε U E e i δ V | + | E e i δ V j = k + 1 N + 1 E e i λ j X j | + | E e i ε U j = 1 k E e i λ j X j | .
Noting that:
C o v ( U , V ) = j = 1 k l = k + 1 N + 1 ε δ λ j λ l C o v ( X j , X l ) = j = 1 k l = k + 1 N + 1 | λ j λ l | C o v ( X j , X l ) .
Therefore, from the induction hypothesis, it follows that:
| E j = 1 n e i λ j X j j = 1 n E e i λ j X j | C [ | ε δ C o v ( U , V ) | + q ( 1 ) ] + C [ k + 1 j l N + 1 | λ j λ l C o v ( X j , X l ) | + j = k + 1 N q ( j ) ] + C [ 1 j l k | λ j λ l C o v ( X j , X l ) | + j = 1 k 1 q ( j ) ] C 1 j l n | λ j λ l C o v ( X j , X l ) | + C j = 1 n 1 q ( j ) .
This completes the proof of Lemma 6. □
Proof of Theorem 1.
By (3), we have E S n 2 c n for sufficiently large n . From (4), we know that m k / n k 0 as k , thus we have m k / n k 1 1 for all sufficiently large k . By Lemma 2, we obtain that for each  ε > 0
P ( max 1 j m k | S n k 1 , j | > ε n k 1 ) 1 ε 2 n k 1 E max 1 j m k | S n k 1 , j | 2 C ( n k 1 ) 1 [ m k E X 1 2 + h 2 ( E X 1 2 ) ] C m k + h 2 n k 1 = C h 2 n k 1 + C ( 1 + m k n k 1 ) m k n k C h 2 n k 1 + C m k n k 0
as k . Hence, to prove (5), we only need to prove
S n k E S n k 2 d N ( 0 , 1 )
as k .
For each j N , we denote:
X ˜ j = j I ( X j < j ) + X j I ( | X j | j ) + j I ( X j > j ) ,
Y j = X ˜ j E X ˜ j ,   Z j = X j Y j .
Then, X ˜ j ,   Y j , and Z j are non-decreasing functions on  X j . Thus, by Lemma 1, we derive that  { X ˜ j } ,   { Y j } , and { Z j } are all AANA sequences, and
E Y j = 0 ,   E Z j = E X j E Y j = 0 ,
Z j = X j X ˜ j + E X ˜ j = ( X j X ˜ j ) E ( X j X ˜ j ) .
Therefore, it follows by Lemma 2 with p = 2 that:
1 n E ( j = 1 n Z j ) 2 1 n ( j = 1 n Z j 2 + h 2 E Z j 2 ) 1 n ( j = 1 n E ( X j X ˜ j ) 2 + C ) = 1 n { j = 1 n E ( ( X j X ˜ j ) 2 ( I ( | X j | j ) ) + I ( | X j | > j ) ) + C } 1 n { j = 1 n E X j 2 I ( | X j | > j ) + j = 1 n j P ( | X j | > j ) + C } 2 n { j = 1 n E X j 2 I ( | X j | > j ) } + C n .
From E X 1 2 < , we know that:
lim n 1 n E ( j = 1 n Z j ) 2 = 0 .
Thus, for each ε > 0 , we have:
P ( 1 n | j = 1 n Z j | > ε ) 1 ε 2 n E ( j = 1 n Z j ) 2 0 .
Since
1 n E ( j = 1 n Y j ) 2 = 1 n E ( j = 1 n ( X j Z j ) ) 2 = 1 n E S n 2 + 1 n ( j = 1 n Z j ) 2 2 n E ( S n j = 1 n Z j ) .
Using the Cauchy–Schwarz inequality and Lemma 2 with p = 2 , it follows that:
1 n | E ( S n j = 1 n Z j ) | ( 1 n E S n 2 ) 1 / 2 ( 1 n E ( j = 1 n Z j ) 2 ) 1 / 2 ( E X 1 2 + h 2 E X 1 2 n ) 1 / 2 ( 1 n E ( j = 1 n Z j ) 2 ) 1 / 2 0
as n . Hence, we derive by (3) that:
lim n E ( j = 1 n Y j ) 2 / n = σ 2 > 0 .
By (6)–(10), we know that to prove (5), we only need to prove:
j = 1 n k Y j / E ( j = 1 n k Y j ) 2 d N ( 0 , 1 )
as k .
It is easy to show that in Theorem 1, if
lim k 1 m k E ( j = 1 m k X n k 1 , j ) 2 = lim k 1 m k E S n k 1 , m k 2 = σ 2 ,
then
lim k 1 m k E ( j = 1 m k Y n k 1 + j ) 2 = σ 2 .
Write
U k = j = 1 m k Y n k 1 + j   and   T k = j = 1 k U j = j = 1 n k Y j .
Then, { U k ; k N } is an AANA random sequence (the proof is similar to that of Theorem 2.2 in Su and Chi [19]). Now, we take a sequence of independent random variables { V k ; k N } , such that V k and U k are identically distributed for all k N .
Let W k = j = 1 k V j . By (11), we obtain that to prove Theorem 1, we only need to prove:
P ( T k / E T k 2 < u ) d Φ ( u ) .
Denote:
γ j = E V j 2 = E U j 2 = E ( l = n j 1 + 1 m j Y l ) 2 ,
Γ k = j = 1 k γ j = j = 1 k E V j 2 = E ( j = 1 k V j ) 2 = E W k 2 .
By (12), we know that γ j / m j σ 2 as j . Hence,
Γ k n k = m 1 ( γ 1 / m 1 ) + + m k ( γ k / m k ) m 1 + + m k σ 2   as   k .
By (10), we have:
E T k 2 / n k = E ( j = 1 n k Y j ) 2 / n k σ 2   as   k .
Thus,
E T k 2 Γ k = E T k 2 n k n k Γ k σ 2 1 σ 2 = 1
as k .
Therefore, in order to prove (13), we only need to show that:
F k ( u ) P ( T k / Γ k < u ) d Φ ( u )
as k for all u R . We will apply Lemma 3 to prove (15).
Let
G k ( u ) P ( W k / Γ k < u ) = P ( W k / E W k 2 < u ) .
For given u R and any ε ( 0 , 1 ] , we construct two functions:
f ε : R R   and   g ε : R R .
Both of the two functions are three times derivative and satisfy the following conditions:
(1)
0 f ε ( x ) 1 ,   0 g ε ( x ) 1 ;
(2)
f ε ( x ) = { 1 , x u , 0 , x u + ε ,   g ε ( x ) = { 1 , x u ε , 0 , x u ;
(3)
For the α ( 0 , 1 ] of condition ( A 3 ).
f ε α C ε 2 α   and   g ε α C ε 2 α ,
where the definition of . α is as follows:
for 0 < α 1 , let f be a real function on R and define:
f α sup { | f ( x ) f ( y ) | / | x y | α ; x R , y R } ,
F α = { f ; R R , f α < } .
In fact, for ε = 1 , there exist functions satisfying the above conditions (see [20]). For 0 < ε < 1 , let:
f ε ( x ) = f 1 ( u + x u ε )   and   g ε ( x ) = g 1 ( u + x u ε ) ,
then they satisfy the above conditions. For each ε ( 0 , 1 ) , it follows that:
E g ε ( T k / Γ k ) F k ( u ) E f ε ( T k / Γ k ) .
So, we have
E g ε ( T k / Γ k ) Φ ( u ) F k ( u ) Φ ( u ) E f ε ( T k / Γ k ) Φ ( u )
and
| F k ( u ) Φ ( u ) | j = i 4 I j ( k , ε ) ,
where
I 1 ( k , ε ) = | E f ε ( T k / Γ k ) E f ε ( W k / Γ k ) | ,
I 2 ( k , ε ) = | E g ε ( T k / Γ k ) E g ε ( W k / Γ k ) | ,
I 3 ( k , ε ) = | E f ε ( W k / Γ k ) Φ ( u ) | = | E f ε ( W k / E W k 2 ) Φ ( u ) | ,
I 4 ( k , ε ) = | E g ε ( W k / Γ k ) Φ ( u ) | = | E g ε ( W k / E W k 2 ) Φ ( u ) | .
Thus, to prove (15), we only need to prove:
I j ( k , ε ) 0
as k , where j = 1 , 2 , 3 , 4 .
Next, we will prove (17) for j = 1 and fixed ε > 0 . By the property of f ε ( x ) , we have:
f ( x ) = f ( x ) = 0   for   x u   or   x u + ε .
From the definition of . α and f ε α c ε 2 α c 1 , we obtain that:
| f ε ( x ) | = | f ε ( x ) f ε ( u ) | c 1 | x u | α   for   u < x < u + ε
and | f ε ( x ) | is bounded in ( u , u + ε ) . Thus, | f ε ( x ) | d x < . Let
a ( t ) = e i t x f ε ( x ) d x = 1 i t e i t x f ε ( x ) d x .
It follows that:
| a ( t ) | 1 | t | | f ε ( x ) | d x c 1 | t | u u + ε | x u | α d x = c 1 | t | 0 ε x α d x = c 1 ε α + 1 | t | ( α + 1 ) = C | t | .
Denote
p k ( t ) = e i t x d F k ( u ) = E exp ( i t Γ k T k ) = E exp ( i t Γ k j = 1 k U j )
By the independence among V 1 , , V k and the identical distribution between V j and U j , we have:
r k ( t ) = e i t x d G k ( u ) = E exp ( i t Γ k W k ) = E e i t Γ k j = 1 k V j = E ( j = 1 k e i t V j Γ k ) = j = 1 k E exp ( i t Γ k V j ) = j = 1 k E exp ( i t Γ k U j ) ,
where G k ( u ) is defined by (16). From lemma 6, it follows that:
| p k ( t ) r k ( t ) | t 2 Γ k ( | C 1 j l k Cov ( U j , U l ) | + C j = 1 k 1 q ( j ) ) C t 2 Γ k ( | E T k 2 j = 1 k E U j 2 | + C h ) = C t 2 ( | E T k 2 Γ k 1 | + C h n k σ 2 ) C t 2 Δ k
as k .
By (14), we have:
Δ k = | E T k 2 / Γ k 1 | + C h n k σ 2 0   as   k .
It is easily to obtain that
I 1 ( k , ε ) = | f ε ( u ) d ( F k ( u ) G k ( u ) ) | = | ( F k ( u ) G k ( u ) ) f ε ( u ) d u | .
Hence, it follows that
I 1 ( k , ε ) = 1 2 π | ( p k ( t ) r k ( t ) ) a ( t ) t d t | C 2 π | p k ( t ) r k ( t ) | t 2 d t
from the Parsevar’s formula (see [24]) (17) and (18).
When Δ k = 0 , it is obviously that I 1 ( k , ε ) = 0 . When Δ k > 0 , let Λ k = Δ k 1 / 2 ; then we obtain that Λ k as k from (17).
Let
J 1 = | t | Λ k | p k ( t ) r k ( t ) | t 2 d t   and   J 2 = | t | Λ k | p k ( t ) r k ( t ) | t 2 d t .
By (19), we have:
J 1 | t | Λ k Δ k d t = 2 Λ k Δ k = 2 Δ k 0   as   k .
From the definitions of p k ( t ) and r k ( t ) , we get | p k ( t ) r k ( t ) | 2 . Thus,
J 2 2 | t | Λ k t 2 d t = 4 / Λ k = 2 Δ k 0   as   k ,
and then
I 1 ( k , ε ) c ( J 1 + J 2 ) 0
as k .
When j = 2 , 3 , 4 , similar to the proof of (20), we have I 2 ( k , ε ) 0 , I 3 ( k , ε ) 0 , and I 4 ( k , ε ) 0 as k .
This completes the proof of Theorem 1. □

3. Asymptotic Normality of M-Estimator

For the model (1), we suppose that ρ is a non-monotonic convex function on R 1 with left derivatives Ψ and right derivatives Ψ + . Select Ψ , such that Ψ ( u ) Ψ ( u ) Ψ + ( u ) for any u R 1 and write L n = i = 1 n x i x i .
Theorem 2.
In the model (1), suppose that  { e i , i 1 }  is an AANA sequence of identically distributed random variables with mixing coefficients  { q ( n ) , n 1 } , and the following conditions are satisfied:
  • ( A 5 )   Ψ ( u ) Ψ ( u ) Ψ + ( u )   ( u R 1 ) ;
  • ( A 6 )  there exists a finite function
    G ( u ) = E Ψ ( e i + u )
    with  E Ψ ( e i ) = G ( 0 ) = 0  and positive derivative  λ  at  u = 0 .
  • ( A 7 )   0 < E Ψ 2 ( e i ) <  and
    lim u 0 E [ Ψ ( e i + u ) Ψ ( e i ) ] 2 = 0 ;
  • ( A 8 )  there exists a  n 0  such that  L n = i = 1 n x i x i  is a  p -order positive definite matrix as  n n 0  and  d n max 1 i n x i L n 1 x i 0  as  n ;
  • ( A 9 )   C o v ( L n 1 2 i = 1 n Ψ ( e i ) x i ) Σ ;
  • ( A 10 )   n = 1 q ( n ) = h <  as  n ;
Then,
L n 1 / 2 ( β ^ n β 0 ) d N ( 0 , λ 2 Σ ) .
Remark 2.
In Theorem 2, except for the extra condition ( A 10 ), other conditions are the same as that of Theorem 1.1 in Rao and Zhao [25].
As independent and NA sequences are special AANA sequences, we can easily obtain the following corollaries.
Corollary 3
([25]). In the model (1), let { e i , i 1 } be an i . i . d . random sequence. Assume that conditions ( A 5 )–( A 8 ) are satisfied. Then, conclusion (23) holds.
Corollary 4.
In the model (1), let  { e i , i 1 }  be an identically distributed NA random sequence. Assume that conditions ( A 5 )–( A 8 ) are satisfied. Then, conclusion (23) holds.
To prove Theorem 2, we need the following lemmas.
Lemma 7
([26]). Let D be an open convex subset in R p and { f n } be a list of random convex functions on D . Assume that f is a real function on D , such that
f n ( u ) a . s . f ( u )   ( or   convergence   in   probability )
for all u D as n . Then, f is a convex function on D and
sup u D 0 | f n ( u ) f ( u ) | a . s . 0   ( or   convergence   in   probability )
for all compact subsets D 0 of D as n , where a . s . stands for almost sure convergence.
Lemma 8.
In the model (1), suppose that all the conditions of Theorem 2 hold. Then,
sup | γ C | | i = 1 n { ρ ( e i x n i γ ) ρ ( e i ) + Ψ ( e i ) x n i γ } λ γ γ / 2 | p 0
and
sup | γ C | | i = 1 n { Ψ ( e i x n i γ ) Ψ ( e i ) } x n i + λ γ | p 0
for all constant  C > 0 , where  | γ |  represents the maximum absolute value of each component of vector  γ  and  P  stands for convergence in probability.
Proof. 
Denote
x n i = L n 1 / 2 x i   ( 1 i n ) ,
f n ( γ ) = i = 1 n { ρ ( e i x n i γ ) ρ ( e i ) + Ψ ( e i ) x n i γ } = i = 1 n 0 x n i γ { Ψ ( e i + u ) Ψ ( e i ) } d u     ( γ R p ) .
By ( A 8 ) , we have max 1 i n x n i 2 = d n 0 as n . So, for each fixed γ , we obtain that max 1 i n | x n i γ | 0 as n . From ( A 6 ) , there exist a bunch of positive numbers, ε n 0 and θ n i ( 1 , 1 ) , such that:
E f n ( γ ) = i = 1 n 0 x n i γ G ( u ) d u = i = 1 n 0 x n i γ { λ u + o | u | } d u = 1 2 λ i = 1 n ( x n i γ ) 2 ( 1 + ε n θ n i ) 1 2 λ γ γ
as n .
There is no effect on the consequence of f n ( γ ) whether x n i γ is positive or negative. Without loss of generality, we suppose x n i γ 0 . From the condition ( A 10 ), Lemma 4, the Schwarz inequality, ( A 7 ) and max 1 i n | x n i γ | 0 , it follows that:
V a r ( f n ( γ ) ) = E [ ( i = 1 n 0 x n i γ ( Ψ ( e i + u ) Ψ ( e i ) ) d u ) + ( i = 1 n 0 x n i γ ( Ψ ( e i + u ) Ψ ( e i ) ) d u ) ] 2 E [ ( i = 1 n 0 x n i γ ( Ψ ( e i + u ) Ψ ( e i ) ) d u ) + ] 2 + E [ ( i = 1 n 0 x n i γ ( Ψ ( e i + u ) Ψ ( e i ) ) d u ) ] 2 C i = 1 n E [ 0 x n i γ ( Ψ ( e i + u ) Ψ ( e i ) ) d u ] 2 C i = 1 n | x n i γ | | 0 x n i γ E [ Ψ ( e i x n i γ ) Ψ ( e i ) ] 2 d u | = o ( 1 ) i = 1 n ( x n i γ ) 2 0 .
Thus, we have:
f n ( γ ) P λ 2 γ γ   as   n .
Therefore, Lemma 8 follows from Lemma 7. □
Proof of Theorem 2.
Let β n 0 = L n 1 / 2 β 0 , x n i = L n 1 / 2 x i ( 1 i n ) . Then, we can rewrite model (1) as:
Y i = x n i β n 0 + e i ,   ( 1 i n , n 1 ) .
Then, i = 1 n x n i x n i = I p and d n = max 1 i n x n i 2 0 as n . Let β ^ n o be the M-estimator of β n 0 . We can easily obtain that β ^ n 0 = L n 1 / 2 β ^ n , where β ^ n is the M-estimator of β 0 in the model (1). Since β 0 = L n 1 / 2 β n 0 and β ^ n = L n 1 / 2 β ^ n 0 , proving
L n 1 / 2 ( β ^ n β 0 ) d N ( 0 , λ 2 Σ )
shows that
β ^ n 0 β n 0 d N ( 0 , λ 2 Σ ) .
Without loss of generality, we suppose that the true parameter β 0 = 0 in the model (1), then β n 0 = 0 . Hence, to prove (25), we only need to prove:
β ^ n 0 d N ( 0 , λ 2 Σ ) .
Denote β ¯ n 0 = λ 1 i = 1 n Ψ ( e i ) x n i . Let c be any p -dimensional unit column vector, then c β ¯ n 0 = λ 1 i = 1 n ( j = 1 p c j x n i j ) Ψ ( e i ) . Suppose x n i γ 0 , otherwise we can discuss it in positive and negative terms. By Theorem 1, we have:
c β ¯ n 0 d N ( 0 , λ 2 c Σ c ) .
And applying the Cramer–Wold theorem, it follows that:
β ¯ n 0 d N ( 0 , λ 2 Σ ) .
From this, we conclude that { β ¯ n 0 } is bounded in probability. That is, for each ε > 0 , there exists a constant C > 0 , such that:
P { | β ¯ n 0 | > C } < ε / 2
as n n 0 . By Lemma 8, we obtain that for each δ > 0
I ( | β ¯ n 0 C | ) sup γ β ¯ n 0 = δ | i = 1 n { ρ ( Y i x n i γ ) ρ ( Y i x n i β ¯ n 0 ) } 1 2 λ γ β ¯ n 0 2 | P 0
as n , where I ( A ) is the indicative function of set A . By (27) and (28), we have
P { inf γ β ¯ n 0 = δ i = 1 n ρ ( Y i x n i γ ) i = 1 n ρ ( Y i x n i β ¯ n 0 ) + 1 4 λ δ 2 } 1 ε
for sufficiently large n . From this and the convex property of ρ , we obtain
P { β ^ n 0 β ¯ n 0 < δ } 1 ε
for sufficiently large n .
By the arbitrariness of ε and δ , we derive that:
β ^ n 0 β ¯ n 0 P 0
as n . Thus, Theorem 2 follows form (26) and (29).
Then, proving Theorem 2 is complete. □

4. Numerical Simulation

In this section, we will carry out a simulation to verify the asymptotic normality of the M-estimator in a linear regression model with AANA errors.
An AANA sequence is given as follows:
sequence   1 :   e i = ( 1 + a i 2 ) 1 / 2 ( η i + a i η i + 1 ) ,   1 i n ,
where η 1 , η 2 , are i . i . d . N ( 0 , 1 ) random variables and a i = 1 / i 2 . The sequence is an AANA sequence but not an NA sequence (see [1]).
We will simulate a regression model:
Y i = x i β 0 + e i ,
where x i = sin ( 5 i ) + 3 u i ,   1 i n ,   i = 1 , 2 , , n ,   u i is a standard uniform distribution U ( 0 , 1 ) , and the random errors are given by sequence 1.
A derivative function Ψ ( ) of function ρ ( ) is given by:
Ψ ( x ) = { ( 1 ( x / 6 ) 2 ) 2 , 1 ( x / 6 ) 2 > 0 ,   0 , 1 ( x / 6 ) 2 0  
(see [27]). The sample sizes n are taken as n = 100 , 500 , 1000 , and 2000 , respectively. We use R software to compute λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) 1000 times and present the histograms and Quantile–Quantile plots of it in Figure 1, Figure 2, Figure 3 and Figure 4. The red curve in Figure 1 is the kernel density estimate curve, and the red straight line is the reference line whose slope is standard deviation and intercept is mean. The red curves and straight lines in the following figures are also the same meaning.
From Figure 1, Figure 2, Figure 3 and Figure 4, we can see that the histograms and Q-Q plots show good fit of the distribution for λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) to standard normal distribution as n increases. The simulation results show the asymptotic normality of the M-estimator in the linear regression model with AANA errors.

5. Conclusions

In this article, we mainly study the asymptotic properties of the estimators in the model (1). Many scholars have obtained the asymptotic properties of the estimators in linear models whose errors are independent (see [9,25]). However, the errors are not independent in many applications. In this paper, we suppose that the errors are AANA random variables, which include independent and NA random variables as special cases. The asymptotic normality of M-estimators for the unknown parameters is investigated under some suitable conditions, and the central limit theorem for AANA sequences of random variables is also derived. The results extend the corresponding ones of independent and NA random variables (see [19]). In addition, for model (1), a simulation is carried out to investigate the numerical performance. The aim of our work was to construct general mixing conditions under the classical setting. Although it might be interesting future work, we do not address sequences that have long-range dependence (see [28,29,30]) herein, where the limiting behavior is no longer classical. This is an interesting subject to investigate the limit properties of the estimators in regression models with long-range dependence errors in future studies.

Funding

This work was supported by the Scientific Research Project of Hubei Provincial Department of Education (No. Q20233003) and the Foundation for Innovative Research Team of Hubei Provincial Department of Education (No. T2022034).

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Chandra, T.K.; Ghosal, S. Extensions of the strong law of number of Marcinkiewicz and Zygmund for dependent variables. Acta Math. Hung. 1996, 71, 327–336. [Google Scholar] [CrossRef]
  2. Chandra, T.K.; Ghosal, S. The strong law of number for weighted averages under dependence assumptions. J. Theor. Probab. 1996, 19, 797–809. [Google Scholar] [CrossRef]
  3. Yuan, D.M.; An, J. Rosenthal type inequalities for asymptotically almost negatively associated random variables and applications. Sci. China Ser. A 2009, 52, 1887–1904. [Google Scholar] [CrossRef]
  4. Hu, X.P.; Fang, G.H.; Zhu, D.J. Strong convergence properties for asymptotically almost negatively associated sequence. Discret. Dyn. Nat. Soc. 2012, 2012, 562838. [Google Scholar] [CrossRef]
  5. Tang, X.F. Some strong laws of large numbers for weighted sums of asymptotically almost negatively associated random variables. J. Inequalities Appl. 2013, 2013, 4. [Google Scholar] [CrossRef]
  6. Zhang, Y.; Liu, X.S.; Hu, H.C. Weak consistency of M-estimator in linear regression model with asymptotically almost negatively associated errors. Commun. Stat.-Theory Methods 2020, 49, 2800–2816. [Google Scholar] [CrossRef]
  7. Huber, P.J. Robust estimation of a location parameter. Ann. Math. Statist. 1964, 35, 73–101. [Google Scholar] [CrossRef]
  8. Huber, P.J. Robust regression: Asymptotics, conjectures and Monte Carlo. Ann. Stat. 1973, 1, 799–821. [Google Scholar] [CrossRef]
  9. Chen, X.R.; Zhao, L.C. M-Methods in Linear Models; Shanghai Scientific and Technical Publishers: Shanghai, China, 1996. [Google Scholar]
  10. Huber, P.J.; Ronchetti, E.M. Robust Statistics, 2nd ed.; John Wiley & Sons: New Jersey, NJ, USA, 2009. [Google Scholar]
  11. Hu, H.C.; Zhang, Y.; Pan, X. Asymptotic normality of DHD estimators in a partially linear model. Stat. Pap. 2016, 57, 567–587. [Google Scholar] [CrossRef]
  12. Goryainov, A.V.; Goryainov, V.B. M-estimates of autoregression with random coefficients. Autom. Remote Control 2018, 79, 1409–1421. [Google Scholar] [CrossRef]
  13. Linke, Y.Y.; Sakhanenko, A.I. Conditions of asymptotic normality of one-step M-estimators. J. Math. Sci. 2018, 230, 95–111. [Google Scholar] [CrossRef]
  14. Wu, Q.Y. Strong consistency of M estimator in linear model for negatively associated samples. J. Syst. Sci. Complex. 2006, 19, 592–600. [Google Scholar] [CrossRef]
  15. Fan, J. Moderate deviations for M-estimators in linear model with-mixing errors. Acta Math. Sin. 2012, 28, 1275–1294. [Google Scholar] [CrossRef]
  16. Cui, H.J.; He, X.M.; Kai, W.N. M-estimation for linear models with spatially-correlated errors. Stat. Probab. Lett. 2004, 66, 383–393. [Google Scholar] [CrossRef]
  17. Muhammad, S.; Sohail, C.; Babar, I. On Some new ridge M-estimators for linear regression models under various error distributions. Pak. J. Stat. 2021, 37, 369–391. [Google Scholar]
  18. Muhammad, S.; Sohail, C.; Muhammad, A. New quantile based ridge M-estimator for linear regression models with multicollinearity and outliers. Commun. Stat. Simul. Comput. 2023, 52, 1418–1435. [Google Scholar]
  19. Su, C.; Chi, X. Some results on CLT for nonstationary NA sequences. Acta Math. Appl. Sin. 1998, 21, 9–21. [Google Scholar]
  20. Butzer, P.L.; Hahn, L. General theorems on rates of convergence in distribution of random variables. 1: General Limit Theorem. J. Multivar. Anal. 1978, 8, 181–201. [Google Scholar] [CrossRef]
  21. Yuan, D.M.; An, J. Laws of large numbers for Ces’aro alpha-integrable random variables under dependence condition AANA or AQSI. Acta Math. Sin. 2012, 28, 1103–1118. [Google Scholar] [CrossRef]
  22. Zhang, L.X. A functional central limit theorem for asymptotically negatively dependent random fields. Acta. Math. Hung. 2000, 86, 237–259. [Google Scholar] [CrossRef]
  23. Ledoux, M.; Talagrand, M. Probability in Banach Space; Springer: Berlin, Germany, 1991. [Google Scholar]
  24. Petrov, V.V. Limit Theorem for the Sum of Independent Random Variables; Science Publishers: Beijing, China, 1987. [Google Scholar]
  25. Rao, C.R.; Zhao, L.C. Approximation to the distribution of M-estimates in linear models by randomly weighted bootstrap. Sankhy Ser. A 1992, 54, 323–331. [Google Scholar]
  26. Anderson, P.K.; Gill, R.D. Cox’s regression model for counting processes: A large sample study. Ann. Stat. 1982, 15, 1100–1120. [Google Scholar] [CrossRef]
  27. Huber, P.J. Finite sample breakdown of M-and P-estimators. Ann. Stat. 1984, 12, 119–126. [Google Scholar] [CrossRef]
  28. Düker, M.C. Limit theorems in the context of multivariate long-range dependence. Stoch. Process. Appl. 2020, 130, 5394–5425. [Google Scholar] [CrossRef]
  29. Kouritzin, M.A.; Sounak, P. On almost sure limit theorems for heavy-tailed products of long-range dependent linear processes. Stoch. Process. Appl. 2022, 152, 208–232. [Google Scholar]
  30. Bai, S.Y.; Taqqu, M.S. Multivariate limit theorems in the context of long range dependence. J. Time Ser. Anal. 2013, 34, 717–743. [Google Scholar] [CrossRef]
Figure 1. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 100 .
Figure 1. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 100 .
Mathematics 11 03858 g001
Figure 2. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 500 .
Figure 2. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 500 .
Mathematics 11 03858 g002
Figure 3. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 1000 .
Figure 3. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 1000 .
Mathematics 11 03858 g003
Figure 4. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 2000 .
Figure 4. Histogram and normal Q-Q plot of λ 1 / 2 L n 1 / 2 ( β ^ n β 0 ) μ with n = 2000 .
Mathematics 11 03858 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y. Asymptotic Normality of M-Estimator in Linear Regression Model with Asymptotically Almost Negatively Associated Errors. Mathematics 2023, 11, 3858. https://doi.org/10.3390/math11183858

AMA Style

Zhang Y. Asymptotic Normality of M-Estimator in Linear Regression Model with Asymptotically Almost Negatively Associated Errors. Mathematics. 2023; 11(18):3858. https://doi.org/10.3390/math11183858

Chicago/Turabian Style

Zhang, Yu. 2023. "Asymptotic Normality of M-Estimator in Linear Regression Model with Asymptotically Almost Negatively Associated Errors" Mathematics 11, no. 18: 3858. https://doi.org/10.3390/math11183858

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop