Next Article in Journal
New Results Achieved for Fractional Differential Equations with Riemann–Liouville Derivatives of Nonlinear Variable Order
Next Article in Special Issue
Analysis of Correlation Bounds for Uniformly Expanding Maps on [0, 1]
Previous Article in Journal
Minor and Major Strain: Equations of Equilibrium of a Plane Domain with an Angular Cutout in the Boundary
Previous Article in Special Issue
Moderate Deviation Principle for Linear Processes Generated by Dependent Sequences under Sub-Linear Expectation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Least Squares Estimator for Gradual Change-Point in Time Series with m-Asymptotically Almost Negatively Associated Errors

1
School of Mathematical Science, Huaibei Normal University, Huaibei 235099, China
2
Department of Applied Mathematics, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Axioms 2023, 12(9), 894; https://doi.org/10.3390/axioms12090894
Submission received: 26 August 2023 / Revised: 14 September 2023 / Accepted: 15 September 2023 / Published: 20 September 2023
(This article belongs to the Special Issue Probability, Statistics and Estimation)

Abstract

:
As a new member of the NA (negative associated) family, the m-AANA (m-asymptotically almost negatively associated) sequence has many statistical properties that have not been developed. This paper mainly studies its properties in the gradual change point model. Firstly, we propose a least squares type change point estimator, then derive the convergence rates and consistency of the estimator, and provide the limit distributions of the estimator. It is interesting that the convergence rates of the estimator are the same as that of the change point estimator for independent identically distributed observations. Finally, the effectiveness of the estimator in limited samples can be verified through several sets of simulation experiments and an actual hydrological example.

1. Introduction

Change point problems originally arose from quality control engineering [1]. Because of the heterogeneities in real data sequences, the problems of change point estimation and detection in real data sequences have drawn attention. Scholars have proposed many methods to solve various data problems (see [2,3,4,5,6,7,8,9]). The purpose of change point detection and estimation is to divide a data sequence into several homogeneous segments, and the theories have been applied in many fields like finance [10], medicine [11], environment [12] and so on. For some special problems, such as hydrological and meteorological problems, most of the change point that occur are gradual rather than abrupt. Therefore, the research on the problems of gradual change is very meaningful.
In earlier research, most theories of gradual changes were derived from two-stage regression model [13,14]. Hušková [15] used the least–squares method to estimate an unknown gradual change point and the model is as follows:
For n > 1 , observations X 1 , X 2 , ⋯, X n shall satisfy:
X t = μ + δ i m n + + e i , 1 i n ,
where ( a ) + = max ( 0 , a ) , μ , δ 0 , m is the location of the change, e 1 , e 2 , , e n are i.i.d. random variables with E e i = 0 , v a r ( e i ) = σ 2 , E | e i | 2 + Δ < for some Δ > 0 . In the same year, Jarušková [16] conducted a log likelihood ratio test of the model on the basis of Hušková [15], and obtained that the asymptotic distribution of the test statistic is Gumbel distribution.
Later, Wang [17] extended the error terms e i from the traditional i.i.d. sequences to the long memory i.i.d sequences and obtained the consistency of the estimator of the mean gradual change point and the limit distribution of the test statistic. Then, Timmermann [18] tested the gradual change in general random processes and update processes respectively, and also obtained the limit distribution of test statistic. It is well known that most of the previous studies tend to focus on the abrupt change. But for some important time series, such as temperature and hydrology, there is a greater possibility of gradual mean change.
Therefore, this paper considers the gradual change problem and the following model based on Hušková [15] is constructed:
X t = μ + δ n t k * n + γ + Y t , 1 t n ,
where n > 1 , X 1 , X 2 , ⋯, X n are observations. ( a ) + = max ( 0 , a ) , γ ( 0 , 1 ) , and k * is unknow change point location. μ , δ n are unknown parameter, μ , δ n 0 with n 1 / 2 | δ n | ( log n ) 1 / 2 . Y 1 , Y 2 , , Y n are random variables with zero mean.
The least squares method is a classic method (see [19,20]). In model (1), the estimators of k * and τ * based on the least–squares–type are, respectively:
k ^ * = min k : k = arg max j U j ( γ ) ; j = 1 , 2 , , n 1 ,
τ ^ * = k ^ * / n ,
where:
U j ( γ ) = t = 1 n ( x t j x ¯ j ) X t t = 1 n ( x t j x ¯ j ) 2 1 / 2 .
For convenience of illustration, x t j and x ¯ j are t j n + γ and 1 n t = 1 n x t j , respectively.
It can be seen that most previous studies have one thing in common, namely that the error terms are independently and identically distributed. However, the constraints of independent sequences are quite strict, and in practical problems, many time series models may not meet the independent conditions. This leads to a classical question of whether the error terms in a model can be generalized to some more general cases. Therefore, the idea of extending the error terms to the m-AANA sequences is proposed in this paper. Before presenting the main asymptotic results, it is necessary to understand the following definition:
Definition 1 ([21]).
There is an fixed integer m 1 , the random variable sequence { Y n , n 1 } is called m-AANA sequence if there exists a non negative sequence q ( n ) 0 as n such that:
C o v { f ( Y n ) , g ( Y n + m , , Y n + k ) } q ( n ) [ V a r ( f ( Y n ) ) V a r ( g ( Y n + m , , Y n + k ) ) ] 1 / 2 ,
for all n 1 , k m and for all coordinatewise nondecreasing continuous functions f and g whenever the variances exist. The sequence { q ( n ) , n 1 } is called the mixing coefficients of { Y n , n 1 } . It is not difficult to see that NA family includes NA [22], m-NA [23], AANA [24], and independent sequences.
Scholars have great interest in the sequences of NA family. For example, NA sequence is opposite to PA (positively associated) sequence, but NA sequence has better property than other existing ND (negatively dependent) sequence: under the influence of increasing functions, the disjoint subset of NA random variable sequence is still NA. Therefore, NA sequences appeared in many literary works. Przemysaw [25] obtained the convergence of partial sums of NA sequences; Yang [26] obtained Bernstein type inequalities for NA sequences; and Cai [27] obtained the Marcinkiewicz – Zygmund type strong law of large numbers of NA sequences.
There are many classical studies about m-NA and AANA sequences as generalizations of NA sequences. Hu [23] discussed the complete convergence of m-NA sequences; Yuan [28] proposed a Marcinkiewics-Zygmund type moment inequality for the maximum partial sum of AANA sequences.
For the m-AANA sequence mentioned in this paper, it is a relatively new concept and its research results are less than those of other sequences in NA family. Ko [29] extended Hájek–Rényi inequalities and the strong law of large numbers of Nam [21] to Hilbert space; Ding [30] proposed the CUSUM method to estimate the abrupt change point in a sequence with the error term of m-AANA process, but for a more general case, he did not discuss the gradual change. Therefore, another reason for constructing the main ideas of this paper is is to fill the research gaps of Ding [30].
The rest is arranged as follows: Section 2 describes the main results. A small simulation study under different parameters and an example is provided in Section 3. Section 4 contains the conclusions and outlooks, and the main results are proved in Appendix A.

2. Main Results

For our asymptotic results, assume the model to satisfy the following assumptions:
Assumption 1.
{ Y n , n 1 } is a sequence of m-AANA random variables with E Y n = 0 , E Y n 2 = σ 2 < , and there is a v > 0 , E | Y n | 2 + v < .
Assumption 2.
The mixing coefficient sequence { q ( n ) , n 1 } satisfies n = 1 q ( n ) , n = 1 q 2 ( n ) < .
Assumption 3.
We note S i , n = j = 1 n Y j + i , S n = S 0 , n . As n ,
E S n 2 n = s 2 < .
In addition, there exists strictly ascending sequence of natural numbers { n k } with n 0 = 0 and m i = n i n i 1 . And for some 0 < α 1 , i = 1 m i n i 1 + α / 2 < , lim i E S n i 1 , m i 2 m i = s 2 < .
Remark 1.
Assumption 1 and 2 are the underlying assumptions, and if these assumptions cannot be met, serious problems such as bias in the estimates, inconsistency in the estimates, and invalidity of the estimates may arise in the proof process.
In addition, what can be verified is that the m-AANA random variables satisfy the central limit theorem when Assumption 1–3 are ture.
Let τ = k / n , τ * = k * / n , we obtain the Theorem 1.
Theorem 1.
If Assumption 1 and 2 are ture, when n ,
| τ ^ * τ * | = o p ( 1 ) .
Theorem 2.
If Theorem 1 holds, when n ,
| k ^ * k * | = O p ( ) ,
= ( δ n 2 n 2 γ ) 1 / ( 2 γ + 1 ) , γ ( 0 , 1 / 2 ) , δ n 1 n 1 / 2 ( log ( n k * ) ) 1 / 2 , γ = 1 / 2 , δ n 1 n 1 / 2 , γ ( 1 / 2 , 1 ) .
Remark 2.
Interestingly, these convergence rates are the same as those in Hušková [31], but independent sequences are contained in m-AANA sequences.
Theorem 3.
Case 1. Assuming that Assumption 1–3 hold true. If γ ( 0 , 1 2 ) , as n , then:
( δ n 2 n 2 γ ) 1 / ( 2 γ + 1 ) ( k ^ * k * ) / ( σ 2 C 7 ) D V γ ,
where V γ = arg max Z γ ( i ) ( ( x + i ) + γ x + γ ) 2 d x / 2 ; i R with Z γ ( i ) ; i R is a Gaussian process with zero mean and covariance function.
C o v ( Z γ ( i ) , Z γ ( s ) ) = ( ( x + i ) + γ x + γ ) ( ( x + s ) + γ x + γ ) d x , i , s R .
Case 2. Assuming that Assumption 1–3 hold true. If γ = 1 2 , as n , then:
δ n ( log ( n k * ) ) 1 / 2 2 n 1 / 2 k ^ * k * ( σ 2 C 8 ) 1 / 2 D V 1 / 2
where V 1 / 2 is a standard normal variable.
Case 3. Assuming that Assumption 1–3 hold true. If γ ( 1 2 , 1 ) , as n , then:
δ n n 1 / 2 k ^ * k * ( σ 2 C 9 ) 1 / 2 D V γ
where V γ is a normal N ( 0 , g 2 ( τ * , γ ) ) random variable with:
g 2 ( τ * , γ ) = ( 1 τ * ) ( 2 γ 1 ) 2 γ + 1 4 ( γ 1 + 2 τ * ) 2 γ 2 + τ * ( 1 + 2 γ ) 1 2 γ 1 ( ( γ 1 ) 2 + τ * ( 2 γ 1 ) ) .
C 7 , C 8 , C 9 < σ 2 are constants and are described in the proofs in Appendix A.

3. Simulations and Example

In this section, assume that there is only one gradual change point of mean at k * in (1), such that:
X t = μ + δ n t k * n + γ + Y t , 1 t n ,
Y 1 , Y 2 , , Y n satisfy:
( Y 1 , Y 2 , , Y n ) = d ω 1 N n ( 0 , I n ) + ω 2 N n ( 0 , Σ n ) ,
where ω 1 , ω 2 0 , ω 1 + ω 2 = 1 , I n is a identity matrix, Σ n satisfies:
Σ n = 1 + 1 / n ρ ρ 2 ρ n 1 ρ 1 + 2 / n ρ ρ n 2 ρ 2 ρ 1 + 3 / n ρ n 3 ρ n 1 ρ 2 ρ 2 n × n
where | ρ | < 1 . It can be verified that { Y 1 , Y 2 , , Y n } is a m-AANA sequence with m = 2 and q ( n ) = O p ( | ρ | n ) . For comparison, we take μ = 1 , τ * = 0.5 , ω 1 = ω 2 = 0.5 , ρ = 0.6 and δ n = ( n 0.1 , 1 , n 0.1 , n 0.2 ) , γ = ( 0.75 , 0.5 , 0.25 ) . It is worth noting that after verification, under these conditions, Σ n is a non singular matrix, which can be used for simulation experiments. Then, 1000 simulation processes are carried out. Figure 1, Figure 2 and Figure 3 are based on the simulation of δ n = n 0.1 , 1 , n 0.1 , n 0.2 from left to right.
In Figure 1, Figure 2 and Figure 3, the ordinate axis represents the value of τ ^ * τ * , and the abscissa is the size of sample n. It is not difficult to find that the larger the gradual coefficient γ , the worse the performance of our estimator. And as n increases, the estimation effect becomes better, this also implies the result in Theorem 1. For different ρ and τ * , similar results can be obtained, which will not be repeated here.
Finally, we do the change-point analysis based on the sequence of monthly average water levels of Hulun Lake in China from 1992 to 2008. For the convenience of description, we subtract the median of the observations. Then, Figure 4 and Figure 5 can be obtained:
It can be seen from Figure 5 that we have no sufficient reason to believe that the water levels of Hulun Lake does not meet the conditions of m-AANA sequence. According to Sun [32], the water levels of Hulun Lake have declined rapidly since 2000, due to the influence of the monsoon, changes in precipitation patterns, and the degradation of frozen soil. Using the method in this paper, Table 1 shows the values of k ^ * under different γ .
The position of 105 represents the year 2000. The average water levels of Hulun Lake have been decreasing since 2000.

4. Conclusions

This paper proposes a least-squares-type estimator of the gradual change point of sequence based on m-AANA noise and study the consistency of the estimator. At the same time, the convergence rates are obtained in Theorem 2:
| k ^ * k * | = O p ( ) ,
= ( δ n 2 n 2 γ ) 1 / ( 2 γ + 1 ) , γ ( 0 , 1 / 2 ) , δ n 1 n 1 / 2 ( log ( n k * ) ) 1 / 2 , γ = 1 / 2 , δ n 1 n 1 / 2 , γ ( 1 / 2 , 1 ] .
Therefore, Theorems 1 and 2 generalize the results in Hušková [31]. Furthermore, due to the asymptotic normality of m-AANA sequences, this paper also derives the limit distributions of the estimator under different γ in Throrem 3. It can be known that the inappropriate γ has a great impact on the change point estimator. If γ 0 , the gradual change point in (2) may be very similar to the abrupt change point, and lose the gradual change properties. If γ 1 , the dispersion of data may be very large, which is not conducive to determining the correct change point position. So we conduct several simulations to verify the results, and the results show that the larger γ is, the worse the estimation effect is, but the consistency is still satisfied. Finally, the paper discusses the gradual change of water levels of Hulun Lake in Section 3, and the estimator successfully finds the position of the change point.
There is also some regret in this paper. For example, for series where the variances cannot be estimated, such as a financial heavy-tailed sequence with a heavy-tailed index κ n < 2 , the location of the change point cannot be obtained using the method in this paper. Therefore, more suitable methods should be promoted in future works. Moreover, we suspect that there may be more common cases in the selection of e i , this is also one of the key points to be solved in the future.

Author Contributions

Methodology, T.X.; Software, T.X.; Writing—original draft, T.X.; Supervision, Y.W.; Funding acquisition, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, 61806155; Natural Science Foundation of Anhui Province, 1908085MF186 and Natural Science Foundation of Anhui Higher Education Institutions of China, KJ2020A0024.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Acknowledgments

The authors are very grateful to the editor and anonymous reviewers for their comments, which enables the authors to greatly improve this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof of Theorem 1.
Let V k ( γ ) = U k 2 ( γ ) U k * 2 ( γ ) , (2) can be equivalent to:
k ^ * = min 1 < k < n k : k = arg max j V j ( γ ) ; j = 1 , 2 , , n 1 ,
after the elementary operation, V k ( γ ) can be divided into five parts:
V k ( γ ) = A k , 1 + A k , 2 + A k , 3 + A k , 4 + A k , 5 ,
where:
A k , 1 = 1 a k k 1 a k * k * t = 1 n ( x t k * x ¯ k * ) Y t 2 ,
A k , 2 = 1 a k k t = 1 n ( x t k x ¯ k ) ( x t k * x ¯ k * ) Y t · t = 1 n ( x t k x ¯ k ) + ( x t k * x ¯ k * ) Y t ,
A k , 3 = 2 δ n a k k * a k k 1 t = 1 n ( x t k x ¯ k ) Y t ,
A k , 4 = 2 δ n t = 1 n ( x t k x ¯ k ) ( x t k * x ¯ k * ) Y t ,
A k , 5 = δ n 2 a k k * 2 a k k a k * k * ,
and a k l = t = 1 n ( x t k x ¯ k ) ( x t l x ¯ l ) , k , l = 1 , 2 , 3 , , n 1 . At this moment, a lemma is needed. □
Lemma A1.
If Assumption 1 and 2 are true, when n , then,
max t = 1 n ( x t k x ¯ k ) Y t a k k 1 / 2 = O p ( log n ) .
Proof of Lemma A1.
max t = 1 n ( x t k x ¯ k ) Y t a k k 1 / 2 = max 1 a k k 1 / 2 t = 1 n x t k Y t 1 a k k 1 / 2 1 n t = 1 n x t k t = 1 n Y t = max 1 n γ a k k 1 / 2 t = k + 1 n ( t k ) γ Y t 1 n γ + 1 a k k 1 / 2 t = k + 1 n ( t k ) γ t = 1 n Y t max 1 n γ a k k 1 / 2 t = k + 1 n ( t k ) γ Y t + max 1 n γ + 1 a k k 1 / 2 t = k + 1 n ( t k ) γ t = 1 n Y t .
Then, another result can be obtained:
a k k n = 1 n t = 1 n ( x t k x ¯ k ) 2 = 1 n t = 1 n x t k 2 x ¯ k 2 = k / n 1 x k n 2 γ d x k / n 1 x k n γ d x 2 + O 1 n n k n γ = 1 2 γ + 1 1 k n 2 γ + 1 1 ( γ + 1 ) 2 1 k n 2 γ + 2 + O 1 n
From Lemma 3.1 in Ko [29] and the proof of Theorem 1 in Ding [30], the following truth can be obtained:
P sup 1 k n 1 1 n k t = k + 1 n Y t > ϵ C ϵ 2 t = 1 n E | Y t | 2 t C 1 ϵ 2 log n .
where C , C 1 are positive constants. Then,
sup 1 k n 1 1 n k t = k + 1 n Y t = O p ( log n ) ,
hence,
t = k + 1 n ( t k ) + γ Y t = t = k + 1 n j = 1 t k j 1 j γ x γ 1 d x Y t = j = 1 n k j 1 j γ x γ 1 d x t = j + k n Y t = O p ( | n k | γ + 1 / 2 log n ) ,
combining Equations (A4)–(A8), the following can be obtained:
max t = 1 n ( x t k x ¯ k ) Y t a k k 1 / 2 C 2 max 1 ( n k ) γ + 1 / 2 t = k + 1 n ( t k ) γ Y t + C 3 max ( n k ) γ + 1 n ( n k ) γ + 1 / 2 t = 1 n Y t = O p ( log n ) .
Similar to the proof of Equation (A5), the following can be obtained:
a k * k * n = 1 2 γ + 1 1 τ * 2 γ + 1 1 ( γ + 1 ) 2 1 τ * 2 γ + 2 + O 1 n ,
a k k * n = 0 1 ( x τ ) + γ ( x τ * ) + γ d x 1 ( γ + 1 ) 2 ( 1 τ ) γ + 1 ( 1 τ * ) γ + 1 + O 1 n ,
from Lemma 2.2 in Hušková [31], uniformly for | k k * | = o ( n ) , as n ,
a k k * a k k n = k k * n γ 1 + 2 τ 2 γ + 2 ( 1 τ ) 2 γ ( 1 + o ( 1 ) ) = O ( n 1 | k k * | ) ,
then,
A k , 5 n = δ n 2 n a k k * 2 a k k a k * k * = δ n 2 ( h ( τ ) h ( τ * ) ) ( 1 + o ( 1 ) ) ,
where:
h ( τ ) = 0 1 ( x τ ) + γ ( x τ * ) + γ d x ( γ + 1 ) 2 ( 1 τ ) γ + 1 ( 1 τ * ) γ + 1 2 ( 2 γ + 1 ) 1 1 τ 2 γ + 1 ( γ + 1 ) 2 1 τ 2 γ + 2 .
Now suppose 0 < ε < min ( τ , 1 τ ) and τ [ 0 , 1 ) . Obviously, max τ [ 0 , 1 ) h ( τ ) = h ( τ * ) , max | k k * | > n ε h ( τ ) < h ( τ * ) . Therefore,
max | k k * | > n ε A k , 5 n = max | τ τ * | > ε δ n 2 ( h ( τ ) h ( τ * ) ) ( 1 + o ( 1 ) ) < 0 . a . s .
By Lemma A1, the following can be obtained:
max | k k * | > n ε | A k , 1 | n = max | k k * | > n ε 1 n a k * k * a k k a k k t = 1 n ( x t k * x ¯ k * ) Y t a k * k * 1 / 2 2 = O p ( n 1 log n ) ,
by the same token,
max | k k * | > n ε | A k , 2 | n O p ( n 1 log n ) ,
max | k k * | > n ε | A k , 3 | n O p ( n 1 / 2 log n ) ,
max | k k * | > n ε | A k , 4 | n O p ( n 1 / 2 log n ) .
By combining (A13)–(A17), it can be found that when | k k * | > n ε , n ,
max | k k * | > n ε 1 n ( A k , 1 + A k , 2 + A k , 3 + A k , 4 + A k , 5 ) < 0 ,
therefore, only when | k k * | < n ε , V k ( γ ) take the maximum value. Theorem 1 has been proved.
Proof of Theorem 2.
Consider A k , 5 first.
A k , 5 = δ n 2 a k k * 2 a k k a k * k * = δ n 2 ( a k k * a k k ) 2 a k k ( a k k + a k * k * 2 a k k * ) .
According to the Lemmas 2.2–2.4 in Hušková [31], for any D > 0 , D n , and D n / n 0 as n . Then, the following can be accessed:
a k * k * + a k k 2 a k k * = O n 2 γ | k k * | 2 γ + 1 , γ ( 0 , 1 / 2 ) , O n 1 | k k * | 2 log ( n k * ) , γ = 1 / 2 , O n 1 | k k * | 2 , γ ( 1 / 2 , 1 ) ,
uniformly for D n > | k k * | > D . The problem can be considered in three parts. When γ ( 0 , 1 / 2 ) , combine (A5), (A11), (A19) and (A20), it is obvious that C 4 , C 5 > 0 ,
C 4 δ n 2 n 2 γ | k k * | 2 γ + 1 A k , 5 C 5 δ n 2 n 2 γ | k k * | 2 γ + 1 ,
uniformly for D n > | k k * | > D .
Next, for A k , 1 ,
max D > | k k * | > M | A k , 1 | δ n 2 n 2 γ | k k * | 2 γ + 1 = max D > | k k * | > M δ n 2 n 2 γ | k k * | 2 γ 1 O ( n 1 | k k * | ) O p ( log n ) = O p | k k * | n 2 γ log n n δ n 2 .
When n ,
max D > | k k * | > M | A k , 1 | δ n 2 n 2 γ | k k * | 2 γ + 1 = o p ( 1 ) ,
similarly,
max D > | k k * | > M | A k , 2 | δ n 2 n 2 γ | k k * | 2 γ + 1 = o p ( 1 ) ,
max D > | k k * | > M | A k , 3 | δ n 2 n 2 γ | k k * | 2 γ + 1 = o p ( 1 ) ,
max D > | k k * | > M | A k , 4 | δ n 2 n 2 γ | k k * | 2 γ + 1 = o p ( 1 ) .
Combining (A21) and (A23)–(A26), when γ ( 0 , 1 / 2 ) , | k ^ * k * | = O p ( ) . For the cases of γ = 1 / 2 and γ ( 1 / 2 , 1 ) , the methods of proof are similar, so the article will not repeat. □
In order to obtain the asymptotic distributions of the estimator, it is necessary to prove several lemmas about random terms.
Lemma A2.
Assuming that Assumption 1 and 2 are true, if γ ( 0 , 1 ) , as n , then,
max 1 k n 1 t = 1 n ( x t k x ¯ k ) Y t t = 1 n ( x t k x ¯ k ) 2 1 / 2 = O p ( ( log n ) 1 / 2 ) ,
max 1 k n ( 1 ε ) t = 1 n ( x t k x ¯ k ) Y t t = 1 n ( x t k x ¯ k ) 2 1 / 2 = o p ( 1 )
for any ε ( 0 , 1 ) .
Proof of Lemma A2.
Prove that (A27) holds, which is equivalent to proving:
max 1 k n 1 t = k + 1 n ( t k ) + γ Y t ( n k ) γ + 1 / 2 = O p ( ( log n ) 1 / 2 ) ,
by (A7), The following are available:
t = k + 1 n ( t k ) + γ Y t = O p ( | n k | γ + 1 / 2 ( log n ) 1 / 2 ) ,
uniformly for 1 k n 1 . And for the proof of (A28), just replace (A7) with max 1 k n 1 t = k + 1 n Y t = O p ( n 1 / 2 ) . □
Then, it is necessary to define that:
Q n γ ( k ) = 1 δ n t = 1 n ( x t k x t k * ) Y t t = 1 n ( x t k x t k * ) 2 , k k * , γ ( 0 , 1 )
and
V n γ k k * = n min γ , 1 / 2 + 1 / 2 1 n 1 / 2 t = 1 n ( x t k x t k * ) Y t , γ 1 2 ,
V n γ k k * = 1 n 1 / 2 n 1 ( log ( n k * ) ) 1 / 2 t = 1 n ( x t k x t k * ) Y t , γ = 1 2 .
It is not difficult to see that:
E Q n γ ( k ) = 0 , k k
and at this point, another lemma is required:
Lemma A3 ([21]).
Let { Y n } be an m-AANA random variables sequence with E Y n = 0 . It has mixing coefficients { q ( n ) , n 1 } . If t = 1 q 2 ( n ) < , there exists a positive constant C p such that:
E max 1 k n t = 1 k Y t p C p m p 1 t = 1 n E | Y t | p ,
where 1 < p 2 .
After simple calculation, the following can be obtained:
V a r ( Q n γ ( k ) ) = ( σ 2 C 6 ) δ n 2 t = 1 n ( x t k x t k * ) 2 1 , k k *
where C 1 is a constant. And similarly, for i R ,
E V n γ ( i ) = 0 ,
V a r ( V n γ ( i ) ) = ( σ 2 C 7 ) | i | 2 γ + 1 ( ( x + 1 ) + γ x + γ ) 2 d x ( 1 + o ( 1 ) ) , γ ( 0 , 1 / 2 ) , ( σ 2 C 8 ) i 2 4 ( 1 + o ( 1 ) ) , γ = 1 / 2 , ( σ 2 C 9 ) i 2 γ 2 ( 1 τ * ) 2 γ 1 2 γ 1 ( 1 + o ( 1 ) ) , γ ( 1 / 2 , 1 ) ,
where C 2 , C 3 and C 4 are constants. Next, we need to break down the discussion into situations.
Lemma A4.
If γ ( 0 , 1 / 2 ) , Assumption 1–3 are true. For any η ( 0 , 1 ) , there exist H η > 0 and n η such that for n > n η ,
P max D n > | k * k | > H η | Q n γ ( k ) | η < η ,
where D n is given in the proof of Theorem 2, and as n ,
V n γ , D D ( σ 2 C 7 ) 1 / 2 Z γ , D ,
where V n γ , D = { V n γ ( i ) , i ( D , D ) } and Z γ , D = { Z γ ( i ) , i ( D , D ) } .
Proof of Lemma A4.
By the proof of Lemma 2.3 in [33], for any H > 0 ,
t = 1 n ( x t k x t l ) 2 M 1 | k l | 2 γ 1 n 2 γ
if k * D n < k < l k * H or k * + H k < l < k * + D n with some M 1 > 0 . Then, for k * D n < k < l k * H ,
E ( Q n γ ( k ) Q n γ ( l ) ) 2 M 2 n 2 γ δ n 2 | k l | 2 γ 1 ( k * k ) 4 γ + 2 + ( k * l ) 2 γ + 1 1 ( k * l ) 2 γ + 1 1 ( k * k ) 2 γ + 1 2 M 3 n 2 γ δ n 2 k * k k * l x 2 d x 2 γ + 1 + k * k k * l x γ 3 / 2 d x 2
with some M 2 , M 3 > 0 . According to Theorem 12.2 of [1], for all H , η > 0 , there exists M 4 > 0 , then:
P max D n > | k * k | > H η | Q n γ ( k ) | η M 4 n 2 γ δ n 2 H k * x 2 d x 2 γ + 1 + H k * x γ 3 / 2 d x 2 M 4 η 2 H 2 γ + 1 .
(A34) be obtained by choosing a sufficiently small H.
By the holding of Assumption 3 and Theorem 12.3 in [1] to prove the convergence of (A35). It suffices to show that j = 1 p b j V n γ ( i j ) V a r j = 1 p b j V n γ ( i j ) 1 / 2 converges in distribution to the N ( 0 , 1 ) distributed random variables with any D < i 1 < < i p < D and b 1 , , b p R and the tightness. Since V n γ ( i ) are the sum of m-AANA random variables, the first expected property can be obtained by applying the central limit theorem. By the tightness conditions from Theorem 12.3 in [33], for D < i 1 < i 2 < D , after a simple calculation,
E ( V n γ ( i 1 ) V n γ ( i 2 ) ) 2 M 5 ( i 2 i 1 ) 2 γ + 1 ( 1 + o ( 1 ) )
with some M 5 > 0 . □
Lemma A5.
If γ = 1 / 2 , Assumption 1–3 are true. For any η ( 0 , 1 ) , there exist H η > 0 and n η such that for n > n η ,
P max D n > | k * k | > H η | Q n γ ( k ) | η < η ,
where D n is given in the proof of Theorem 2, and as n ,
V n γ , D D { i Y , i ( D , D ) } ,
where Y is a normal N ( 0 , ( σ 2 C 8 ) / 4 ) random variable.
Proof of Lemma A5.
The proof of (A40) is very similar to (A34) and is omitted here. To prove the tightness, then,
E ( V n γ ( i 1 ) V n γ ( i 2 ) ) 2 = σ 2 C 8 2 ( log ( m k * ) ) 2 j = 1 k * n k * j i 1 + 1 / 2 j i 2 + 1 / 2 2 ( 1 + o ( 1 ) ) = ( σ 2 C 8 ) | i 1 i 2 | 2 2 ( log ( m k * ) ) 2 j = 1 k * n k * j i 1 + 1 / 2 j i 2 + 1 / 2 2 ( 1 + o ( 1 ) ) = ( σ 2 C 8 ) | i 1 i 2 | 2 4 ( 1 + o ( 1 ) )
and
E ( V n γ ( i 1 ) V n γ ( i 2 ) ) = ( σ 2 C 8 ) i 1 i 2 4 ( 1 + o ( 1 ) ) , i 1 , i 2 R .
Therefore, the limit distribution is Gaussian with zero mean and the covariance function v γ ( i 1 , i 2 ) = ( σ 2 C 8 ) i 1 i 2 4 ( 1 + o ( 1 ) ) , which implies (A41). □
Lemma A6.
If γ ( 1 / 2 , 1 ) , Assumption 1–3 are true. For any η ( 0 , 1 ) , there exist H η > 0 and n η such that for n > n η ,
P max D n > | k * k | > H η | Q n γ ( k ) | η < η ,
where D n is given in the proof of Theorem 2, and as n ,
V n γ , D D { i Y , i ( D , D ) } ,
where Y is a normal N ( 0 , γ 2 ( σ 2 C 9 ) ( 1 τ * ) 2 γ 1 2 γ 1 ) random variable.
Proof of Lemma A6.
The proof of (A44) is very similar to (A34) and is omitted here. To prove the tightness, then,
E ( V n γ ( i 1 ) V n γ ( i 2 ) ) 2 = ( σ 2 C 9 ) n 2 2 1 n j = 1 k * n k * j i 1 n + γ j i 2 n + γ 2 ( 1 + o ( 1 ) ) = ( σ 2 C 9 ) ( i 1 i 2 ) 2 γ 2 ( 1 τ * ) 2 γ 1 2 γ 1 ( 1 + o ( 1 ) )
and
E ( V n γ ( i 1 ) V n γ ( i 2 ) ) = ( σ 2 C 9 ) i 1 i 2 γ 2 ( 1 τ * ) 2 γ 1 2 γ 1 ( 1 + o ( 1 ) ) , i 1 , i 2 R .
Therefore, the limit distribution is Gaussian with zero mean and the covariance function v γ ( i 1 , i 2 ) = ( σ 2 C 9 ) i 1 i 2 ( γ 2 ( 1 τ * ) 2 γ 1 ) / 2 γ 1 , which implies (A45). □
Proof of Theorem 3.
The proofs of cases 1 and 2 and 3 in Theorem 3 are similar, so we only prove case 1 here.
To derive the limit distribution of 1 ( k ^ * k * ) , we combine (A9)–(A11), for k = k * + i ,
A k , 1 + A k , 2 + A k , 3 + A k , 4 + A k , 5 = 2 V n γ k k * | i | 2 γ + 1 ( ( x + 1 ) + γ x + γ ) 2 d x + o p ( 1 ) ,
(A48) links with (A35) can imply the result in case 1 of Theorem 3. □

References

  1. Page, E.S. Continuous inspection schemes. Biometrika 1954, 41, 100–115. [Google Scholar] [CrossRef]
  2. Csörgo, M.; Horváth, L. Limit Theorems in Change-Point Analysis; Wiley: Chichester, UK, 1997. [Google Scholar]
  3. Horváth, L.; Hušková, M. Change-point detection in panel data. J. Time Ser. Anal. 2012, 33, 631–648. [Google Scholar] [CrossRef]
  4. Horváth, L.; Rice, G. Extensions of some classical methods in change point analysis. Test. 2014, 23, 219–255. [Google Scholar] [CrossRef]
  5. Xu, M.; Wu, Y.; Jin, B. Detection of a change-point in variance by a weighted sum of powers of variances test. J. Appl. Stat. 2019, 46, 664–679. [Google Scholar] [CrossRef]
  6. Gao, M.; Ding, S.S.; Wu, S.P.; Yang, W.Z. The asymptotic distribution of CUSUM estimator based on α-mixing sequences. Commun. Stat. Simul. Comput. 2020, 51, 6101–6113. [Google Scholar] [CrossRef]
  7. Jin, H.; Wang, A.M.; Zhang, S.; Liu, J. Subsampling ratio tests for structural changes in time series with heavy-tailed AR(p) errors. Commun. Stat. Simul. Comput. 2022, 2022, 2111584. [Google Scholar] [CrossRef]
  8. Hušková, M.; Prášková, Z.; Steinebach, G.J. Estimating a gradual parameter change in an AR(1)–process. Metrika 2022, 85, 771–808. [Google Scholar] [CrossRef]
  9. Tian, W.Z.; Pang, L.Y.; Tian, C.L.; Ning, W. Change Point Analysis for Kumaraswamy Distribution. Mathematics 2023, 11, 553. [Google Scholar] [CrossRef]
  10. Pepelyshev, A.; Polunchenko, A. Real-time financial surveillance via quickest change-point detection methods. Stat. Interface 2017, 10, 93–106. [Google Scholar] [CrossRef]
  11. Ghosh, P.; Vaida, F. Random change point modelling of HIV immunologic responses. Stat. Med. 2007, 26, 2074–2087. [Google Scholar] [CrossRef]
  12. Punt, A.E.; Szuwalski, C.S.; Stockhausen, W. An evaluation of stock-recruitment proxies and environmental change points for implementing the US Sustainable Fisheries Act. Fish Res. 2014, 157, 28–40. [Google Scholar] [CrossRef]
  13. Hinkley, D. Inference in two-phase regression. J. Am. Stat. Assoc. 1971, 66, 736–763. [Google Scholar] [CrossRef]
  14. Feder, P.I. On asymptotic distribution theory in segmented regression problems. Ann. Stat. 1975, 3, 49–83. [Google Scholar] [CrossRef]
  15. Hušková, M. Estimators in the location model with gradual changes. Comment. Math. Univ. Carolin. 1998, 1, 147–157. [Google Scholar]
  16. Jarušková, D. Testing appearance of linear trend. J. Stat. Plan. Infer. 1998, 70, 263–276. [Google Scholar] [CrossRef]
  17. Wang, L. Gradual changes in long memory processes with applications. Statistics 2007, 41, 221–240. [Google Scholar] [CrossRef]
  18. Timmermann, H. Monitoring Procedures for Detecting Gradual Changes. Ph.D. Thesis, Universität zu Köln, Köln, Germany, 2014. [Google Scholar]
  19. Schimmack, M.; Mercorelli, P. Contemporary sinusoidal disturbance detection and nano parameters identification using data scaling based on Recursive Least Squares algorithms. In Proceedings of the 2014 International Conference on Control, Decision and Information Technologies (CoDIT), Metz, France, 3–5 November 2014; pp. 510–515. [Google Scholar]
  20. Schimmack, M.; Mercorelli, P. An Adaptive Derivative Estimator for Fault-Detection Using a Dynamic System with a Suboptimal Parameter. Algorithms 2019, 12, 101. [Google Scholar] [CrossRef]
  21. Nam, T.; Hu, T.; Volodin, A. Maximal inequalities and strong law of large numbers for sequences of m-asymptotically almost negatively associated random variables. Commun. Stat. Theory Methods 2017, 46, 2696–2707. [Google Scholar] [CrossRef]
  22. Block, H.W.; Savits, T.H.; Shaked, M. Some concepts of negative dependence. Ann. Probab. 1982, 10, 765–772. [Google Scholar] [CrossRef]
  23. Hu, T.C.; Chiang, C.Y.; Taylor, R.L. On complete convergence for arrays of rowwise m-negatively associated random variables. Nonlinear Anal. 2009, 71, e1075–e1081. [Google Scholar] [CrossRef]
  24. Chandra, T.K.; Ghosal, S. The strong law of large numbers for weighted averages under dependence assumptions. J. Theor. Probab. 1996, 9, 797–809. [Google Scholar] [CrossRef]
  25. Przemysaw, M. A note on the almost sure convergence of sums of negatively dependent random variables. Stat. Probab. Lett. 1992, 15, 209–213. [Google Scholar]
  26. Yang, S.C. Uniformly asymptotic normality of the regression weighted estimator for negatively associated samples. Stat. Probab. Lett. 2003, 62, 101–110. [Google Scholar] [CrossRef]
  27. Cai, G.H. Strong laws of weighted SHINS of NA randomvariables. Metrika 2008, 68, 323–331. [Google Scholar] [CrossRef]
  28. Yuan, D.M.; An, J. Rosenthal type inequalities for asymptotically almost negatively associated random variables and applications. Chin. Ann. Math. B 2009, 40, 117–130. [Google Scholar] [CrossRef]
  29. Ko, M. Hájek-Rényi inequality for m-asymptotically almost negatively associated random vectors in Hilbert space and applications. J. Inequal. Appl. 2018, 2018, 80. [Google Scholar] [CrossRef]
  30. Ding, S.S.; Li, X.Q.; Dong, X.; Yang, W.Z. The consistency of the CUSUM–type estimator of the change-Point and its application. Mathematics 2020, 8, 2113. [Google Scholar] [CrossRef]
  31. Hušková, M. Gradual changes versus abrupt changes. J. Stat. Plan. Inference 1999, 76, 109–125. [Google Scholar] [CrossRef]
  32. Sun, Z.D.; Huang, Q.; Xue, B. Recent hydrological dynamic and its formation mechanism in Hulun Lake catchment. Arid Land Geogr. 2021, 44, 299–307. [Google Scholar]
  33. Billingsley, P. Convergence of Probability Measures; Wiley: New York, NY, USA, 1968. [Google Scholar]
Figure 1. The box plots of τ ^ * τ * with γ = 0.75 , δ n = n 0.1 , 1 , n 0.1 , n 0.2 , and τ * = 1 / 2 .
Figure 1. The box plots of τ ^ * τ * with γ = 0.75 , δ n = n 0.1 , 1 , n 0.1 , n 0.2 , and τ * = 1 / 2 .
Axioms 12 00894 g001
Figure 2. The box plots of τ ^ * τ * with γ = 0.5 , δ n = n 0.1 , 1 , n 0.1 , n 0.2 , and τ * = 1 / 2 .
Figure 2. The box plots of τ ^ * τ * with γ = 0.5 , δ n = n 0.1 , 1 , n 0.1 , n 0.2 , and τ * = 1 / 2 .
Axioms 12 00894 g002
Figure 3. The box plots of τ ^ * τ * with γ = 0.25 , δ n = n 0.1 , 1 , n 0.1 , n 0.2 , and τ * = 1 / 2 .
Figure 3. The box plots of τ ^ * τ * with γ = 0.25 , δ n = n 0.1 , 1 , n 0.1 , n 0.2 , and τ * = 1 / 2 .
Axioms 12 00894 g003
Figure 4. The plot graph of monthly average water levels of Hulun Lake from 1992 to 2008.
Figure 4. The plot graph of monthly average water levels of Hulun Lake from 1992 to 2008.
Axioms 12 00894 g004
Figure 5. Autocorrelation function.
Figure 5. Autocorrelation function.
Axioms 12 00894 g005
Table 1. The values of k ^ * based on water level.
Table 1. The values of k ^ * based on water level.
γ = 0.1 γ = 0.15 γ = 0.2 γ = 0.25 γ = 0.3 γ = 0.35 γ = 0.4
k ^ * = 105 k ^ * = 105 k ^ * = 105 k ^ * = 105 k ^ * = 105 k ^ * = 105 k ^ * = 105
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, T.; Wei, Y. A Least Squares Estimator for Gradual Change-Point in Time Series with m-Asymptotically Almost Negatively Associated Errors. Axioms 2023, 12, 894. https://doi.org/10.3390/axioms12090894

AMA Style

Xu T, Wei Y. A Least Squares Estimator for Gradual Change-Point in Time Series with m-Asymptotically Almost Negatively Associated Errors. Axioms. 2023; 12(9):894. https://doi.org/10.3390/axioms12090894

Chicago/Turabian Style

Xu, Tianming, and Yuesong Wei. 2023. "A Least Squares Estimator for Gradual Change-Point in Time Series with m-Asymptotically Almost Negatively Associated Errors" Axioms 12, no. 9: 894. https://doi.org/10.3390/axioms12090894

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop